BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR...

5
BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng Mao Yuefeng Chen Yuhong Li Tao Xiong Yuan He Hui Xue Alibaba Group, China {mxf164419, yuefeng.chenyf, daniel.lyh, weilue.xt, heyuan.hy, hui.xueh}@alibaba-inc.com ABSTRACT The task of Language-Based Image Editing (LBIE) aims at generating a target image by editing the source image based on the given language description. The main challenge of LBIE is to disentangle the semantics in image and text and then combine them to generate realistic images. Therefore, the editing performance is heavily dependent on the learned representation. In this work, conditional generative adversar- ial network (cGAN) is utilized for LBIE. We find that existing conditioning methods in cGAN lack of representation power as they cannot learn the second-order correlation between two conditioning vectors. To solve this problem, we propose an improved conditional layer named Bilinear Residual Layer (BRL) to learning more powerful representations for LBIE task. Qualitative and quantitative comparisons demonstrate that our method can generate images with higher quality when compared to previous LBIE techniques. Index TermsGenerative adversarial networks, Bilin- ear, Language-based image editing 1. INTRODUCTION The task of Language-Based Image Editing (LBIE) aims at manipulating a source image semantically to match the given description well. LBIE has seen applications to domains as diverse as Computer-Aided Design (CAD), Fashion Genera- tion and Virtual Reality (VR) [1]. As illustrated in Fig 1, us- ing LBIE technique, one can automatically modify the color, texture or style for a given design drawing by language in- structions instead of the traditional complex processes. Nevertheless, LBIE is still challenging due to the follow- ing two difficulties: i). the model should find the areas in image which are relevant to the given text description; ii). the relations of disentangled semantics in image and text de- scription should be learned for a better generation of realis- tic image. To tackle these problems, several methods have been proposed [1, 2, 3, 4, 5], and most of them utilize the generative models, e.g., GANs [6]. [1, 2] divide LBIE into two subtasks: language-based image segmentation and image generation. Specifically, Zhu et al. [2] performes LBIE to “re- dress” the person with the given outfit description, while at the The lady wore a white sleeveless dress + Fig. 1. LBIE for fashion generation. same time keeping the wearer and his posture or expression unchanged. They use a two stages GAN that outputs a seman- tic segmentation map as intermediate step, which is further used to render the final image with precise regions and tex- tures at the second step. Some other approaches [3, 4, 5] can achieve LBIE without any segmentation map or explicit spa- tial constraints by adversarially train a conditional GAN [7]. Among them, [5] is the seminal work and it uses concatena- tion operation to condition the image generation process with text embeddings. [3, 4] follow up this framework, and replace the concatenation operation with Feature-wise Linear Modu- lation (FiLM), which is a more efficient and powerful method as a generalization of concatenation. In this work, we first theoretically analyse these works which edit the image based on fused visual-text representa- tions using different conditioning methods. We found that all these conditioning methods can be modeled by a univer- sal form of bilinear transformation based on [8]. However, all these methods are lack of representation power as they cannot learn the second-order correlation between two conditioning embeddings. To solve this problem, we present an improved conditoning method named Bilinear Residual Layer which can strike a happy compromise between representation ef- fectiveness and model efficiency. We have both theoretically and experimentally proved that the Bilinear Residual Layer can provide richer representations than previous approaches. Quantitative and qualitative results on Caltech-200 bird [9], Oxford-102 flower [10] and Fashion Synthesis datasets [2] suggest that our approach can generate images with higher quality when compared to previous LBIE techniques. 2. METHOD In this section, we first theoretically analyse existing condi- tioning methods in cGANs. Then an improved conditional arXiv:1903.07499v1 [cs.CV] 18 Mar 2019

Transcript of BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR...

Page 1: BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng

BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USINGCONDITIONAL GENERATIVE ADVERSARIAL NETWORKS

Xiaofeng Mao Yuefeng Chen Yuhong Li Tao Xiong Yuan He Hui Xue

Alibaba Group, China{mxf164419, yuefeng.chenyf, daniel.lyh, weilue.xt, heyuan.hy, hui.xueh}@alibaba-inc.com

ABSTRACT

The task of Language-Based Image Editing (LBIE) aimsat generating a target image by editing the source image basedon the given language description. The main challenge ofLBIE is to disentangle the semantics in image and text andthen combine them to generate realistic images. Therefore,the editing performance is heavily dependent on the learnedrepresentation. In this work, conditional generative adversar-ial network (cGAN) is utilized for LBIE. We find that existingconditioning methods in cGAN lack of representation poweras they cannot learn the second-order correlation between twoconditioning vectors. To solve this problem, we propose animproved conditional layer named Bilinear Residual Layer(BRL) to learning more powerful representations for LBIEtask. Qualitative and quantitative comparisons demonstratethat our method can generate images with higher quality whencompared to previous LBIE techniques.

Index Terms— Generative adversarial networks, Bilin-ear, Language-based image editing

1. INTRODUCTION

The task of Language-Based Image Editing (LBIE) aims atmanipulating a source image semantically to match the givendescription well. LBIE has seen applications to domains asdiverse as Computer-Aided Design (CAD), Fashion Genera-tion and Virtual Reality (VR) [1]. As illustrated in Fig 1, us-ing LBIE technique, one can automatically modify the color,texture or style for a given design drawing by language in-structions instead of the traditional complex processes.

Nevertheless, LBIE is still challenging due to the follow-ing two difficulties: i). the model should find the areas inimage which are relevant to the given text description; ii).the relations of disentangled semantics in image and text de-scription should be learned for a better generation of realis-tic image. To tackle these problems, several methods havebeen proposed [1, 2, 3, 4, 5], and most of them utilize thegenerative models, e.g., GANs [6]. [1, 2] divide LBIE intotwo subtasks: language-based image segmentation and imagegeneration. Specifically, Zhu et al. [2] performes LBIE to “re-dress” the person with the given outfit description, while at the

The lady wore a white sleeveless dress

+

Fig. 1. LBIE for fashion generation.

same time keeping the wearer and his posture or expressionunchanged. They use a two stages GAN that outputs a seman-tic segmentation map as intermediate step, which is furtherused to render the final image with precise regions and tex-tures at the second step. Some other approaches [3, 4, 5] canachieve LBIE without any segmentation map or explicit spa-tial constraints by adversarially train a conditional GAN [7].Among them, [5] is the seminal work and it uses concatena-tion operation to condition the image generation process withtext embeddings. [3, 4] follow up this framework, and replacethe concatenation operation with Feature-wise Linear Modu-lation (FiLM), which is a more efficient and powerful methodas a generalization of concatenation.

In this work, we first theoretically analyse these workswhich edit the image based on fused visual-text representa-tions using different conditioning methods. We found thatall these conditioning methods can be modeled by a univer-sal form of bilinear transformation based on [8]. However, allthese methods are lack of representation power as they cannotlearn the second-order correlation between two conditioningembeddings. To solve this problem, we present an improvedconditoning method named Bilinear Residual Layer whichcan strike a happy compromise between representation ef-fectiveness and model efficiency. We have both theoreticallyand experimentally proved that the Bilinear Residual Layercan provide richer representations than previous approaches.Quantitative and qualitative results on Caltech-200 bird [9],Oxford-102 flower [10] and Fashion Synthesis datasets [2]suggest that our approach can generate images with higherquality when compared to previous LBIE techniques.

2. METHOD

In this section, we first theoretically analyse existing condi-tioning methods in cGANs. Then an improved conditional

arX

iv:1

903.

0749

9v1

[cs

.CV

] 1

8 M

ar 2

019

Page 2: BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng

layer called Bilinear Residual Layer (BRL) is proposed inSec 2.2. Finally, we introduce overall framework in Sec 2.3.

2.1. Overview of conditioning methods

Conditioning is a general-purpose operation and can be usedfor different tasks, e.g., conditional image generation [11, 12]and cross-modality distillation [13]. The most commonlyused approach in conditional GANs is concatenation. For-mally, denote If ∈ RD and Ic ∈ RD′

as the output ofprevious layer and conditioning feature respectively, whereD and D′ are the dimensionality of features. The concatedrepresentation [If Ic] ∈ RD+D′

can be further encoded by amatrix W = [Wf ;Wc], Wf ∈ RD×O and Wc ∈ RD′×O arethe corresponding weights for If and Ic. O is the output di-mension. Formally, we can get the following transformation:

Io = [If Ic] [Wf ;Wc] = IfWf + IcWc (1)

where Io is the output tensor. Equation 1 suggests that con-catenation based conditioning method amounts to adding afeature-wise bias on the unconditional output IfWf . There-fore, some other approaches [14, 15] suggest to add condi-tional bias directly instead of concatenation.

Recently, some works [16] have validated that deep mod-els could mimic the human attention mechanism by gatingeach feature using a value between 0 and 1. Inspired by this,Perez et al. [17] proposes a more general conditioning methodnamed feature-wise linear modulation (FiLM), which rescalesthe features by adding multiplicative interactions:

Io = (IfWf )� (IcW c) + IcWc (2)

W c ∈ RD′×O is the weight for learning rescaling coeffi-cients. From this formulation, we can conclude that concate-nation is a special case of FiLM when IcW c = 1, where 1is a matrix of ones. FiLM has shown its superiority over con-ventional concatenation method and has been widely appliedto the multimodal interaction.

However, concatenation and FiLM only apply a lineartransformation between the input and conditional features.In this work, we go a step further and generalize these lin-ear methods to the more powerful bilinear version, which canprovide richer representations than linear models by learningthe second-order interaction. In bilinear model, the ith featurein output Io can be calculated as

Ioi = IfWiITc (3)

Wi ∈ RD×D′is a weight matrix for the output feature Ioi .

Interestingly, we have found FiLM can be presented by bilin-ear transformations. Denote the weights corresponding to theith output feature in Wf , W c and Wc as wfi , wci and wci .The FiLM transformation for Ioi = (Ifwfi)(Icwci) + Icwci

can be represented by

If (wfiwTci +Wi

′︸ ︷︷ ︸Wi

)ITc (4)

where IfWi′ = wT

ci , Wi′ can be constructed by randomly

choosing a nonzero element Ifk in If , we have

Wi′ =

0 0 0 · · ·...

...wci1

Ifk

wci2

Ifk· · · wc

iD′Ifk

......

0 0 0 · · ·

(5)

where elements in the Wi′ are 0 except for the kth row.

Obviously, the rank of matrix wfiwTci and Wi

′ are both 1.So we have Rank(Wi) ≤ 2∗. The constructed formulationindicates that FiLM is equivalent to bilinear transformationwith transformation matrix Wi is sparse and has the rankno greater than 2. From a theoretical perspective, it illus-trates that bilinear transformations can provide more fine-grained conditioning representations than the concatenationand FiLM.

2.2. Bilinear Residual Layer

We propose Bilinear Residual Layer (BRL) for learning con-ditional bilinear representations as illustrated in dashed boxof Fig 2. Similar to FiLM, we add shortcuts to guaranteethe model’s capability to learn identical mapping. As a con-sequence, our bilinear residual layer can automaticly decidewhether or not the model needs to incorporate the condition-ing information in the later layers.

However, the representational power of bilinear featurescomes with the cost of very high-dimensional model parame-ters , which require substantial computing and large quantitiesof training data to fit [18]. For example, the dimensionality ofW is |D × D′ × O| which is cubical expansion. To reducethe dimensionality of model parameters, our approach adoptsa low-rank bilinear method [19] to reduce the rank of Wi.Based on this idea, Ioi can be rewritten as follows:

Ioi = IfWiITc = IfUiVi

T ITc = IfUi � IcVi (6)

where Ui ∈ RD×d and Vi ∈ RD′×d are the decomposedsubmatrices and they restrict the rank of Wi to be at mostd ≤ min(D,D′). Then the final feature vector Io can beprojected by P ∈ RO×d as follows:

Io = P(IfU� IcV) (7)

Moreover, Our bilinear residual layer is a general condi-tion layer, and it is applicable not only for LBIE, but also forother conditional models or applications, e.g., text-to-imagegeneration [20]. In following sections, we will present theoverall framework of our work and we denote the bilinearresidual layer as F for convenience.

∗Properties of rank: https://en.wikipedia.org/wiki/Rank (linear algebra)

Page 3: BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng

GRU

GRUthe

lady

GRU

GRU

GRU

GRU

wore

a

GRU

pink

sleeveless

dress

ImageEncoder

BRL 1

BRL 2

BRL N

Decoder

Discriminator

FC

Conv 1

BN 1

ReLU

Conv 2

BN 2

ReLU

V

Inputs

U

P

Low-rank bilinear

Text

En

cod

er

Fusing module

ϕ

down

up

Fig. 2. Overview of our network architecture. Detail of ourbilinear residual layer is presented in the dashed box.

2.3. Overall framework

We follow the work of Dong et al. [5] which utilizes thecGAN to learn the target mapping conditioning based onimage and text description. As shown in Fig 2, the networkconsists of a generator G and a discriminator D. The gen-erator has three modules: encoding module, fusing moduleand decoding module. Encoding module contains pre-trainedencoders ϕ and φdown, and they are used to extract text andimage features respectively. We adopt the procedure in [21]to pre-train the text encoder ϕ and use parameters of conv1-4layers in VGG16 as the feature extractor φdown for image.The text and image features are then fed in the followingfusing module, which can be seen as a conditioning layerto compromise the semantics of multiple modalities. Thefinal decoding module φup upsamples the fused feature to ahigh-resolution images. Finally, the discriminator is a clas-sifier which takes the generated image and text embeddingsas input and output the probability whether the descriptionmatches the image.

Formally, given an original image-text pair <x, t>, t isthe text matching with the image x. Suppose that we use de-scription text t to manipulate the image x, typically t is a textrelevant to x. The generator can transform the image accord-ing to text embedding ϕ(t) and output

G(x, ϕ(t)) = φup(F(φdown(x), ϕ(t))) (8)

the discriminator D is trained to distinguish semantically dif-ferentiated image-text pairs. To this end, we need to take amismatching text t as negative sample. Original pair <x, t>,

current editing pair <x, t> and negative pair <x, t> are fedinto discriminator D to minimizing

LD = E(x,t)∼pdata

[D(x, ϕ(t))2

]+ E(x,t)∼pdata

[(D(x, ϕ(t))− 1)2

]+ E(x,t)∼pdata

[D(G(x, ϕ(t)), ϕ(t))2

] (9)

here the objective of the first and second terms is to classifynegative and original real-world image-text pairs. The thirdterm makes D to identify the synthesized image with its edit-ing text as mismatching as possible. Alternately to the train-ing of D, the generator G is trained to generate more seman-tically similar images with the editing text t:

LG = E(x,t)∼pdata

[(D(G(x, ϕ(t)), ϕ(t))− 1)2

](10)

In this work, t and t are selected from the text descriptions ofother images in the dataset.

3. EXPERIMENTS

We conduct experiments on Caltech-200 bird dataset [9],Oxford-102 flower dataset [10] and Fashion Synthesis dataset[2]. The bird dataset has 11,788 images with 200 classesof birds. We split it to 160 training classes and 40 testingclasses. The flower dataset has 8,189 images with 102 classesof flowers, and we split it to 82 training classes and 20 test-ing classes. The fashion dataset has much more classes with78,979 images totally. We choose 3200 classes from 4119 fortraining and the rest for testing.

3.1. Implementation details

The source code has been released†. Our encoder ϕ for textdescriptions is a recurrent network. Given the pair of imageand text <x, t>, the method in [21] is used to pre-train thetext encoder to minimize the pair-wise ranking loss. This pre-trained text encoder encodes the text description t into visual-semantic text representation ϕ(t), which will be further usedin the adversarial training process as detailed in Sec 2.3.

For image encoder, it receives images with size of 64×64as input and output features with dimension of 16×16×512.Text encoder encodes descriptions to the text embeddingswith dimensionality of 128. Our fusing module consists of 4(i.e., N in the Fig. 2) bilinear residual layers. To implementthe low-rank bilinear method, we duplicate the text embed-dings to be of dimension 16×16×128, so as to keep the samespatial size with image features. Then the dimensions ofboth text and image features are reduced to d (cf. Section2.2) by using 1×1 convolutions. The decoding module con-sists of several upsampling layers that transform the learnedrepresentations into 64×64 images. For the discriminator,we first apply convolutional layers to encode the images†https://github.com/vtddggg/BilinearGAN for LBIE

Page 4: BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng

The lady was wearing a blue short-sleevedblouse.

This little bird is mostly white with a black superciliary and primary.

This flower has petals that are yellow at the edges and spottedorange near the center.

Original image Baseline Our method (Bil-R256)FiLMEditing text

Fig. 3. Qualitative comparisons.

into feature representations. We then concatenate the imagerepresentation with text embeddings, then apply two convo-lutional layers to produce final probabilities. Note that we useconcatenation to conditioning for limiting the discriminatorcapability to prevent the mode collapse effect.

To train the generator and discriminator, we adopt theAdam optimizer with momentum of 0.5. The learning rateis 0.0002. We set batch size to 64 for all three experimentsand number of iterative epochs to 600 for birds and flowerssynthesis, 200 for fashion synthesis. The parameters of VGGpart were fixed during training the generator. The trainingtakes about 1 day to converge on a single Tesla P100 GPU.

3.2. Qualitative comparison

We compare our proposed model with the baseline [5] (i.e.,concatenation) and FiLM on three datasets. The results areshown in Fig 3. Baseline method fails to transform the detailattributes based on the given description because the learnedrepresentations are not powerful as it does not contain enoughdetail information. For example, the generated images bybaseline method in editing flowers demonstrate the model haslearned the colors of yellow and orange, but it is unaware ofthe location of these colors. Meanwhile, when original im-age has a complex background (e.g. 3th and 4th samples infirst row), the model will fall into mode collapse and outputthe same meaningless image. On the contrary, our methodcan capture the specific semantic changes in detail, which isattribute to our richer bilinear representations. It correctly dis-entangles semantically related objects from some messy im-ages and prevent the occurrence of mode collapse. As a con-sequence, our approach can successfully generate meaningfulimages subject to the text description.

3.3. Quantitative comparison

We choose inception score (IS) for quantitative evaluation.Inception score is a well-known metric for evaluating GANs.IS can be computed by IS = exp(ExDKL(p(y|x)‖p(y))),where x denotes one generated sample, and y is the label pre-dicted by the inception model. The better models which gen-

Methods Caltech bird Oxford flower FashionBaseline 1.92±0.05 5.03±0.62 8.65±1.33FiLM [4] 2.59±0.11 4.83±0.48 8.78±1.43Bil-R2 2.60±0.11 4.93±0.39 9.30±1.48Bil-R64 2.63±0.17 5.40±0.62 10.94±2.28

Bil-R256 2.76±0.08 6.26±0.44 11.63±2.15

Table 1. The comparison of IS score of methods

erate diverse and meaningful images can get larger inceptionscore. In this experiment, we use the test dataset for evalua-tion. We first finetune the inception model with test imagesfor classification. Then, for every test class, we randomlychoose an image and text description (e.g. if test dataset has40 classes, 40 images and 40 descriptions are selected). Theimages are generated by inputting every pair of images anddescriptions (e.g. 40 images and 40 descriptions can generate40×40 edited images).

The results are shown in Table 1. To explore the influ-ence of rank constraint d, we set d = 2, 64, 256 and get threevariants: Bil-R2, Bil-R64 and Bil-R256. The Bil-R256 getsthe highest IS in all three tasks. Interestingly, the baselinemethod has higher IS than FiLM on Oxford flower datasetbecause flower editing is simple and is not very dependent onthe power of learned representation. For more complicatedbird and fashion editings, our method gets the highest IS andachieves better performance with the increasing of d. Experi-mental result suggests that the learned bilinear representationis more powerful and do help to generate images with higherquality.

4. CONCLUSION

In this work, we propose a conditional GAN based encoder-decoder architecture to semantically manipulate images bytext descriptions. A general condition layer called BilinearResidual Layer (BRL) is proposed to learn more powerfulbilinear representations for LBIE. BRL is also applicablefor other common conditional tasks. Our evaluation resultson Caltech-200 bird dataset, Oxford-102 flower dataset andFashion Synthesis dataset achieve plausible effects and out-perform the state-of-art methods on LBIE.

Page 5: BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING … · BILINEAR REPRESENTATION FOR LANGUAGE-BASED IMAGE EDITING USING CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS Xiaofeng

5. REFERENCES

[1] Jianbo Chen, Yelong Shen, Jianfeng Gao, JingjingLiu, and Xiaodong Liu, “Language-based image edit-ing with recurrent attentive models,” arXiv preprintarXiv:1711.06288, 2017.

[2] Shizhan Zhu, Sanja Fidler, Raquel Urtasun, Dahua Lin,and Chen Change Loy, “Be your own prada: Fashionsynthesis with structural coherence,” in Computer Vi-sion (ICCV), 2017 IEEE International Conference on.IEEE, 2017, pp. 1689–1697.

[3] Varun Manjunatha, Mohit Iyyer, Jordan Boyd-Graber,and Larry Davis, “Learning to color from language,”in Proceedings of the 2018 Conference of the NorthAmerican Chapter of the Association for ComputationalLinguistics: Human Language Technologies, Volume 2(Short Papers), 2018, vol. 2, pp. 764–769.

[4] Mehmet Gunel, Erkut Erdem, and Aykut Erdem,“Language guided fashion image manipulationwith feature-wise transformations,” arXiv preprintarXiv:1808.04000, 2018.

[5] Hao Dong, Simiao Yu, Chao Wu, and Yike Guo, “Se-mantic image synthesis via adversarial learning,” in Pro-ceedings of the IEEE International Conference on Com-puter Vision, 2017, pp. 5706–5714.

[6] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza,Bing Xu, David Warde-Farley, Sherjil Ozair, AaronCourville, and Yoshua Bengio, “Generative adversar-ial nets,” in Advances in neural information processingsystems, 2014, pp. 2672–2680.

[7] Mehdi Mirza and Simon Osindero, “Conditional gener-ative adversarial nets,” arXiv preprint arXiv:1411.1784,2014.

[8] Vincent Dumoulin, Ethan Perez, Nathan Schucher, Flo-rian Strub, Harm de Vries, Aaron Courville, and YoshuaBengio, “Feature-wise transformations,” Distill, 2018,https://distill.pub/2018/feature-wise-transformations.

[9] C. Wah, S. Branson, P. Welinder, P. Perona, and S. Be-longie, “The Caltech-UCSD Birds-200-2011 Dataset,”Tech. Rep., 2011.

[10] M-E. Nilsback and A. Zisserman, “Automated flowerclassification over a large number of classes,” in Pro-ceedings of the Indian Conference on Computer Vision,Graphics and Image Processing, Dec 2008.

[11] Qiang Huang, Philip Jackson, Mark D Plumbley, andWenwu Wang, “Synthesis of images by two-stage gen-erative adversarial networks,” in 2018 IEEE Interna-tional Conference on Acoustics, Speech and Signal Pro-cessing. IEEE, 2018.

[12] Zhu-Liang Chen, Qian-Hua He, Wen-Feng Pang, andYan-Xiong Li, “Frontal face generation from multi-ple pose-variant faces with cgan in real-world surveil-lance scene,” in 2018 IEEE International Conferenceon Acoustics, Speech and Signal Processing (ICASSP).IEEE, 2018, pp. 1308–1312.

[13] Siddharth Roheda, Benjamin S Riggan, Hamid Krim,and Liyi Dai, “Cross-modality distillation: A case forconditional generative adversarial networks,” in 2018IEEE International Conference on Acoustics, Speechand Signal Processing (ICASSP). IEEE, 2018, pp.2926–2930.

[14] Aaron van den Oord, Sander Dieleman, Heiga Zen,Karen Simonyan, Oriol Vinyals, Alex Graves, NalKalchbrenner, Andrew Senior, and Koray Kavukcuoglu,“Wavenet: A generative model for raw audio,” in 9thISCA Speech Synthesis Workshop, pp. 125–125.

[15] Aaron van den Oord, Nal Kalchbrenner, Lasse Espeholt,Oriol Vinyals, Alex Graves, et al., “Conditional im-age generation with pixelcnn decoders,” in Advancesin Neural Information Processing Systems, 2016, pp.4790–4798.

[16] Fei Wang, Mengqing Jiang, Chen Qian, Shuo Yang,Cheng Li, Honggang Zhang, Xiaogang Wang, and Xi-aoou Tang, “Residual attention network for image clas-sification,” in Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, 2017, pp.3156–3164.

[17] Ethan Perez, Florian Strub, Harm De Vries, Vincent Du-moulin, and Aaron Courville, “Film: Visual reasoningwith a general conditioning layer,” in http://www. aaai.org/Conferences/AAAI/aaai. php, 2018.

[18] Shu Kong and Charless Fowlkes, “Low-rank bilinearpooling for fine-grained classification,” in Computer Vi-sion and Pattern Recognition (CVPR), 2017 IEEE Con-ference on. IEEE, 2017, pp. 7025–7034.

[19] Jin-Hwa Kim, Kyoung-Woon On, Woosang Lim,Jeonghee Kim, Jung-Woo Ha, and Byoung-Tak Zhang,“Hadamard product for low-rank bilinear pooling,”arXiv preprint arXiv:1610.04325, 2016.

[20] Han Zhang, Tao Xu, and Hongsheng Li, “Stackgan:Text to photo-realistic image synthesis with stacked gen-erative adversarial networks,” in 2017 IEEE Interna-tional Conference on Computer Vision (ICCV). IEEE,2017, pp. 5908–5916.

[21] Ryan Kiros, Ruslan Salakhutdinov, and Richard SZemel, “Unifying visual-semantic embeddings withmultimodal neural language models,” arXiv preprintarXiv:1411.2539, 2014.