Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary...

5
Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla [email protected] Jogendra Kundu * [email protected] Venkatesh Babu R [email protected] Video Analytics Lab Department of Computational and Data Sciences Indian Institute of Science Bangalore, INDIA ABSTRACT Freehand sketching is an inherently sequential process. Yet, most approaches for hand-drawn sketch recognition either ignore this sequential aspect or exploit it in an ad-hoc man- ner. In our work, we propose a recurrent neural network architecture for sketch object recognition which exploits the long-term sequential and structural regularities in stroke data in a scalable manner. Specifically, we introduce a Gated Recurrent Unit based framework which leverages deep sketch features and weighted per-timestep loss to achieve state- of-the-art results on a large database of freehand object sketches across a large number of object categories. The in- herently online nature of our framework is especially suited for on-the-fly recognition of objects as they are being drawn. Thus, our framework can enable interesting applications such as camera-equipped robots playing the popular party game Pictionary with human players and generating sparsified yet recognizable sketches of objects. Keywords object recognition, sketch, recurrent neural networks, deep features, sequence classification 1. INTRODUCTION The process of freehand sketching has long been employed by humans to communicate ideas and intent in a minimal- ist yet almost universally understandable manner. In spite of the challenges posed in recognizing them [13], sketches have formed the basis of applications in areas of forensic analysis [11], electronic classroom systems [15], sketch-based retrieval [20, 13] etc. Sketching is an inherently sequential process. The pro- liferation of pen and tablet based devices today enables us * equal contribution as first author ACM ISBN . DOI: to capture and analyze the entire process of sketching, thus providing additional information compared to passive pars- ing of static sketched content. Yet, most sketch recognition approaches either ignore the sequential aspect or lack the ability to exploit it [20, 7, 18]. The few approaches which attempt to exploit the sequential sketch stroke data do so either in an unnatural manner [23] or impose restrictive con- straints (e.g. Markov assumption) [1]. In our work, we propose a recurrent neural network archi- tecture for sketch object recognition which exploits the long- term sequential and structural regularities in stroke data in a scalable manner. We make the following contributions: We propose the first deep recurrent neural network ar- chitecture which can recognize freehand sketches across a large number (160) of object categories. Specifically, we introduce a Gated Recurrent Unit (GRU)-based framework (Section 3.1) which leverages deep sketch features and weighted per-stroke loss to achieve state- of-the-art results. We show that the choice of deep sketch features and re- current network architecture both play a crucial role in obtaining good recognition performance (Section 4.3). Via our experiments on sketches with partial temporal stroke content, we show that our framework recognizes the largest percentage of sketches (Section 4.3). Given the on-line nature of our recognition framework, it is especially suited for on-the-fly interpretation of sketches as they are drawn. Thus, our framework can enable interest- ing applications such as camera-equipped robots playing the popular party game Pictionary [16] with human players, gen- erating sparsified yet recognizable sketches of objects [17], interpreting hand-drawn digital content in electronic class- rooms [15] etc. 2. RELATED WORK To retain focus, we review approaches exclusively related to recognition of hand-drawn object sketches. Early datasets tended to contain either a small number of sketches and/or object categories [20, 14]. In 2012, Eitz et al. [7] released a dataset containing 20000 hand-drawn sketches across 250 categories of everyday objects. The dataset, currently the largest sketch object dataset available, provided the first opportunity to attempt the sketch object recognition prob- lem at a relatively large-scale. Since its release, a num- ber of approaches have been proposed to recognize free- hand sketches of objects. The initial performance of hand- crafted feature-based approaches [13, 18] has been recently arXiv:1608.03369v1 [cs.CV] 11 Aug 2016

Transcript of Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary...

Page 1: Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla ravikiran@grads.cds.iisc.ac.in

Enabling My Robot To Play Pictionary : Recurrent NeuralNetworks For Sketch Recognition

Ravi Kiran [email protected]

Jogendra Kundu∗

[email protected] Babu R

[email protected] Analytics Lab

Department of Computational and Data SciencesIndian Institute of Science

Bangalore, INDIA

ABSTRACTFreehand sketching is an inherently sequential process. Yet,most approaches for hand-drawn sketch recognition eitherignore this sequential aspect or exploit it in an ad-hoc man-ner. In our work, we propose a recurrent neural networkarchitecture for sketch object recognition which exploits thelong-term sequential and structural regularities in strokedata in a scalable manner. Specifically, we introduce a GatedRecurrent Unit based framework which leverages deep sketchfeatures and weighted per-timestep loss to achieve state-of-the-art results on a large database of freehand objectsketches across a large number of object categories. The in-herently online nature of our framework is especially suitedfor on-the-fly recognition of objects as they are being drawn.Thus, our framework can enable interesting applications suchas camera-equipped robots playing the popular party gamePictionary with human players and generating sparsified yetrecognizable sketches of objects.

Keywordsobject recognition, sketch, recurrent neural networks, deepfeatures, sequence classification

1. INTRODUCTIONThe process of freehand sketching has long been employed

by humans to communicate ideas and intent in a minimal-ist yet almost universally understandable manner. In spiteof the challenges posed in recognizing them [13], sketcheshave formed the basis of applications in areas of forensicanalysis [11], electronic classroom systems [15], sketch-basedretrieval [20, 13] etc.

Sketching is an inherently sequential process. The pro-liferation of pen and tablet based devices today enables us

∗equal contribution as first author

ACM ISBN .

DOI:

to capture and analyze the entire process of sketching, thusproviding additional information compared to passive pars-ing of static sketched content. Yet, most sketch recognitionapproaches either ignore the sequential aspect or lack theability to exploit it [20, 7, 18]. The few approaches whichattempt to exploit the sequential sketch stroke data do soeither in an unnatural manner [23] or impose restrictive con-straints (e.g. Markov assumption) [1].

In our work, we propose a recurrent neural network archi-tecture for sketch object recognition which exploits the long-term sequential and structural regularities in stroke data ina scalable manner. We make the following contributions:• We propose the first deep recurrent neural network ar-

chitecture which can recognize freehand sketches acrossa large number (160) of object categories. Specifically,we introduce a Gated Recurrent Unit (GRU)-basedframework (Section 3.1) which leverages deep sketchfeatures and weighted per-stroke loss to achieve state-of-the-art results.• We show that the choice of deep sketch features and re-

current network architecture both play a crucial role inobtaining good recognition performance (Section 4.3).• Via our experiments on sketches with partial temporal

stroke content, we show that our framework recognizesthe largest percentage of sketches (Section 4.3).

Given the on-line nature of our recognition framework, itis especially suited for on-the-fly interpretation of sketchesas they are drawn. Thus, our framework can enable interest-ing applications such as camera-equipped robots playing thepopular party game Pictionary [16] with human players, gen-erating sparsified yet recognizable sketches of objects [17],interpreting hand-drawn digital content in electronic class-rooms [15] etc.

2. RELATED WORKTo retain focus, we review approaches exclusively related

to recognition of hand-drawn object sketches. Early datasetstended to contain either a small number of sketches and/orobject categories [20, 14]. In 2012, Eitz et al. [7] releaseda dataset containing 20000 hand-drawn sketches across 250categories of everyday objects. The dataset, currently thelargest sketch object dataset available, provided the firstopportunity to attempt the sketch object recognition prob-lem at a relatively large-scale. Since its release, a num-ber of approaches have been proposed to recognize free-hand sketches of objects. The initial performance of hand-crafted feature-based approaches [13, 18] has been recently

arX

iv:1

608.

0336

9v1

[cs

.CV

] 1

1 A

ug 2

016

Page 2: Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla ravikiran@grads.cds.iisc.ac.in

Figure 1: Overview of our sketch-recognition framework. Alexnet-based deep features (x1 ,x2 , . . . ,xt , . . .) fromeach cumulative stroke image form the input to the GRU (blue), obtaining the corresponding predictionsequence (y1 ,y2 , . . . ,yt , . . .). The per-timestep loss lt is computed w.r.t ground-truth (green arrow) by the lossfunction (yellow box). This loss is weighted by a corresponding wt (shown as proportionally sized gray circles)and backpropagated (purple curved arrow) for the corresponding time step t. Best viewed in color.

surpassed by deep feature-based approaches [17, 19], cul-minating in an custom-designed Convolutional Neural Net-work dubbed SketchCNN which achieved state-of-the-art re-sults [23]. The approaches mentioned above are primarilydesigned for static, full-sketch object recognition. In con-trast, another set of approaches attempt to exploit the se-quential stroke-by-stroke nature of hand-drawn sketch cre-ation [1, 22]. For example, Arandjelovic and Sezgin [1]propose a Hidden Markov Model (HMM)-based approachfor recognizing military and crisis management symbol ob-jects. Although mentioned above in the context of staticobject recognition, a variant of the SketchCNN [23] can alsohandle sequential stroke data. In fact, the authors demon-strate that exploiting the sequential nature of sketching pro-cess improves the overall recognition rate. However, giventhat CNNs are not inherently designed to preserve sequen-tial “state”, better results can be expected from a frame-work which handles sequential data in a more natural fash-ion. The approach we present in our paper aims to do pre-cisely this. Our framework is based on Gated RecurrentUnit (GRU) networks recently proposed by Cho et al. [3].GRU architectures share a number of similarities with themore popular Long Short Term Memory Networks [9] includ-ing the latter’s ability to perform better [8] than traditionalmodels (e.g. HMM) for problems involving long and compli-cated sequential structures. To the best of our knowledge,recurrent neural networks have not been utilized for onlinesketch recognition.

3. OUR RECOGNITION FRAMEWORK3.1 Overview

Sketch creation involves accumulation of hand-drawn strokesover time. Thus, we require our recognition framework tooptimally exploit object category evidence being accumu-lated on a per-stroke basis as well as temporally. Moreover,the variety in sketch-based depiction and intrinsic represen-tational complexity of objects results in a large range forstroke-sequence lengths. Therefore, we require our recogni-

tion framework to address this variation in sequence lengthsappropriately. To meet these requirements, we employ GatedRecurrent Unit (GRU) networks [3]. Our choice of GRU ar-chitecture is motivated by the observation that it involveslearning a smaller number of parameters and performs bet-ter compared to LSTM in certain instances [4] including, asshall been seen (Section 4), our problem of sketch recogni-tion as well.

A GRU network learns to map an input sequence X =(x1,x2 . . .xN) to an output sequence Y = (y1,y2 . . .yN).This mapping is performed by the following transformationswhich are applied at each time step:

rt = σ(Wxrxt +Whrht−1 + br) (1)

zt = σ(Wxzxt +Whzht−1 + bz) (2)

ht = tanh(Wxhxt + U(rt � ht−1) + bh) (3)

ht = (1− zt)� ht−1 + zt � ht (4)

yt = Whyht (5)

Here, xt and yt represent the t-th input and t-th outputrespectively, h represents the “hidden” sequence state of theGRU whose contents are regulated by parameterized gating

units r, z, h and � represents the elementwise dot-product.The subscripted W s, bs and U represent the trainable pa-rameters of the GRU. Please refer to Chung et al. [5] fordetails.

For each sketch, information is available at temporal strokelevel. We use this to construct an image sequence S =(s1, s2 . . . sN ) of sketch strokes cumulatively accumulatedover time. Thus, sN represents the full, final object sketchand N represents the number of sketch strokes or equiva-lently, time-steps (see Figure 1). To represent the strokecontent for each si ∈ S, we utilize deep features obtainedwhen si is provided as input to Alexnet1 [12]. The resulting

1Specifically, we remove classification layer and use outputsfrom final fully-connected layer of the resulting net as fea-

Page 3: Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla ravikiran@grads.cds.iisc.ac.in

Figure 2: Comparison of online recognition per-formance for various classifiers. Our architecturerecognizes the largest % of sketches at all levels ofsketch completion. Best viewed in color.

deep feature sequence X = (x1,x2 . . . ,xN) forms the inputsequence to GRU (see Figure 1). The GRU unit contains3600 hidden units and its output is densely connected to afinal softmax layer for classification. For better generaliza-tion, we include a dropout layer before the final classifica-tion layer which tends to benefit recurrent networks havinga large number of hidden units. We used a dropout of 0.5in our experiments.

3.2 TrainingOur architecture produces an output prediction yt for ev-

ery time-step t, 1 ≤ t ≤ N in the sequence. By comparingthe predictions yt with the ground-truth, we can determinethe corresponding loss lt for a fixed loss function (shown asa yellow box in Figure 1). This loss is weighted by a cor-responding wt and backpropagated) for the correspondingtime step t. For the weighing function, we use

wt = e−α(1−tN

) (6)

Thus, losses corresponding to final stages of sequence areweighted more to encourage correct prediction of the fullsketch. Also, since wt is non-zero, our design incorporateslosses from all steps of the sequence. This has the net effectof encouraging correct predictions even in the early stages ofthe sequence. Overall, this feature enables our recognitionframework to be accurate and responsive right from the be-ginning of the sketching process (Section 4) in contrast withframeworks which need to wait for the sketching to finish be-fore analysis can begin. We additionally studied variationsof the weighing function given in Equation (6) – using thefinal sequence member loss (i.e. wt = 0 ∀t 6= N) and linearlyweighted losses (i.e. wt = t

N). We found that exponentially

weighted loss2 gave superior results.To address the high variation of sequence length across

sketches, we create batches of sketches having equal sequencelength (i.e. N (Sec. 3.1)). These batches of varying size arerandomly shuffled and delivered to the recurrent networkduring training. For each batch, categorical-cross-entropyloss is generated for each sequence by comparing the pre-dictions with the ground-truth. The resulting losses areweighted (Equation (6)) on a per-timestep basis as describedpreviously and back-propagated through the corresponding

tures.2We used α = 10 for our experiments.

CNNRecurrentNetwork #Hidden

Avg.Acc

Alexnet-FC GRU 3600 85.1%Alexnet-FC LSTM 3600 82.5%

SketchCNN [23] - - 81.4%Alexnet-FT - - 83.9%

SketchCNN-Sch-FC LSTM 3600 78.8%SketchCNN-Sch-FC GRU 3600 79.1%

Table 1: Average recognition accuracy (rightmostcolumn) for various architectures. #Hidden refersto the number of hidden units used in recurrent net-work. We obtain state-of-the-art results for sketchobject recognition.

sequence during training. We used stochastic gradient de-scent with a learning rate of 0.001 for training.

3.3 Category predictionSuppose for a given input sequence X = (x1,x2 , . . . ,xN),

the corresponding outputs at the softmax layer are p1,p2, . . . ,pN.Note that in our case, xt ∈ Rd where d is the dimension ofdeep feature xt and pt ∈ RK where K is the number ofobject categories (160). To determine the final category la-bel cX, we perform a weighted sum-pooling of the softmaxoutputs as cX = arg maxj

∑Nt=1 pt

jwt where wt is as givenin Equation (6) and 1 ≤ j ≤ K. We explored various othersoftmax output pooling schemes – last sequence member-based prediction (cX = arg maxj pN

j), max-pooling (cX =

arg maxj [maxt

ptj ]), mean-pooling (cX = arg maxj

∑Nt=1 pt

j/N).

From our validation experiments, we found weighted sum-pooling to be the best choice overall.

4. EXPERIMENTS

4.1 DataIn addition to obtaining the best results among approaches

using handcrafted features, the work of Rosalia et al. [18] wasespecially instrumental in identifying a 160-category sub-set of the TU Berlin dataset which could be unambiguouslyrecognized by humans. Consequently, our experiments arebased on this curated 160-category subset of sketches. Fol-lowing Rosalia et al. [18], we use 56 sketches from each of the160 categories. To ensure principled evaluation, we split the56 sketches of each category randomly into sets containing57%, 18% and 25% of sketches to be used for training, valida-tion and testing respectively3. Additionally, we utilized thevalidation set exclusively for making choices related to ar-chitecture and parameter settings and performed a one-shotcomparative evaluation of ours and competing approacheson the test set.

4.2 Other architecturesWe compared our performance with the following archi-

tectures:Alexnet-FT: As a baseline experiment, we fine-tuned

Alexnet using our 160-category training data. To ensuresufficient data, we augmented the training data on the linesof Sarvadevabhatla et al. [17]. We also used the final fully-connected 4096-dimensional layer features as input to our

3Thus, we have 32, 10 and 14 sketches from each categoryin the training, validation and test sets respectively.

Page 4: Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla ravikiran@grads.cds.iisc.ac.in

Figure 3: Examples of sketches misclassified by our framework. The ground-truth category is top label (blue)while the bottom label is our prediction (pink). Most of the misclassifications are reasonable errors.

recurrent architectures. We shall refer to such usage byAlexnet-FC.

SketchCNN: This is essentially the deep architecture ofYu et al. [23] but retrained for the categories and splits men-tioned in Section 4.1. Since CNNs do not inherently store“state”, the authors construct six different sub-sequence strokeaccumulation images which comprise the channels of the in-put representation to the CNNs. It comprises of five dif-ferent CNNs, each trained for five different scaled versionsof 256 × 256 sketches. The last fully-connected layer’s 512-dimensional features from all the five CNNs are processedusing a Bayesian fusion technique to obtain the final classi-fication.

For our experiments, we also concatenated the 512 dimen-sional features from each scale of SketchCNN as the inputfeature to the recurrent neural network architectures thatwere evaluated. However, only the full 256× 256 sketch wasconsidered as the input to CNN (i.e. single-channel). For therest of the paper, we refer to the resulting 2560-dimensionalfeature as SketchCNN-SCh-FC.

Recurrent architectures: We experimented with thenumber of hidden units, the number of recurrent layers, thetype of recurrent layers (i.e. LSTM or GRU), the trainingloss function (Section 3.2) and various pooling methods forobtaining final prediction in terms of individual sequencemember predictions (Section 3.3).

We built the software framework for our proposed archi-tecture using Lasagne [6] and Theano [2] libraries. We alsoused MatConvNet [21] and Caffe [10] libraries for experi-ments related to other competing architectures.

4.3 ResultsOverall performance: Table 1 summarizes the over-

all performance in terms of average recognition accuracy forvarious architectures. As can be seen, our GRU-based archi-tecture (first row) outperforms SketchCNN by a significantmargin even though it is trained on only 57% of the totaldata. We believe our good performance stems from (a) beingable to exploit the sequential information in a scalable andefficient manner via recurrent neural networks (b) the supe-riority of the deep sketch features provided by Alexnet [17]compared to the SketchCNN-FC features. The latter can beclearly seen when we compare the first two rows of Table 1with the last two rows. In our case, the performance of GRUwas better than that of LSTM when Alexnet features were

used. Overall, it is clear that the choice of (sketch) featuresand the recurrent network both play a crucial role in obtain-ing state-of-the-art performance for the sketch recognitiontask.

On-line recognition: We also compared the various ar-chitectures for their ability to recognize sketches as they arebeing drawn (i.e. on-line recognition performance). For eachclassifier, we determined the fraction of test sketches tx cor-rectly recognized when only the first x% of the temporalsketch strokes are available. We varied x between 20 to 100in steps of 20 and plotted tx as a function of x. The resultscan be seen in Figure 2. Intuitively, the higher a curve onthe plot, the better its online recognition ability. As can beseen, our framework consistently recognizes a larger fractionof sketches at all levels of sketch completion (except for verysmall x) relative to other architectures.

Semantic information: To determine the extent to whichour architecture captures semantic information, we exam-ined the performance of the classifier on misclassified sketches.As can be seen in Figure 3, most of the misclassifications arereasonable errors (e.g. guitar is mistaken for violin) anddemonstrate that our framework learns the overall semanticsof the object recognition problem.

5. CONCLUSIONIn this paper, we have presented our deep recurrent neural

network architecture for freehand sketch recognition. Ourarchitecture has two prominent traits. Firstly, its design ac-counts for the inherently sequential and cumulative natureof human sketching process in a natural manner. Secondly,it exploits long-term sequential and structural regularities instroke data represented as deep features. These two traitsenable our system to achieve state-of-the-art recognition re-sults on a large database of freehand object sketches. Wehave also shown that our recognition framework is highlysuitable for on-the-fly interpretation of sketches as they arebeing drawn. Our framework source-code and associateddata (pre-trained models) can be accessed at https://github.com/val-iisc/sketch-obj-rec.

6. ACKNOWLEDGEMENTSWe thank NVIDIA for their grant of Tesla K40 GPU.

7. REFERENCES

Page 5: Enabling My Robot To Play Pictionary : Recurrent Neural ... · Enabling My Robot To Play Pictionary : Recurrent Neural Networks For Sketch Recognition Ravi Kiran Sarvadevabhatla ravikiran@grads.cds.iisc.ac.in

[1] R. Arandjelovic and T. M. Sezgin. Sketch recognitionby fusion of temporal and image-based features.Pattern Recognition, 44(6):1225–1234, 2011.

[2] F. Bastien, P. Lamblin, R. Pascanu, J. Bergstra, I. J.Goodfellow, A. Bergeron, N. Bouchard, andY. Bengio. Theano: new features and speedimprovements. Deep Learning and UnsupervisedFeature Learning NIPS 2012 Workshop, 2012.

[3] K. Cho, B. van Merrienboer, C. Gulcehre,F. Bougares, H. Schwenk, and Y. Bengio. Learningphrase representations using RNN encoder-decoder forstatistical machine translation. CoRR, abs/1406.1078,2014.

[4] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio.Empirical evaluation of gated recurrent neuralnetworks on sequence modeling. arXiv preprintarXiv:1412.3555, 2014.

[5] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio. Gatedfeedback recurrent neural networks. arXiv preprintarXiv:1502.02367, 2015.

[6] S. Dieleman, M. Heilman, J. Kelly, M. Thoma,K. Rasul, E. Battenberg, H. Weideman, S. K.Sønderby, C. Raffel, J. Degrave, et al. Lasagne: Firstrelease. 2015.

[7] M. Eitz, J. Hays, and M. Alexa. How do humanssketch objects? ACM Trans. Graph. (Proc.SIGGRAPH), 31(4):44:1–44:10, 2012.

[8] A. Graves, M. Liwicki, S. Fernandez, R. Bertolami,H. Bunke, and J. Schmidhuber. A novel connectionistsystem for unconstrained handwriting recognition.Pattern Analysis and Machine Intelligence, IEEETransactions on, 31(5):855–868, 2009.

[9] S. Hochreiter and J. Schmidhuber. Long short-termmemory. Neural computation, 9(8):1735–1780, 1997.

[10] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev,J. Long, R. Girshick, S. Guadarrama, and T. Darrell.Caffe: Convolutional architecture for fast featureembedding. arXiv preprint arXiv:1408.5093, 2014.

[11] B. F. Klare, Z. Li, and A. K. Jain. Matching forensicsketches to mug shot photos. Pattern Analysis andMachine Intelligence, IEEE Transactions on,33(3):639–646, 2011.

[12] A. Krizhevsky, I. Sutskever, and G. E. Hinton.Imagenet classification with deep convolutional neuralnetworks. In Advances in neural informationprocessing systems, pages 1097–1105, 2012.

[13] Y. Li, Y.-Z. Song, and S. Gong. Sketch recognition byensemble matching of structured features. BMVC,2013.

[14] A. Lovett, M. Dehghani, and K. Forbus. Efficientlearning of qualitative descriptions for sketchrecognition. In Proceedings of the 20th InternationalQualitative Reasoning Workshop, 2006.

[15] B. Meyer, K. Marriott, A. Bickerstaffe, andL. Knipping. Intelligent diagramming in the electroniconline classroom. In Human System Interactions,2009. HSI’09. 2nd Conference on, pages 177–183.IEEE, 2009.

[16] Pictionary. https://en.wikipedia.org/wiki/Pictionary.

[17] R. K. Sarvadevabhatla and V. B. Radhakrishnan. Eye

of the dragon: Exploring discriminatively minimalistsketch-based abstractions for object categories. InProceedings of the 23rd ACM International Conferenceon Multimedia, MM ’15, pages 271–280, New York,NY, USA, 2015. ACM.

[18] R. G. Schneider and T. Tuytelaars. Sketchclassification and classification-driven analysis usingfisher vectors. ACM Trans. Graph., 33(6):174:1–174:9,Nov. 2014.

[19] O. Seddati, S. Dupont, and S. Mahmoudi. Deepsketch:deep convolutional neural networks for sketchrecognition and similarity search. In Content-BasedMultimedia Indexing (CBMI), 2015 13th InternationalWorkshop on, pages 1–6. IEEE, 2015.

[20] Z. Sun, C. Wang, L. Zhang, and L. Zhang.Query-adaptive shape topic mining for hand-drawnsketch recognition. In Proceedings of the 20th ACMinternational conference on Multimedia, pages519–528. ACM, 2012.

[21] A. Vedaldi and K. Lenc. Matconvnet: Convolutionalneural networks for matlab. In Proceedings of the 23rdAnnual ACM Conference on Multimedia Conference,pages 689–692. ACM, 2015.

[22] X. Wang, H. Xu, and H. Wang. On-line sketchrecognition for course of action diagrams. InMechatronics and Automation (ICMA), 2010International Conference on, pages 465–469. IEEE,2010.

[23] Q. Yu, Y. Yang, Y.-Z. Song, T. Xiang, andT. Hospedales. Sketch-a-net that beats humans.BMVC, 2015.