Copyright by Sung Ju Hwang 2013

Size: px
Start display at page:

Download "Copyright by Sung Ju Hwang 2013"

Transcription

1 Copyright by Sung Ju Hwang 2013

2 The Dissertation Committee for Sung Ju Hwang certifies that this is the approved version of the following dissertation: Discriminative Object Categorization with External Semantic Knowledge Committee: Kristen Grauman, Supervisor Fei Sha J. K. Aggarwal Raymond Mooney Pradeep Ravikumar

3 Discriminative Object Categorization with External Semantic Knowledge by Sung Ju Hwang, B.S., M.A. DISSERTATION Presented to the Faculty of the Graduate School of The University of Texas at Austin in Partial Fulfillment of the Requirements for the Degree of DOCTOR OF PHILOSOPHY THE UNIVERSITY OF TEXAS AT AUSTIN August 2013

4 Dedicated to my mom Hyunsook Park

5 Acknowledgments I wish to first thank my advisor Kristen Grauman, who always had great passion in research and insightful ideas in object recognition, and gave me much freedom to pursue any research ideas. She also led me to the direction of a better researcher with seriousness in every aspect, and incredible attention to details. All the accomplishments I have made would not have been possible without her, and I feel extremely lucky to be her student. I also want to thank greatly to my co-adviser Fei Sha, who guided me through the path of machine learning with rigor in mathematics and equally incredible attention to details as Kristen, and especially for his help on the methodology that made the ideas to come to life as working algorithms. I am also very grateful to my thesis committee members Professor Raymond Mooney, Professor J. K. Aggarwal, and Professor Pradeep Ravikumar for their insightful comments and suggestions. I also want to thank my labmates with whom I spent the most time for the last five years. I want to thank Sudheendra for being an exemplary student to follow the step, Yong Jae for his availability to discuss even the vaguest idea and being a good friend, Jaechul for sharing his insight and experience in computer vision. Thanks Adriana for always smiling and being kind, and thank Chao-yeh with for being able to chat for anything, and helping me v

6 out with courseworks. Thank Sunil for his jolliness, and thank Lu Zheng for sharing his experience on how to prepare for job search. Thank Dinesh for always staying until late with me and discussing on great ideas and Aron for being a good company in group outings. My friends in Austin also deserve great thanks for not leaving me lonely on any days during my Ph.D. Thanks to Jong Wook Kim, who gave me a ride to school on my defense date when my car went out of battery, and also helped me many times with personal emergencies. Thank Eunho Yang for being a good friend for more than ten years, thank Yunshik Choi for great jokes and stories, and Jawook Huh for being a good exercise partner and helping me with anything, thank Jayoung Song for her company at hard times. Special thanks to Songyi Lee, for her love and patient to wait for me through the long years, and finally, I want to thank my parents Buhyun hwang and Hyunsook Park for their unconditional love and support they have shown me over the last five years, as well as to my brother Sung Min Hwang for his love and care. I especially dedicate this thesis to my mother Hyunsook Park who is currently fighting with an ovarian cancer, and hope this accomplishment can giver her a light of hope to win over the long battle that could await her. vi

7 Discriminative Object Categorization with External Semantic Knowledge Publication No. Sung Ju Hwang, Ph.D. The University of Texas at Austin, 2013 Supervisor: Kristen Grauman Visual object category recognition is one of the most challenging problems in computer vision. Even assuming that we can obtain a near-perfect instance level representation with the advances in visual input devices and low-level vision techniques, object categorization still remains as a difficult problem because it requires drawing boundaries between instances in a continuous world, where the boundaries are solely defined by human conceptualization. Object categorization is essentially a perceptual process that takes place in a humandefined semantic space. In this semantic space, the categories reside not in isolation, but in relation to others. Some categories are similar, grouped, or co-occur, and some are not. However, despite this semantic nature of object categorization, most of the today s automatic visual category recognition systems rely only on the category labels for training discriminative recognition with statistical machine vii

8 learning techniques. In many cases, this could result in the recognition model being misled into learning incorrect associations between visual features and the semantic labels, from essentially overfitting to training set biases. This limits the model s prediction power when new test instances are given. Using semantic knowledge has great potential to benefit object category recognition. First, semantic knowledge could guide the training model to learn a correct association between visual features and the categories. Second, semantics provide much richer information beyond the membership information given by the labels, in the form of inter-category and category-attribute distances, relations, and structures. Finally, the semantic knowledge scales well as the relations between categories become larger with an increasing number of categories. My goal in this thesis is to learn discriminative models for categorization that leverage semantic knowledge for object recognition, with a special focus on the semantic relationships among different categories and concepts. To this end, I explore three semantic sources, namely attributes, taxonomies, and analogies, and I show how to incorporate them into the original discriminative model as a form of structural regularization. In particular, for each form of semantic knowledge I present a feature learning approach that defines a semantic embedding to support the object categorization task. The regularization penalizes the models that deviate from the known structures according to the semantic knowledge provided. The first semantic source I explore is attributes, which are humanviii

9 describable semantic characteristics of an instance. While the existing work treated them as mid-level features which did not introduce new information, I focus on their potential as a means to better guide the learning of object categories, by enforcing the object category classifiers to share features with attribute classifiers, in a multitask feature learning framework. This approach essentially discovers the common low-dimensional features that support predictions in both semantic spaces. Then, I move on to the semantic taxonomy, which is another valuable source of semantic knowledge. The merging and splitting criteria for the categories on a taxonomy are human-defined, and I aim to exploit this implicit semantic knowledge. Specifically, I propose a tree of metrics (ToM) that learns metrics that capture granularity-specific similarities at different nodes of a given semantic taxonomy, and uses a regularizer to isolate granularity-specific disjoint features. This approach captures the intuition that the features used for the discrimination of the parent class should be different from the features used for the children classes. Such learned metrics can be used for hierarchical classification. The use of a single taxonomy can be limited in that its structure is not optimal for hierarchical classification, and there may exist no single optimal semantic taxonomy that perfectly aligns with visual distributions. Thus, I next propose a way to overcome this limitation by leveraging multiple taxonomies as semantic sources to exploit, and combine the acquired complementary information across multiple semantic views and granularities. This allows us, for ix

10 example, to synthesize semantics from both Biological, and Appearance - based taxonomies when learning the visual features. Finally, as a further exploration of more complex semantic relations different from the previous two pairwise similarity-based models, I exploit analogies, which encode the relational similarities between two related pairs of categories. Specifically, I use analogies to regularize a discriminatively learned semantic embedding space for categorization, such that the displacements between the two category embeddings in both category pairs of the analogy are enforced to be the same. Such a constraint allows for a more confusing pair of categories to benefit from a clear separation in the matched pair of categories that share the same relation. All of these methods are evaluated on challenging public datasets, and are shown to effectively improve the recognition accuracy over purely discriminative models, while also guiding the recognition to be more semantic to human perception. Further, the applications of the proposed methods are not limited to visual object categorization in computer vision, but they can be applied to any classification problems where there exists some domain knowledge about the relationships or structures between the classes. Possible applications of my methods outside the visual recognition domain include document classification in natural language processing, and gene-based animal or protein classification in computational biology. x

11 Table of Contents Acknowledgments Abstract List of Tables List of Figures v vii xv xvi Chapter 1. Introduction The need for semantic knowledge in object categorization Learning discriminative object recognition models with semantic regularization Leveraging attributes to guide feature learning Learning disjoint features on a taxonomy Combining complementary information from multiple taxonomies Transferring knowledge between related category pairs with analogies Chapter 2. Related Work Semantic knowledge in object categorization Attributes in visual recognition Taxonomies for multiclass object classification Analogies in recognition Leveraging and combining information from multiple semantic views Discriminative learning methods and regularization Multitask learning for learning the structures between tasks Metric learning for learning discriminative features Learning to combine features with multiple kernel learning 28 xi

12 2.2.4 Embedding and manifold learning for object categorization Feature selection with regularization Chapter 3. Leveraging Attributes to Guide Feature Learning Approach Basic setup and notation Learning shared features via regularization Convex optimization Extension to kernel classifiers Other extensions Results Impact of sharing features Impact of disjoint training images Selecting relevant attributes Semantically meaningful predictions Discussion Chapter 4. Learning Disjoint Features on a Taxonomy Approach Distance metric learning Sparse feature selection for metric learning Learning a tree of metrics (ToM) with disjoint visual features Results Proof of concept on synthetic dataset Visual recognition experiments Per-node accuracy and analysis of the learned representations Hierarchical multi-class classification accuracy Discussion xii

13 Chapter 5. Combining Complementary Information in Multiple Taxonomies Approach Learning a semantic kernel forest Learning class-specific kernels across taxonomies Numerical optimization Experiments Image datasets Taxonomies Baseline methods for comparison Implementation details Results Discussion Chapter 6. Transferring Knowledge between Related Category Pairs with Analogies Analogy-preserving Semantic Embedding (ASE) Encoding analogies Automatic discovery of analogies Discriminative learning of the ASE Numerical optimization Results Automatic discovery of analogies Visual recognition with ASE Completing a visual analogy Discussion Chapter 7. Future Work Unified framework for different types of semantic knowledge Learning from more complex semantic relations Exploiting first-order logical formulas A deeper semantic model Scalable approaches to object categorization xiii

14 7.3.1 Approximating the whole category space with few categories Iterative, incremental learning of the categories Chapter 8. Conclusion 147 Bibliography 151 xiv

15 List of Tables 3.1 Object prediction accuracies of Sharing+Attributes and baselines on the 50-class animals dataset (AWA), as a function of training set size Object prediction accuracies of Sharing+Attributes and baselines on the 8-class scene dataset (OSR), as a function of training set size [Object prediction accuracies for Sharing+Attributes and NSO, as a function of which image pool is used for the attribute tasks.]object prediction accuracy as a function of which image pool is used for the attribute tasks, on the 10-class AWA subset Attributes selected by ToM+Disjoint for various superclass objects in AWA Multi-class hierarchical classification accuracy and semantic similarity of ToM and baselines, on the AWA-ATTR and AWA- PCA datasets Multi-class hierarchical classification accuracy and semantic similarity of ToM and baselines, on the VEHICLE-20 datasets Attribute groups used to build each taxonomy for AWA-10 and ImageNet Multi-class classification accuracy of semantic kernel forest and baselines, on all datasets, across 5 train/test splits Multiclass classification accuracy of ASE and baselines Top-k class prediction accuracy, given an analogy with an unknown class in the form p:q=r:? Sample analogy completion results xv

16 List of Figures 1.1 Various semantic models for object categorization The overview of the thesis work Concept figure for our proposed feature sharing method between object and attribute classifiers Example images for Animals with Attributes dataset Example images for Outdoor Scene Recognition dataset Hinton diagram of the matrix Θ Accuracy on AWA and OSR classes Mutual Information experiment results Confusion matrices Example predictions by our proposed feature sharing method Graphical representations of DAP, our method, and DSLDA Concept figure for Tree of Metrics ToM experiment on Synthetic dataset Examples images for VEHICLE-20 dataset Semantic hierarchy for AWA and the per-node accuracy improvements of ToM+regularizations relative to Euclidean distance Semantic hierarchy for VEHICLE-20 and the per-node accuracy gains using ToM+regularizations Concept figure for Semantic Kernel Forests Example images for ImageNet-20 dataset Taxonomies for the AWA-10 and ImageNet-20 datasets Per-class accuracy improvements of each individual taxonomy and the semantic kernel forest over the raw feature kernel baseline Confusion matrices from semantic kernel forest Example β k s to show the characteristics of the l 1 and hierarchical regularizers for semantic kernel forest xvi

17 6.1 Concept of the analogy-preserving semantic embedding (ASE) Geometry of ASE Example analogies discovered from attributes Confusion reduction using ASE-C AWA-50 categories projected to the 2D space using each embedding method

18 Chapter 1 Introduction Humans have the natural ability to categorize objects. Objects in the physical world are grouped into a category through the process of perception and recognition. The goal of an automatic object category recognition system is to implement the same ability on a machine. Object categorization at the general level is different in nature from recognition at the instance level, for instance, from recognizing the category of concrete, homogeneous classes such as numbers or characters. In addition to the fundamental difficulties of visual recognition due to the difficulties of segmentation, variance in lighting and pose, clutter, and occlusion, there exists another, and more difficult problem of how to generalize over heterogeneous object instances. What makes us think of a chihuahua and a dalmatian as the same general object category dog? A baby or a member of an isolated tribe who has never seen either of them may have no idea that the two animals belong to the same category at the first sight. Gradually, they might learn that the two animals are similar in some sense by first observing the characteristics of each instance, and identifying the similarities between the observed characteristics of each instance, but still, the 1

19 observation of the visual similarities is not sufficient to classify them into a same category. Only after telling them that the two animals belong to the same category dog, they can associate the general object category with the commonalities that they observe. These common traits could be appearancebased such as having some specific shape of the snout, or behavior-based, such as being friendly and loyal to humans. Most current supervised learning-based automatic visual object category recognition systems work similarly, and use the category labels to learn the recognition models with statistical machine learning techniques. First, the features (characteristics) are extracted from an image, and are organized into an image descriptor that best describes the given image (object). Then, a decision function is learned to map the constructed descriptors to their category labels. The learned decision function can be later used for the category prediction of a novel test instance. Currently, discriminative learning approaches dominate the literature due to their strong empirical performance. Discriminative approaches have shown much success in object recognition for many years. Earlier methods such as logistic classifier [77], boosted classifier [39, 106], and the neural network [46], have shown to be useful in visual object recognition for specific objects such as faces [106], and characters [39]. For more challenging problem of general object category recognition, kernel methods such as support vector machine (SVM) [23] have shown much success owing to kernel trick, which allows to find non-linear classification boundaries in the original space by learning linear classifiers in a high- 2

20 dimensional feature mapping space. The state of the art recognition results on challenging datasets such as Caltech-101 and Caltech-256 [49] are obtained by some of the kernel combination methods that learn both the classifiers and the optimal combination of the kernels, such as multiple kernel learning [103], or LP-Boost [42]. Latent SVM [36], a variant of SVM that models object parts as latent variables, holds state-of-the art results in object detection. After the introduction of large-scale visual recognition datasets such as ImageNet [27], that involves the category recognition of nearly all existing general object categories, kernel methods became lackluster for their high computation and space overhead. Still, the state-of-the art results on these datasets are obtained from discriminative approaches, either by learning a low-dimensional embedding along with hierarchical classifier [11], or improving the input image descriptor by discriminatively learning mappings from each feature to codewords [117] while keeping the classifier relatively simple. However, all of these are limited in that, the only information they leverage is that the instances that have the same category label are different from the others with different labels. They view the object categories as independent, isolated entities that have no relation to others. Some recent work treats the category space as interdependent such as in structured output learning [99] and multitask learning [17], and such a structured output model have shown some success in object categorization [29]. However, important semantic information is still missing in these models. 3

21 In this thesis, I consider an important question: how can external semantic knowledge help better learn a discriminative recognition model for object categorization? 1.1 The need for semantic knowledge in object categorization The most fundamental reason why external knowledge is critical in the understanding of objects in the category level, is that the categories are semantic entities defined and perceived by humans. As the correctness of the categorization depends on the perceptual similarity of the recognition result, performing object recognition on the semantic level is a more robust way. A purely statistical model that only utilizes the class label information could be misled into learning incorrect associations between visual features and the category. For example, suppose that the model wants to recognize the category horse, but all the images available are images of a horse jumping over a fence with a person riding on it. With only image-level labels provided, the model might learn to associate visual features describing people and fences to the category horse. However, with semantic knowledge, we know that the horse is a four-legged animal, with distinct physical features of the equine, which could be all utilized to correctly associate the visual features describing horses. However, this is not the only possible advantage of using external semantic knowledge. Another advantage is that we can access much richer knowledge about the world. We humans have good knowledge about the world we 4

22 live in, and we can make use of our knowledge by associating the categories with the known concepts, unlike the traditional object recognition system that has to make decisions based only on the provided training examples. Suppose that we want the system to recognize the class hawk, but it has only seen them flying in the sky. Then, how would it recognize a hawk in a close distance? The external knowledge about the category hawk provides much information that is not present in the training set. We know that a hawk is a bird, a bird has feathers, predator birds have strong beaks, and associate the visual input to these known concepts, to recognize this animal we have never seen as a hawk. This is possible because while the categories are discrete concepts, the human semantic space they exist in is a continuous, interdependent space, where each object category does not exist in isolation, but in relation to others. Thus, an object category can be associated with other categories and semantic concepts, whether they are observed in the training set or not. Finally, relational semantic knowledge scales with the number of categories. This is the opposite situation to visual-only statistical models, for which having a larger number of categories only means more confusions. The conventional non-semantic categorization models have shown some success on small scale datasets, as each object category is visually distinct, and there is less information between the classes. (Consider a dataset consisting of four classes, car, pedestrian, monitor and keyboard). However, as the size of the dataset grows larger and the categories become more fine-grained, the categorization problem becomes more difficult as the visual space becomes more dense and 5

23 crowded, and there exists more overlap in the visual feature space between the categories. For example, categorizing different subspecies of birds [116] could be difficult as all birds have beaks and wings. Yet, this densely populated feature space is beneficial with semantic knowledge leveraged, as it means having more instances for higher-level concept learning, and being able to identify the similarities and differences more clearly. For example, suppose where we want to distinguish an otter from a beaver. They are visually very confusing and if we do not know where to focus, the classification of the two categories is difficult. Suppose, however, that we are given new categories weasel and hamster, as well as knowledge that otters and weasels are both musteline mammals, and beaver and hamster are both rodents. This gives us a critical hint on where to focus by the identified common features between the categories grouped as the same the distinct body shape of the musteline (long and sleek body) and the rodent (short body), rather than the background, pose, and many others. Further, assume that all object categories are related to each other. Then the set of all categories will form a fully connected graph adding a category will introduce the same number of linkage to the number of existing categories. The number of linkage where the relational information lies between the categories then will grow in O(C 2 ) where C is the number of categories, which can all contribute to better discrimination. To recap, the benefit of using external semantic knowledge in object recognition, as opposed to the traditional vision-only model, is threefold. 6

24 Dalmatian Siamese Cat Leopard Wolf Horse Visual Space Semantic Space (a) No Semantics Dalmatian Dalmatian Siamese Cat Spotted Siamese Cat Canine Carnivore Leopard domestic Leopard Feline Animal Wolf meatteath Wolf Horse longleg Horse Visual Space Semantic Space Visual Space Semantic Space (b) Attributes (c) Taxonomy Dalmatian Canine Biological Carnivore Dog Analogical Siamese Cat Feline Cat Leopard:cat = wolf:dog Animal Leopard Leopard Tameness Wolf Wild Animal Wolf Leopard:cat =?:Horse Horse Pet Domestic Tameness Horse Visual Space Semantic Space (d) Multiple Taxonomies Visual Space Semantic Space (e) Analogy Figure 1.1: Various semantic models for object categorization. (a) Traditional recognition model treats each object categories as isolated, independent entities that have no relation among themselves. (b)-(e), The proposed semantic models relate categories and other semantic concepts in the semantic space. 7

25 First, semantics help learn correct associations between the visual features and the category membership. Second, the external semantic knowledge enables to associate unobserved concepts that are crucial in the understanding and the characterization of the categories to the observed. Third, a semantic aware method can benefit from semantic relationships such that increasing number of categories would introduce more relationships for better learning, in contrast to the traditional model which suffers from more confusion. Overall, we can utilize the mass of knowledge about the known world as the semantic world is a continuous, interdependent space, the knowledge could be exploited from, or transferred through their relations. Traditional vision-only recognition model, on the contrary, is confined to the use of only the instances provided for training. The goal of this thesis is to explore how to exploit this external semantic knowledge, to learn discriminative models for visual object recognition in the object category level. 1.2 Learning discriminative object recognition models with semantic regularization In this section, I will give an overview of the entire thesis, while addressing what semantic knowledge to use, and how to incorporate them into the learning of discriminative categorization models. I will first start by explaining how to incorporate general semantic knowledge into a discriminative learning framework. 8

26 The approach I take in leveraging the semantics in learning is a structural regularization method [122, 61]. I introduce a regularization term that penalizes learning models that deviate from known structures defined by the given type of semantic knowledge, to augment the discriminative learning objective. This allows to leverage the power of existing discriminative learning methods while also learning semantically meaningful models that conform to human knowledge about the world; thus, we will be able to obtain a model that is discriminative yet semantic. First, let us formally define the learning problem for object categorization. Given N training instances composed of descriptor-label pairs, D = {(x i, y i )} N n=1, where x Rn is the image descriptor (or features) describing the i-th visual instance, and category labels y i {1,...,C} where C is the maximum number of categories, the learning objective for each category model j is to learn the parameter w j for the label prediction function f(x, w j ), whose optimal value can be obtained by minimizing the classification loss l(x i, y i, w j ) for each instance i defined by f(x, w j ) over all N training instances. The following shows a generic form of this categorization model learning problem. (1.1) minimize {w j } N i C l(x i, y i, w j ) j As aforementioned, this does not impose any relations between each independent categorization model w j, and thus ignores vast human knowledge to relate and group categories. The regularized discriminative learning model 9

27 I employ for imposing semantics to this model has the following problem formulation: (1.2) minimize {w j },φ N i C l(φ(x i ), y i, w j ) + λω({w j }) j The above differs from the basic categorization model learning problem in Equation 1.1 in two aspects. 1) It contains a transformation φ(x), which in most cases is learned alongside the classifier parameter w, that will transform the instances in a low-level input feature space to a higher-level common semantic space where the categories are associated to one another. 2) The categorization model learning is regularized with a semantic structural regularizer Ω({w j }) on the set of parameters {w j }, where λ balances its effect with the classification loss. The desired outcomes of this regularized learning are discriminative categorization models that minimize both the classification loss and penalty defined on prior knowledge, as well as new features φ(x) from the learned transformation φ. Due to the second aspect where the features are learned as by-products of the categorization model learning, my methods can be also viewed as feature learning methods. While the learned features are optimized for the specific categorization model learned, they could be also treated as stand-alone features, and can be used for tasks other than object categorization, such as matching or retrieval. 10

28 The key component in this model is the regularization term Ω({w j }) that provides structural constraints to the learned models and also to a learned transformation φ(x), which vary depending on the specific type of the semantic knowledge provided. Then, what kind of semantic knowledge is available for us to exploit? The semantic knowledge can come in various forms. The form could be either fixed such as groupings of the categories or arbitrary as in natural language descriptions. In this thesis, I specifically exploit the types of semantic knowledge that have fixed forms; that is, the structural constraints from the models are consistent throughout different semantic instances. I focus on semantic sources to augment the information provided with the surface category labels. The first of these semantic sources is attributes (Figure 1.1 (b)), which are semantic concepts that are shared by different object categories. They are general concepts which can span through different categories or instances, such as black, longleg, fast, or has wheels. The second semantic source is a taxonomy (Figure 1.1 (c)) which groups leaf-level classes into hierarchically inclusive groups. Further, as there exists no single taxonomy that is optimal, since the semantic relations among the categories differ for each semantic perspective, we consider semantic taxonomies in multiple semantic views (Figure 1.1,(d)). The last type of semantic knowledge visited in this thesis is an analogy (Figure 1.1 (e)), which captures high-level relational similarities between two pairs of categories with the equality constraint. Figure 1.2 shows the overview of this thesis. I allocate separate chapters for four pieces of work that have been published to major conferences [58, 55, 11

29 Category Dalmatian Visual Features Recognition model Regularization Semantic Knowledge Attributes (Chapter 3) Dalmatian Spots Longlegs Domestic Fast Taxonomy (Chapter 4, 5) Carnivore Canine Feline Visual entity Learned Features Dalmatian Wolf Siam. Cat Leopard Analogy (Chapter 6) : = : Chow Chow Dalmatian Lion Leopard Figure 1.2: The overview of the thesis work. 56, 57]. Each chapter shows how to exploit each type of semantic knowledge to regularize a specific type of discriminative categorization model for improved object categorization performance. The validation of the proposed methods categorization performance on several categorization datasets that include different types of categories such as animal [65], scenes [79], and general objects [27], show that these different types of semantic knowledge are indeed helpful in achieving better classification performance over the state-of-the art discriminative learning methods. Thus, the proposed methods can potentially be adopted to any visual recognition systems where such discriminative learning methods are used internally, to 12

30 improve upon their performance. The only requirement in using the proposed models is the provision of some domain knowledge on the set of categories. Such domain knowledge is usually inexpensive to obtain compared to perinstance labels, as it requires defining the models on the set of categories, which is not affected with the number of training instances. Also, semantic sources such as attributes and taxonomies are abundant at least for general object categories, further minimizing additional human effort. In the next subsections, I will give a brief preview of each chapter Leveraging attributes to guide feature learning The first type of semantic knowledge I exploit is semantic attributes. An attribute is a human describable property of an object, that is either visual such as spots and longleg, or semantic as domestic and fast. In the original work of [65, 34] where semantic attributes are introduced and in most of the follow-up works [63, 102, 109, 12, 86], attributes are mostly treated as midlevel features that bridge the lower-level visual features and high-level classes, and each attribute model is independently trained. However, this separation of object class (category) classifiers and attribute classifier training does not consider the fact that the object classifiers and attribute classifiers are trained on the same set of visual features, and are inherently related to each other. I instead propose to use attributes as a means to relate different object categories through learning a common, low-dimensional representation that is shared between object and attribute classifiers. 13

31 The learning of the shared features between object and attributes classifiers is achieved through group sparsity regularization. The (2,1)-norm regularizer favors shared weights, by enforcing grouping between different classifiers with the l 2 -norm regularization, and sparse feature selection with the l 1 -norm regularization within the same classifier. The resulting regularized model learns a feature space that is more semantically meaningful and achieves significant improvements over two challenging datasets of animals and outdoor scenes Learning disjoint features on a taxonomy Then, I move my attention to the second form of semantic knowledge, taxonomy. A taxonomy a human-defined hierarchical grouping of object categories, and popular examples are Wordnet [35], and the phylogenetic tree of life. Most previous work using semantic taxonomies focused either on its hierarchical structure that enables efficient classification [72, 50, 11], or on the explicit semantic information such as tree-hop distances between the classes [113, 37]. Instead, I focus on information implicitly provided from the parent-child relationships, specifically, the intuition that the features used to characterize the parent-level category should be different from the features used to characterize its children. For example, a wheel-shaped patch is useful when discriminating between a ship and a wheeled vehicle, but is not useful when discriminating between bicycle, car, and motocycle. The objective here is to focus only on the features that are useful for the discrimination of the 14

32 categories at a specific semantic granularity. To achieve this goal, we learn metrics for each node of the taxonomy, and then perform disjoint regularization between the metrics. We call this method tree of metrics (ToM). I propose a novel disjoint regularizer that requires the metrics at a node and its children to compete for features, by minimizing the l 2 -norm of the sum of the diagonals of two metrics, as it prevents the two metrics from having high value for the same feature dimension. The competition results in the isolation of the features that are discriminative for each semantic granularity. The proposed method is evaluated on two challenging datasets containing animals and vehicles. The resulting ToM model achieves better classification accuracy with k-nearest neighbor method, compared to a single metric model or flat multi-metric models. Also, the model with the proposed disjoint regularizer outperforms non-regularized models Combining complementary information from multiple taxonomies I further extend the scope of the external semantic sources to contain multiple semantic views represented by the semantic taxonomies. The motivation of the idea is that there exists no single optimal taxonomy, as the utility of the taxonomy depends on each task and view. For example, the taxonomy defined on the biological origin would group the class dog and wolf into the same superclass differentiated from the superclass containing cat and leopard, while the taxonomy defined on tameness would group the classes dog and cat 15

33 as the same. The idea is to exploit such complementary information present in these taxonomies, to learn a better (combined) semantic representation. To this end, I propose semantic kernel forests, which capture semantic similarities between instances at different views and different semantic granularities, and use multiple kernel learning (MKL) to learn the optimal combination of these feature spaces. In addition to the usual l 1 -norm regularizer for MKL to select only the useful kernels, I introduce a hierarchical regularizer based on the hinge loss, to favor upper-level metrics to select kernels that capture more high-level semantic differences. The resulting regularized MKL model outperforms the single kernel SVM, non-semantic MKL, perturbed taxonomy and single taxonomy MKL baselines, and the added hierarchical regularizer results in improved classification accuracy Transferring knowledge between related category pairs with analogies Finally, I explore a new type of semantic knowledge, analogies. While analogies have been explored to some extent in psychology and artificial intelligence [44, 45, 74, 75, 108], no prior work exploits them for categorization. Analogies provide the relational similarities between two pairs of categories. For example, in the analogy lion:tiger = horse:zebra, the common relationship would be that the latter is the striped version of the former, without the mane. I show how such a relational similarity can be interpreted into a geometrical constraint in a hypothetical category space, such that the difference between 16

34 the first pair of categories should be the same as the difference between the second pair of categories. This equality constraint will result in a more confused pair of categories benefiting from well-separated categories that share the same relationship. I encode this into a regularization term to regularize the geometry of the discriminatively learned category embedding space. The resulting analogy-preserving semantic embedding (ASE) outperforms the embedding that is discriminatively learned without any semantics or learned only with class-similarity constraints encoded as distances. ASE also outperforms others on the analogy completion task, where the task is to predict the object class that sensibly completes an analogy based on the three given classes: p:q = r:?. In the next chapter, I will describe the related work in two perspectives of how to utilize each type of semantic knowledge for object categorization, and how to augment the learning methods to incorporate the obtained semantic information. In later chapters, I will go over each method and also will describe possible future research directions in the context of semantic approaches for object categorization. 17

35 Chapter 2 Related Work My thesis work tackles two main issues. The first is what semantic knowledge to use and in what sense, and the second is how to incorporate the learned semantics in learning of a discriminative object recognition model. In this chapter, I will describe related work in these two perspectives: utilization of semantic knowledge in visual recognition, and discriminative learning methods for categorization. 2.1 Semantic knowledge in object categorization External semantics beyond object class labels are rarely used in today s object recognition systems, but recent work has begun to investigate new ways to integrate richer knowledge, such as attributes and taxonomies. My work introduced in the next three chapters focuses on exploiting these two types of semantic knowledge Attributes in visual recognition Attributes are human describable characteristics of an instance, which could be either visual or semantic [65, 34, 38]. Recent work shows that at- 18

36 tributes are useful in a variety of settings. First, they are independently useful to describe familiar and unfamiliar things (e.g., the leopard is spotted and furry, whether or not we know to call it a leopard [34, 38]), or to search through large image/video collections in semantic terms [102]. Second, they enable new zero-shot learning paradigms, where one can build an object model on the fly [65]. Third, they can serve as mid-level features to an object classification layer; having learned to predict the presence of each attribute, one can build supervised object models on top of those predictions [63, 65, 34, 110]. Usually attribute-object associations are manually specified, but some work explores ways to obtain them automatically [83, 109, 12, 86]. Notably, nearly all models using attributes for recognition learn them independently. On relating object and attributes, the indirect attribute prediction model [65] offers a way to regularize attribute predictions based on object predictions; however, the attribute-object connections are set by human-given definitions, and so the two are not jointly learned. The novel multiple instance learning (MIL) approach in [107] jointly trains attribute and object detectors with weakly labeled data, with a constraint that both models should agree on localization (e.g., if an image is tagged blue cap, both MIL classifiers should prefer to select positive training instances from the same location). In contrast, in my work (Chapter 3), I use the attributes to influence the feature space construction, not training instance selection. There is also some work that aims to use attributes to improve object classification performance. The method in [110] integrates attribute- and 19

37 object-based cues into a structured latent SVM model: the attribute labels are left as latent variables on the training data, and the objective is to minimize object prediction loss. In contrast, I show the value in discovering a single shared representation such that both attribute and object tasks can be predicted well. Thus, while [110] implicitly discovers object-attribute relationships, my work exploits the two simultaneously as explicit tasks. Doubly supervised latent Dirichlet allocation (DSLDA) [1], which is a recently proposed generative topic model that has both supervised attributes and latent shared features in the intermediate layer, is also highly relevant to my work. Such a hybrid supervised-latent intermediate layer model can benefit from both the explicit high-level semantic attributes as in [65] and learned shared latent features that account for (possibly) non-semantic high-level topics. However, DSLDA separates the latent shared feature learning from attributes, and does not infuse semantic knowledge from attributes into the shared feature learning as our model does. This limits its use as a feature learning method compared to ours, which can produce semantic, shared features as outputs Taxonomies for multiclass object classification Hierarchical taxonomies have natural appeal for object categorization, and researchers have studied ways to discover such structure automatically [95, 10, 50, 69], or to integrate known structure to train classifiers at different levels [72, 124]. The emphasis is generally on saving prediction time (by traversing the tree from its root) or combining decisions, whereas we propose to influence 20

38 feature learning based on these semantics. While semantic structure need not always translate into helping visual feature selection, the correlation between WordNet semantics and visual confusions observed in [26] supports our use of the knowledge base in this work. The machine learning community has also long explored hierarchical classification (e.g., [62, 73, 16]). Of this work, our goals most relate to [62] which focus on a very small set of features at each node of a taxonomy, during the hierarchical classification process. However, our focus is on learning features discriminatively and biasing toward a disjoint feature set via regularization. Most work in object recognition that leverages category hierarchy does so for the sake of efficient classification [72, 50, 11, 28, 41]. Making coarse to fine predictions along a tree of classifiers efficiently rules out unlikely classes at an early stage. Since taxonomies need not be ideal structures for this goal, recent work focuses on novel ways to optimize the tree structure itself [11, 28, 41], while others consider splits based on initial inter-class confusions [50]. A parallel line of work explores unsupervised discovery of hierarchies for image organization and browsing, from images alone [95, 10] or from images and tags [68]. Whereas all such work exploits tree structures to improve efficiency (whether in classification or browsing), my goal is for externally defined semantic hierarchies to enhance recognition accuracy. More related to the problem setting tackled in this thesis are techniques that exploit the inter-class relationships in a taxonomy [71, 98, 37, 26, 105]. One idea is to combine the decisions of classifiers along the semantic hierar- 21

39 chy [71, 124]. Alternatively, the semantic distance between nodes can be used to penalize misclassifications more meaningfully [26], or to share labeled exemplars between similar classes [37]. Metric learning and feature selection can also benefit from an object hierarchy, either by using a taxonomy-induced loss for structured sparsity [61], or by sharing parameters between metrics along the same path [105]. My approaches to leveraging taxonomies (Chapter 4 and 5) are different from the existing work in that I mainly focus on the exploitation of the implicit information present in the parent-child relations, and learning a granularityspecific feature space based on it Analogies in recognition Some existing work in cognitive science and AI has explored analogies in various contexts, different from my work in this thesis. Gentner et al. [44] study analogies in light of human cognition. They define an analogy as a relational similarity over two pairs of entities, and contrast it with the more superficial similarity defined by attributes. Based on this intuition, they suggest a conceptual structural mapping engine that enables analogical reasoning [45]. Recognizing that such generic analogies require high-level logical reasoning that may be problematic for an automated prediction system, Miclet et al. suggest focusing on the analogical dissimilarity between entities in the same semantic universe [74]. They exploit analogical dissimilarity to do direct logical inference when one of the entities is unknown. My work focuses on sim- 22

40 ilarly scoped analogies the semantic universe of object categories. In contrast to their logical inference model, however, I propose geometric constraints to enforce analogical proportions in a learned embedding. While my main idea is to use analogies in an embedding, I also show how to automatically discover categories that have analogical relationships using their attribute descriptions. In this respect, there is a connection to structural transfer learning work that discovers mappings between domains [75, 108]. However, while that work aims to associate distinct source and target domains (e.g., computer viruses and human viruses), we aim to detect parallel associations within the same domain, and then use those pairings to constrain feature learning. In graphics, inferring the filter relating two input images allows the automatic creation of image analogies [53]; I deal with analogies on visual data, but my idea of using them to regularize the representation is different and original. The idea of capturing higher-order relationships as vector differences in a semantic space and using a learned space to answer an analogy question in a recently published work [76] is similar to mine. However, my main objective is on improving object categorization performance rather than on predicting categories that form an analogy. Also, my method encodes the analogical relationships between category pairs explicitly into the learned semantic embedding space through regularization, while [76] does not present any means for such supervised learning for analogical relationships and soley rely on inherent 23

41 analogical relationships in the semantic space. Such an implicit unsupervised model could be less powerful even for the analogy completion task they are targeting Leveraging and combining information from multiple semantic views Combining information from multiple views of data is a well-researched topic in the machine learning, multimedia, and computer vision communities. In multi-view learning, the training data typically consists of paired examples coming from different modalities e.g., text and images, or speech and video; basic approaches include recovering the underlying shared latent space for both views [52, 68], bootstrapping classifiers formed independently per feature space [15, 21], or accounting for the view dependencies during clustering [30, 51]. When the classification tasks themselves are grouped, multi-task learning methods leverage the parallel tasks to regularize parameters learned for the individual classifiers or features (e.g., [5, 70, 58]). Broadly speaking, the problem visited in Chapter 5 has a similar spirit to such settings, since we want to leverage multiple parallel taxonomies over the data; however, the goal of aggregating portions of the taxonomies during feature learning is quite distinct. More specifically, while previous methods attempt to find a single structure to accommodate both views, our method seeks complementary information from the semantic views and assembles taskspecific discriminative features. The topic of multiple taxonomies was also 24

42 visited in [91], but their focus was on the construction of multiple taxonomies from the semantic attributes. In contrast, my focus is on exploiting predefined multiple taxonomies, where the end product is a single discriminative feature space targeted for categorization. 2.2 Discriminative learning methods and regularization From the machine learning perspective, my proposed methods can be viewed as structural regularization methods in learning discriminative models. They build on several successful existing machine learning methods namely multitask learning, metric learning, multiple kernel learning, and large margin embedding and augment the models with semantic knowledge through the means of regularization. In this section, I give a brief overview on the backgrounds of these discriminative learning and regularization methods Multitask learning for learning the structures between tasks Multitask learning refers to a class of methods that exploits the task structure among related classification tasks, to obtain better generalization ability. In the original work of [17] where multitask learning is first introduced, classifiers for different classification tasks were jointly learned by sharing the hidden units in the neural network, which are activated similarly positive for similar task outputs, and negatively for dissimilar task output. However, in general, we can refer to any method that can relate different classifiers together so that each classifier is affected by the others as multitask learning. 25

43 There are two predominant directions to pursue mutltiask learning: parameter sharing and feature sharing. Which sharing to use depends on the task. For example, for multitask learning with class classifiers and attributes, a plausible assumption is that there are invariant visual features tied to semantics, which both object classifiers and attribute classifiers use, thus rendering feature sharing as more sensible. For multiple kernel learning with taxonomies that assign weights to each node that are shared by different categories, parameter sharing would make more sense. One could differentiate different tasks as main, and auxiliary, depending on which task is the main target. For most object recognition methods, object category recognition is the main task, and different data and tasks are used as auxiliary, such as text [84, 70] or pattern matching [3]. My object-attribute feature sharing model is the first to explore multitask learning with attributes, which (relative to other sources of auxiliary tasks) has potential advantages of intrinsic task relevance and supervision reuse. Furthermore, I focus on disjoint sharing for the disjoint visual feature learning with taxonomies where the learners compete for features rather than trying to share them Metric learning for learning discriminative features Metric learning is an embedding method that learns the metric space that preserves certain distances among the training instances. It has been a subject of extensive research in recent years, in both vision and learning. Good visual metrics can be trained with boosting [92, 6], feature weight learn- 26

44 ing [40], or Mahalanobis metric learning methods [64, 59, 111]. An array of Mahalanobis metric learners has been developed in the machine learning literature [47, 25, 112]. In my work of Tree of Metrics [55] (Chapter 4), I learn a discriminative local metric at each node on a taxonomy. The idea of using multiple local metrics to cover a complex feature space is not new [114, 85, 111, 20]; however, in contrast to ToM, existing methods resort to clustering or (flat) class labels to determine the partitioning of training instances to metrics. Most methods treat the partitioning and metric learning processes separately, but some recent work integrates the grouping directly into the learning objective [6], or trains multiple metrics jointly across tasks [82]. No previous work explores mapping the semantic hierarchy to a ToM, nor couples metrics across the hierarchy levels, as we propose. To show the impact, in experiments in Chapter 4, we directly compare to a state-ofthe-art approach for learning multiple metrics. Previous metric learning work integrates feature learning and selection via a regularizer for sparsity [119], as I exploit for the ToM approach here. However, whereas prior work targets sparsity in the linear transformed space, ours targets sparsity in the original feature space, and, most importantly, also includes a disjoint sparsity regularizer. The advantage in doing so is that our learner will be able to return both discriminative and interpretable feature dimensions, as we demonstrate in our results. Transformed feature spaces while suitably flexible if only discriminative power is desired add layers that complicate interpretability, not only to models for individual classifiers but 27

45 also (more seriously) to tease apart patterns across related categories (such as parent-child) Learning to combine features with multiple kernel learning The support vector machine has shown much success in recent years in many applications, including object recognition, thanks to the kernel trick that enables learning of non-linear class boundaries by first transforming the points in the original feature space to a high-dimensional space using some function and learning a linear classifier in the resulting space [101]. While we use the term high dimensional space, most of the kernel methods actually operate in the Hilbert space that preserves similarities between training instances. This trait is also advantageous as it provides the flexibility as to how to compute the similarities. One kernel (matrix) could be computed based on similarities in the contour shape, and another kernel could be computed based on the similarities in color. Then, the problem arises on how to combine the kernels so that the combined kernel would optimally capture similarities in the category space. The simplest way is to just average them. Or, the combination weights could be learned by cross-validation. Multiple kernel learning [8], was originally proposed as the extension of the kernel-based support vector machine to solve the kernel combination problem, by simultaneously learning the classifier and the kernel combination, and it has shown much success in visual object recognition [104, 42]. The predominant direction in the research of multiple kernel learning 28

46 in machine learning has been on exploring the ways to efficiently optimize the original additive kernels. How to generate the base kernels for combination has been mostly a secondary issue. For effective combination, finding a non-linear kernel combination has shown some progress in recent years, such as product of kernels [104], polynomial kernels [22], and Hadamard product of kernels [66]. Still, how to generate the kernels remains as a domain-specific application problem. Most kernels are generated by differentiating the parameters for the radial basis function kernels, or computing on different features. The proposed semantic kernel forest (Chapter 5) also employs a form of MKL, but rather than pool kernels stemming from different low-level features or kernel hyperparameters, it pools kernels stemming from different semantic sources. Furthermore, it adds a novel regularizer that exploits the hierarchical structure from which the kernels originate Embedding and manifold learning for object categorization The analogy-preserving semantic embedding (ASE) I propose in Chapter 6 is an instance of an embedding method whose objective is to learn a representation that preserves certain topologies or properties in the original topological object. Most existing embedding methods aim to preserve the distances between data points, either globally [32] or locally [87, 115]. Label embeddings learned for object or document categorization also aim to preserve distances, but with further constraints to promote the discriminability of labeled classes [113]. Recent embedding methods preserve not only the ge- 29

47 ometry of local neighborhoods, but also higher-order properties like category clusters [94] or graph structure [93]. In my analogy-based embedding method, I also aim to preserve more far-reaching structures. However, my method is distinct in that it enforces the relative distances between semantically related pairs of instances Feature selection with regularization Identifying and using good features is critical to the robustness of a classification model, and there has been extensive work in this direction in machine learning. Regularization is a term for a general technique in statistical machine learning to introduce additional constraints, or in other words, penalty terms, in the learning model to avoid overfitting to the bias in the training sample [97, 90]. A popular regularization method for learning classification or regression model is a sparsity-inducing norm regularization, that enables to select features. Lasso [97] uses l 1 -norm penalty term to favor sparse solutions for the training of classifiers or regressors. This enables to select features that are more useful and suppress noisy terms, resulting in a robust classifier that better generalizes. Ridge regression [90] regularizes the coefficient of the model using l 2 -norm, suppressing the coefficient from growing to infinite. It cannot zero-out the parameters to mathematical zeros as lasso does, but can correlate feature dimensions by shrinking correlated features simultaneously. The elastic net [123] uses a convex combination of both l 1 - and l 2-30

48 penalties, resulting in sparse solutions while also shrinking correlated factors at the same time. This is called group sparsity, and further explored in the mixed-norm regularization. A group lasso performs l 2 -regularization along the feature dimension, and performs l 1 -regularization of these l 2 -regularized groups. This results in group sparsity, which makes correlated features drop out together. In my multtiask learning method with semantic attributes, we use this (2, 1)-norm as the objective (while solving the alternative problem that is convex). Most group-sparsity regularization works by promoting sharing among the different learners. However, in some scenarios, making each learner to compete instead of share could be beneficial. Exclusive lasso [122], aims to minimize the l 2 -norm of each dimension of the classifiers that are l 1 -regularized (lasso), making each classifier to compete for a feature dimension. The disjoint regularizer used for the tree of metrics shares the same spirit, but promotes competition between two metrics instead of two classifiers. Taxonomy-based regularization also has gained some limited attention recently. Tree-guided group lasso [61] uses the l 2 -norm to identify shared parts, and l 1 -norm regularization to obtain sparse selection of its children. Orthogonal transfer [121] leverages the intuition that classification among subcategories should not consider the factors that are already considered at upper levels, by constraining the parent and children classifiers to be orthogonal to each other. ToM is based on the same intuition but targets metric learning, and enables true selection of features using sparsity regularization and a dis- 31

49 joint regularizer that minimizes the l 2 -norm of the diagonal. The proposed semantic kernel forest also introduces a structured regularization is based on the intuition that higher level classification should be considered as more important (as it is tied to more number of lower-level classification problems), which is implemented into a hinge-loss regularizer. The main novelty of my work in the machine learning, is in showing how to translate the abstract external domain knowledge into concrete structural constraints between classifiers, that sum up to regularizers to augment the discriminative learning objective, to learn discriminative yet semantic models (and features). This process is domain-agnostic as the requirement is only on the structures of the knowledge. Thus not just visual recognition models, but any classification models where such specific type of domain knowledge are available, can benefit from my method; the augmented model will enable leveraging the power of the existing discriminative classification learning algorithms while also utilizing the vast and complex domain knowledge that will guide the learning into a more correct direction. They will also less overfit to the training set biases compared to purely statistical approaches that rely only on the labels, which will result in improved accuracy from better generalization. 32

50 Chapter 3 Leveraging Attributes to Guide Feature Learning The first semantic source I explore is semantic attributes. Attributes are human-understandable properties shared among object categories (e.g., glassy, has legs), and they are a compelling way to introduce high-level semantic knowledge into predictive models. As discussed in the previous chapter, recent work shows that attributes are valuable in several interesting scenarios, ranging from description of generic images or unfamiliar objects [38, 34, 102], to zeroshot transfer learning [65], to intermediate features that aid in distinguishing people, objects, and scenes [63, 65, 34, 110]. Existing approaches to attribute-based recognition assume that the attributes role is primarily to focus learning effort on properties that will be reusable for many categories of interest, and to elegantly integrate human knowledge into discriminative models. As such, attribute classifiers are learned independently from object classifiers, and then their predictions are treated as mid-level features that bridge low-level image features and high-level object classes. However, segregating supervision about attributes from supervision about objects may restrict their impact. In particular, in conventional mod- 33

51 polar bear dalmatian white spots Object class classifier ym1 ymm ya1 yaa Attributes classifier D(M+A) Shared features!! u1 u2 u3 ud!! x1 x2 x3 Input visual feature xd Figure 3.1: In my object-attribute feature sharing model, object categories and their human-defined visual attributes share a lower-dimensional representation (dashed lines indicate zero-valued connections), thereby allowing the attribute-level supervision to regularize the learned object models. els, even though attributes influence object predictions, the attribute-labeled training data does not directly introduce new information when discriminatively learning the objects. I explore how learning visual attributes in concert with object categories can strengthen recognition. The assumption is that both types of prediction tasks rely on some shared structure in the original image descriptor space. In other words, patterns among those generic visual properties that humans elect to name may reveal information about which low-level cues are valuable to object recognition in the most general case, whether the objects of interest exhibit those attributes or not. Thus, rather than treat attributes as intermediate features, I propose an approach to discover this structure and learn a shared lower-dimensional representation amenable to discriminative 34

52 models for either one (see Figure 6.1) 1. In effect, I show how human-defined semantics (as revealed by attributes) can regularize training for object classifiers. Given a low-level visual feature space together with attribute- and object-labeled image data, my method learns a feature subspace for all labeling tasks based on a joint loss function that favors common sparsity. The optimization process alternates between regularizing towards shared features, and retraining task-specific classifiers based on those features. Our technique directly builds on a multi-task feature learning algorithm developed in [2], where it was applied to collaborative filtering of consumer data. To improve its scalability, we provide a more efficient kernelized implementation and linear algebra shortcuts for dealing with large matrices. Additionally, while in [2] all tasks are assumed to have the same label space, our setting entails non-overlapping label spaces (attributes, objects), for which feature-sharing is expected to be more challenging. It is well-known that the success of multi-task learning or feature sharing hinges on the assumption that the input tasks are indeed related. Why should the assumption hold in our case? What makes attributes special as auxiliary tasks for object learning? Intuitively, their relation is intrinsic, since attributes are by definition shared among object categories. Many object-level distinctions can be made using a vocabulary of relevant properties, suggest- 1 The work introduced in this chapter was published in [58]. 35

53 ing that a representation sufficient to distinguish the properties would also be relevant for the objects (e.g., a child learning to discriminate cows from other animals might focus on the visual properties a cow exclusively has but other animals do not). In fact, in early visual processing, it is known that the human visual system discovers some sparse coding using a feature vocabulary of low-level filters [80]. More abstractly, we expect that structure among a wide span of attribute classifiers could reveal information about which low-level features are valuable to human understanding of the visual world. That is, even attributes that are not relevant to distinguishing a particular object may still help to constrain the space of image descriptors suitable for higher-level recognition problems. Finally, there is a practical incentive for treating attributes as auxiliary tasks regarding supervision cost: for many attributes, knowing the real world object-attribute relationship is sufficient to transfer object-level image labels to attribute-level labels (i.e., all buildings are manmade, so if we have a labeled image of a building, it is also an image of the manmade attribute). 2 In short, the contribution of this chapter is threefold: 1) design of a method for feature sharing between object and attribute prediction tasks; 2) validation of the method s effectiveness with experiments on two datasets that feature sharing can offer noted improvements in accuracy for target object categorization tasks; and 3) exploration to what extent different attributes are 2 This is the case for many binary attributes, but of course not all attributes (e.g., some bicycles are red, some are blue). 36

54 useful for a target task, and provide some initial ideas for automatic selection of relevant attributes to limit training costs. 3.1 Approach I describe in detail the approach we take to learn shared features between objects and their attributes. My work directly builds on a previous approach [2]. Being mindful of desired large-scale learning settings, however, we extend the method by providing faster and more scalable numerical techniques. Additionally, we adapt the models to handle classification tasks where the label sets are disparate. I start by describing the basic setup for learning features from multiple tasks, and then explain how the problem can be cast as convex optimization for both linear and kernel classifiers. Finally, I discuss extensions and improvements I have developed in order to apply the approach Basic setup and notation There are two groups of classification tasks. We aim to improve object classification accuracy; thus, we refer to the objects as the main task, and the attribute classifiers as auxiliary tasks. Note that the two groups have different sets of labels. We use multi-class support vector machines (SVMs) for the main task [24]. Let M denote the number of object classes, x n R D denote the n-th feature vector in the training data and y n its class label. The multi-class SVM has M 37

55 parameter vectors {wm }M m=1, one for each class. In the most basic setting, we consider linear discriminants which are parameterized by w m R D. Let W denote the matrix whose columns are w m. To identify W, we minimize a loss function that maximizes the discriminant w T y n x n, W = arg min n l({w T m x n} M m=1, y n) + γ m w m 2 2 where γ 0 is a tradeoff parameter that regularizes the model complexity, using the parameter s 2-norm. For learning A auxiliary tasks, we use y na to denote the label for the a-th auxiliary task and w a for the corresponding model parameter. Our auxiliary tasks are binary classification of attributes. We use the squared hinge loss for these tasks. For simplicity, the notation assumes that both the main task and auxiliary tasks are trained on the same feature vectors. However, this is not mandatory, as we demonstrate in our results. We use t ranging from 1 to T = (M + A) to index all parameter vectors for the main and auxiliary tasks. To avoid unnecessary notation clutter, with a slight abuse, we use M t=1 l(wt t x n, y nt ) in lieu of l({w T m x n} M m=1, y n), namely, the true object function for the main task Learning shared features via regularization Conventionally, all T parameters {w m } T t=1 are learned by independently training (1+A) classifiers. For linear discriminants such as wmx T n, the resulting parameter often reveals how effective features are. For instance, a zero-valued 38

56 element w mi indicates that the i-th feature of x n does not play a role in classifying objects. Thus, intuitively, for related tasks, we expect their parameters to reveal similar sparsity patterns. Furthermore, we hypothesize that shared patterns will enable more effective parameter training for example, reducing feature space dimensionality, thus improving classification performance. How can we identify such common patterns across tasks? This desideratum is achieved in two steps. The first is to transform the original features to a shared feature space U T x n U for all tasks [2, 4]. The second step is to learn models in the space of U and promote a common sparsity pattern in the new parameters. Concretely, we express the discriminant in {θ t } such that w t = Uθ t. Analogously to W, we collect all θ t in Θ R D T. We jointly optimize all loss functions, but regularized with Θ s (2, 1)-norm, (3.1) Θ, U = arg min t The norm is given by Θ 2,1 = D d=1 l(θt T UT x n, y nt ) + γ Θ 2 2,1 n t θ2 td. An important property of this norm is that it computes the 2-norm of parameter values in each dimension across tasks. Consequently, for any dimension d, the regularization attains the minimum if and only if the corresponding parameters are all zero: θ td = 0 for all t. Therefore, the regularization would choose the Θ with the smallest number of non-zero rows. The discriminant θ T t U T x n depends only on nonzero elements of θ t. Thus equation 3.1 yields solutions that use a subset of features that are commonly effective for all tasks. Similar ideas have also been explored in other 39

57 settings [120, 78]. The optimization of Equation (3.1) is challenging due to the nonsmoothness of the regularization term. We next describe the alternating minimization algorithm proposed in [2] Convex optimization The optimization algorithm of [2] starts by identifying equation 3.1 with its equivalent form W, Ω = arg min l(wt T x n, y nt ) t n (3.2) + γ wt T Ω 1 w t + γǫ Trace(Ω 1 ), t where Ω R D D is constrained to be a positive definite matrix with bounded trace Trace (Ω) = 1. ǫ 1 is a smoothing parameter for numerical stability and benign convergence properties (cf. Theorem 3 in [2]). Ω s role can be understood more clearly by relating the solutions to the two problems in equation 3.1 and equation 3.2: { (3.3) W = U Θ, Ω = U Diag Θ d 2 Θ 2,1 } D d=1 U T where the operator Diag( ) converts its D-element arguments as elements of a diagonal matrix. Θ d 2 is the 2-norm of Θ s d-th row: t θ2 td. Intuitively, the diagonal measures relatively how much each row of Θ is non-zero. Therefore, the matrix Ω measures relative effectiveness of each feature dimension. 40

58 Further insight could be gained by drawing an analogy to the maximum a posteriori (MAP) estimator when the prior distribution for the parameter w t is a Gaussian N(w t 0;Σ 1 ). The regularization term of the MAP estimator is in the form wt TΣ 1 w t. Therefore, intuitively, Ω functions as an estimator of the covariance structure, computed from all parameters w t (or equivalently, θ t ), over all tasks. Equation 3.2 is computationally advantageous for it is a convex optimization. To solve it, we alternatively minimize over {w t } and Ω while holding the other fixed. When Ω is fixed, each w t can be identified as (3.4) w t = arg min n l(w T t x n, y nt ) + γw T t Ω 1 w t. With two simple variable substitutions, the optimization takes the standard form of l 2 -norm regularization: (3.5) ŵ t = arg min n l(ŵ T t z n, y nt ) + γ ŵ t 2 2, (3.6) z n Ω 1/2 x n, ŵ t Ω 1/2 w t. When the parameters {w} are fixed, the optimal Ω that minimizes equation 3.2 has a closed-form solution: (3.7) Ω = (WW T + ǫi) 1/2 Trace[(WW T + ǫi) 1/2 ]. The alternating minimization procedure monotonically decreases the objective function until the optimum solution is reached. Algorithm 1 lists the key steps. We set the hyperparameters γ and ǫ using a validation data set. 41

59 Algorithm 1 Learning Shared Features for Linear Classifier [2] Require: training data (x n, {y nt }),ǫ,γ Ensure: W, Ω 1: Initialize Ω with a scaled identity matrix 1 D I 2: while W still changes between two iterations do 3: Compute transformed variables according to Equation (3.6) 4: Solve ŵ t according to Equation (3.5) 5: Compute w t as w t = Ω 1/2 ŵ t 6: Update Ω according to Equation (3.7) 7: end while Extension to kernel classifiers The feature learning framework can be extended to kernel-based nonlinear classifiers. We apply the kernel construction of [2]. Let K(x n, x n ) denote the kernel function between two original feature vectors x n and x n. The kernel induces a nonlinear feature mapping φ(x n ) H R H. We perform feature learning in this new space H. To kernelize, note that the optimal parameter W R H T for the models is a linear combination of (training) feature vectors. This can be understood intuitively by observing that Equation (3.5) is the standard formulation of an SVM; therefore the solution {ŵt } is a linear combination of feature vectors. The same statement is also true for W, as the two are linearly related as in Equation (3.6). It is computationally convenient to express W using the basis V of the feature space H: W = V α (we have adopted a slightly different notation from [2] by adhering to the standard nomenclature in SVMs). We assume the number of basis vectors in V is B < N where N is the total number of 42

60 feature vectors. The matrix α is the linear combination matrix, each column for a task. The basis V can be computed from the kernel matrix formed from training feature vectors, for instance, through eigendecomposition or Gram- Schmidt (G-S) orthogonalization. We use the latter technique for its slightly lower computational overhead. Concretely, we randomly choose B training feature vectors S and express the basis in the linear combination of those features, V = Φ S B, where the matrix Φ S s columns are the nonlinear features computed from the chosen training instances. The matrix B R B B stores the linear combination coefficients, computed by the G-S process. The parameter W is also linearly represented, as W = Φ S Bα. Analogous to Equation (3.2), the optimal α is then: α, Ω = arg min l(α T t z n, y nt ) t n (3.8) + γ α T t Ω 1 α t + γǫ Trace(Ω 1 ). t where α t is the t-th column of α. z n = B T k S (x n ) is the transformed data, resulting from the linear discriminant in the feature space H, (3.9) w T t φ(x n) = (Bα t ) T Φ T S φ(x n) = α T t BT k S (x n ), where the vector k S (x n ) R B consists of the elements of the kernel function k(x n, x b ) = φ(x b ) T φ(x n ). The optimization problem Equation (3.8) is now readily solvable using techniques described previously. Key steps are given in Algorithm 2. 43

61 Algorithm 2 Learning Features for a Kernel Classifier Require: training data (x n, {y nt }),ǫ,γ, and B Ensure: α, Ω, B 1: Formulate kernel matrix K 2: Compute the basis B, S Gram-Schmidt(K, B) 3: Transform data according to Equation (3.9) and S 4: α,ω Algorithm 1((z n, {y nt }),ǫ,γ) Other extensions I propose several additional extensions, addressing issues that naturally arise in our setting. Modeling disparate sets of labels As opposed to [2], the main task and auxiliary tasks here have different sets of labels and different types of loss functions. Thus, we use two regularizers, one for each group. In the linear classifier case, our optimization takes the form, W, Ω = arg min l(wt T x n, y nt ) + ǫ Trace(Ω 1 ) t n (3.10) M T + γ M wt T Ω 1 w t + γ A wt T Ω 1 w t t=1 t=m+1 where γ M is used for the main task and γ A for auxiliary tasks. When γ A is set to zero, the optimization learns shared features from parameters for all object classes, without attributes. We term this setup as Sharing-Obj. When γ M is constrained to be the same as γ A, we recover equation 3.2. Handling high-dimensional features The alternating minimization algorithm described in Section depends on re-estimating Ω and computing its 44

62 square root Ω 1/2 with equation 3.3 and equation 3.6. For the high-dimensional features used in our setting, directly computing these quantities is costly. We exploit the low-rank property of Ω to circumvent this challenge. Note that the matrix W has T columns and D T rows. Thus, W can be factorized with thin singular value decomposition: W = LSR T, where L R D T and R R T T are W s (partial) left and right eigenvectors. The diagonal matrix S R T T is composed of W s singular values {σ i (W)} T i=1. With some algebraic manipulation, we identify the eigenvalues of Ω: (3.11) λ i (W) = ( σ 2i (W) + ǫ ) /ρ, λ(ǫ) = ǫ/ρ (3.12) ρ = T σ2 (W) + ǫ + ǫ [D T]. i=1 The eigenvectors in L and the subspace orthogonal to them span precisely Ω s column space. This yields, (3.13) Ω = L Diag ( {λ i (W)} T i=1) L T + λ(ǫ)(i LL T ). The matrix Ω 1/2 can be formulated similarly, replacing λ i (W) and λ(ǫ) with their square roots. Choosing the kernel basis For the kernelized version, one needs to choose B basis vectors to expand the kernel feature space, as described in Section We use two simple heuristics. We choose B large enough such that the performance of using the B basis vectors for individual task learning is close to the performance of our baseline system s. The individual task learning is set up as 45

63 antelope beaver cow dalmatian elephant german sheperd horse lion persian cat polar bear rhinoceros zebra white spots Figure 3.2: Examples images for AwA (Animals with Attributes) dataset. Top two rows: Object categories. Bottom row: Attributes. a linear classifier using the transformed feature vectors Equation (3.9), while the baseline system s are kernel-based nonlinear classifiers using the original features. For the Gram-Schmidt process, we choose B/M feature vectors randomly from each of M classes. This gives balanced coverage of different features, and in practice works better than purely randomly selecting without taking object class into consideration. 3.2 Results I validate my approach against relevant baselines, and report results on object categorization, the main target task. 46

64 street tallbuilding mountain insidecity forest coast opencountry highway natural open diagonal-plane Figure 3.3: Examples images for OSR (Outdoor Scene Recognition) dataset. Top two rows: Object categories. Bottom row: Attributes. Datasets We consider two datasets: the Animals with Attributes dataset (AWA) [65], and the Outdoor Scene Recognition dataset (OSR) [79]. AWA contains 30,475 images, 50 animal classes, and 85 attributes. 3 Each image is labeled by the animal and attributes present. OSR has 2,688 images, 8 scene classes, and 6 attributes as given in [79]: natural, open, perspective, size, diagonal plane, and depth. See Figures 3.2 and 3.3. We asked another vision researcher to make the assignment from attributes to scenes. We apply random train-test splits, ensuring balance among object classes. Throughout, we use object to refer to an animal or scene. Baselines We consider two baselines: a traditional multi-class object recognition approach using an SVM with 3 For all methods, we use the 59 attributes exceeding 70% accuracy as reported in [65], since some are unpredictable from the given features. 47

65 a χ 2 kernel computed on image features, which we refer to as No sharing- Object, or NSO. an approach that treats attributes as intermediate features, which we call No sharing-attribute, or NSA. For NSA, we train SVMs on image features to predict attribute labels, and then treat their outputs as features to a multi-class logistic regression classifier. This baseline follows the basic direct attribute prediction (DAP) approach defined in [65]. We use LIBSVM. Image features All methods use the same original image features. For AWA, we use the six (SIFT, rgsift, PHOG, SURF, LSS, RGB) provided with the dataset, each up to 2688-D. For OSR we generate 512-D Gist and 45-D LAB color histograms. We average the kernels computed over multiple feature types. Note that both datasets permit global descriptors, since there is one primary object of interest per image. To test with multi-object images, one would apply a window-based detector Impact of sharing features First we evaluate the object recognition accuracy of our approach and the baselines. Our approach gets the same training images for both the attribute and object tasks. We form four training splits of increasing size (10% to 60%), and reserve the rest for validation and testing (20% each). We demonstrate two variants of our approach: Sharing-Obj, where we learn a common 48

66 50-class Animals Dataset Method / % train data 10% 20% 40% 60% No sharing-obj. (NSO) No sharing-attr. (NSA) Sharing-Obj. (Ours) Sharing+Attr. (Ours) % gain over NSO 14.92% 11.75% 8.21% 6.06% % gain over NSA 18.37% 19.63% 16.00% 16.86% Table 3.1: Accuracy on the 50-class animals dataset (AWA), as a function of training set size. Learning shared representations with our approach significantly improves generalization on the novel test set, and can be most pronounced when labeled training data is limited. representation for all object classes simultaneously, corresponding to γ A = 0 in Equation (3.10), and Sharing+Attributes, where we learn the space for all objects and attributes, corresponding to γ A = γ M. Table 3.1 and Table 3.2 shows the results. Our feature sharing approach offers significant improvements over both No sharing baselines, and we obtain the best results when jointly learning with both the objects and attributes. The last two rows summarize gains of Sharing+Attributes over the baselines. Our improvements over the NSO baseline are perhaps most informative, since the general approach taken by NSO (multiple image features, kernel combination, nonlinear SVM) is typical in state-of-the-art image recognition techniques. While the margin between our Sharing-Object and Sharing+Attributes variants is smaller than the margin between not sharing at all versus sharing, the impact of attributes is clear and consistent. A one-tailed paired t-test on the 60% training split confirms that the accuracy gain with attribute tasks 49

67 8-class Scene Dataset Method / % train data 10% 20% 40% 60% No sharing-obj. (NSO) No sharing-attr. (NSA) Sharing-Obj. (Ours) Sharing+Attr. (Ours) % gain over NSO 1.73% 2.34% 3.44% 3.90% % gain over NSA 35.17% 38.39% 41.97% 43.16% Table 3.2: Object prediction accuracies of Sharing+Attributes and baselines on the 8-class scene dataset (OSR), as a function of training set size. %&"!'(&# %&"!'(&#!"#$#!"#$# Figure 3.4: Hinton diagram of the matrix Θ in the initial and last iterations of Alg. 2. Each square is a matrix entry, and area reflects the entry s magnitude. For clarity only a partial matrix is shown, for the first 30 features (horizontally) and the first 10 object classes (vertically). The matrix at the last iteration is much sparser. is statistically significant (for α = 5% on AWA and α = 1% on OSR). By separately tuning the γ M and γ A regularization weights, we expect even better performance; we simply let them be equal to save computation time. Interestingly, on the larger AWA set, the gain using our method are largest for smaller labeled data pools, supporting our claim that attribute feature sharing can have a beneficial regularization effect for object learning. This is an encouraging result, particularly since obtaining attribute labels on object-labeled data has minimal additional overhead for many attribute 50

68 Classification Accuracy No sharing Object No sharing Attributes Sharing+Attributes 0 dalmtn raccoon skunk collie leopard gorilla giraffe plrbear rabbit zebra ox panda fox rhino grizbear h.whale hippo elephant buffalo bobcat lion beaver weasel sheep tiger walrus mouse per.cat rat mnkey siam.cat ki.whale chimp mole chihua deer hamster otter moose dolphin pig horse g.shprd wolf squirrel bat antelope bl.whale cow seal Classification Accuracy street tallbuilding mountain insidecity forest No sharing Object No sharing Attributes Sharing+Attributes coast opencountry highway Figure 3.5: Accuracy on AWA (top) and OSR (bottom) classes. Our approach outperforms methods that learn objects (No sharing-object) or attributes (No sharing- Attributes) independently. types, as discussed previously. Figure 3.4 visualizes the shared features over iterations, showing how we converge to a common sparse set. Figure 3.5 breaks out the prediction accuracy per object category on both datasets. We improve accuracy for 33 of the 50 AWA classes, and yield correct predictions for some classes the baselines miss completely (e.g., beaver, rat). On OSR, the absolute accuracy is higher overall, due to the smaller multi-way decision. However, NSA suffers due to the insufficiency of the attribute vocabulary; it happens that the scenes tallbuilding and insidecity have exactly the same attribute definitions. In contrast, our approach accounts 51

69 for attributes while still learning features sufficient to make the distinction. One might ask whether some arbitrary grouping of object classes into tasks might also have similar benefits. That is, are our gains due to the attributes meaning, or could it be a sort of error-correcting code effect? To analyze this, we test a baseline where each object s attribute labels are randomly reassigned to other attributes, and then apply our method (for five such random assignments on the 60% training split). On OSR, we find this baseline offers no improvement over Sharing-Object (decreasing accuracy by 0.06). On AWA, the baseline improves over Sharing-Object (by 0.97 on average), but by less than sharing with real attributes (which increases accuracy by 1.79). This indicates the attribute semantics are indeed a factor in our method s success. 4 In the remaining text, I report the results using Sharing+Attributes, and focus on the AWA data, since it is 11 larger and has a richer set of attributes Impact of disjoint training images Our model is flexible to the source of object- and attribute-labeled data, and we can train the tasks on disjoint sets of images. This is relevant when one has a large set of existing attribute-labeled data, and wants to use it to regularize the training process for a new set of object models. 4 Looking closely at the AWA data, we see that the baseline s small gain made with randomly assigned attribute labels may be misleading. Because the classes are fine-grained, any random assignment of labels can overlap with meaningful attributes; the 85 attribute labels in AWA are certainly not exhaustive for the 50 animals. 52

70 Image source for attributes Method Same Disjoint All No sharing-object (NSO) Sharing+Attribute % gain 4.67% 4.56% 5.56% Table 3.3: [Object prediction accuracies for Sharing+Attributes and NSO, as a function of which image pool is used for the attribute tasks.]object prediction accuracy as a function of which image pool is used for the attribute tasks, on the 10-class AWA subset. Thus, we next examine the impact of which images are used as the auxiliary attribute tasks to train the object classifiers. We select 10 classes (the same as [65]) to train the object classifiers, and test three variations for learning the attributes: 1) the same images used for the objects, 2) a disjoint set of images containing object classes outside of the 10, and 3) all images, the union of the previous two. Table 3.3 shows the results. Interestingly, I see that our method performs similarly whether the attribute data overlaps or not (see first two columns). This suggests that the value of the attributes is not simply having deeper/stronger labels on the very same training examples; rather, it is the fact that we identify a common space where both types of labels are well predicted. The table also indicates that more attribute-labeled images is helpful (cf. last column) Selecting relevant attributes Having tested the impact of which images have attribute labels, next we consider the impact of which attribute classes are leveraged as auxiliary 53

71 tasks. Presumably, not all attributes will benefit feature sharing, and as usual in multi-task learning some may be detrimental. Even if all attributes were relevant to some degree, we may want to be selective to save training costs. Thus, I explore a simple form of automatic attribute selection in which we rank all attributes by their mutual information (MI) with the 10 animals 5. Figure 3.6 (left) displays the computed MI, from the most informative attributes (e.g., spots, which chimps and pigs lack, but leopards and pandas have) to the least (e.g., none of the 10 animals fly ). Figure 3.6 (right) shows the impact of using the MI scores to select attributes for sharing. Both dotted curves denote our method, but one uses the k most informative attributes, and the other uses the k least informative attributes. 6 The most interesting cases are for lower values of k. (For higher values of k, the most and least sets overlap more, and they are identical at k = 85.) The results show that using the 20 attributes with the highest MI yields the best accuracy, while using the lowest 20 is slightly worse than using none whatsoever. Further, we see that more attribute classes do not necessarily always help. These findings plus the fact that training time increases linearly with k (see solid green line, right axis), suggest it is practical to choose in- 5 chimp, panda, leopard, persian cat, hippo, whale, raccoon, rat, seal 6 Note, we simply fix the γ and ǫ parameters for all cases, in order to see the effect of the attribute selection in isolation. 54

72 spots meatteeth agility furry fast lean fish coastal newworld stalker tunnels longneck horns flys oldworld Mutual information Accuracy (%) Selecting relevant attributes Lowest mutual information 400 Highest mutual information k Training time (seconds) Figure 3.6: Left: Mutual information scores. Right: Object classification accuracy and training time as a function of the number of attribute tasks included. chimp panda leoprd cat pig hippo whale raccn rat seal chimp panda leoprd cat pig hippo whale raccn rat seal chimp panda leoprd cat pig hippo whale raccn rat seal chim mp pand da leoprd cat pig po c ale cn p hipp wha racc rat eal r se (a) True semantic ties chimp panda leoprd cat pig hippo whale raccn rat seal (b) NSO confusions chimp panda leoprd cat pig hippo whale raccn rat seal (c) Our confusions Figure 3.7: Confusions made by the baseline (b) and our method (c) relative to human-given object relationships (a). telligently. This result also shows the potential for performing task selection outside of the feature sharing learning procedure Semantically meaningful predictions Finally, we analyze to what extent the semantics we introduce by jointly training objects and attributes are manifest in our method s predictions. Figure 3.7 compares the confusion matrices for our method (c) and NSO (b). To 55

73 judge the reasonableness of their errors, in (a) we depict the true relationships between all pairs of the 10 objects. To obtain this matrix, we use human subjects ratings collected in [81] about the relative strength of association between the 85 attributes and 50 objects in AWA. For each object, we create a vector of its 85 property strengths, and then compute the pairwise χ 2 kernel values between all such vectors. Brighter boxes indicate greater true association in (a), and higher confusion in (b,c). Thus, if a method captures semantics well, its confusion matrix will look more like (a). First, we notice that our method boosts accuracy for most classes, raising the mean diagonal from 66.9% to 68.9%. Second, we see that the pairs for which our method most reduces confusions (e.g., pig vs. rat) are more distinctive semantically. On the flip side, some closely related pairs become confused by our method (e.g., raccoon vs. cat). Figure 3.8 shows example animal category and attribute predictions, compared alongside NSO and NSA. 3.3 Discussion In this chapter, I showed that by learning a common feature space suitable to either attribute or object tasks, the classifiers can obtain noticeably stronger object recognition performance. I demonstrated the proposed method s improved generalization accuracy and its potential to make more predictable errors in terms of human-defined semantics. 56

74 NSO polar+bear NSA strong big walks oldworld fast solitary meatteeth NSA polar+bear Ours strong big oldworld walks fast ground solitary Ours dalmatian NSO dolphin NSA fast active toughskin chewteeth forest ocean swims NSA walrus Ours fast active toughskin fish forest meatteeth strong Ours grizzly+bear (a) Dalmatian (b) Grizzly Bear NSO otter NSA solitary quadrapedal fast paws active claws small NSA otter Ours fast quadrapedal solitary ground active gray tail Ours hippopotamus NSO grizzly+bear NSA strong inactive vegetation quadrapedal slow walks big NSA rhinoceros Ours strong toughskin slow walks vegetation quadrapedal inactive Ours moose (c) Hippopotamus (d) Moose NSO giant+panda NSA quadrapedal oldworld walks ground furry gray chewteeth NSA rabbit Ours quadrapedal oldworld ground walks tail gray furry Ours rhinoceros NSO cow NSA oldworld quadrapedal walks ground chewteeth furry forest NSA deer Ours oldworld quadrapedal walks ground chewteeth furry forest Ours wolf (e) Elephant (f) Fox Figure 3.8: Example predictions by our method (right column in each), No Sharing- Attributes (NSA, middle columns), and No Sharing-Objects (NSO, object prediction under each image). Attributes are those with 7 highest positive decision values, by ours or NSA (red attributes incorrect). (a)-(d) illustrate good results, and (e)- (f) show failure cases that highlight our method s tendency to make semantically meaningful errors. The enforced sharing via mixed norm regularization results in discarding features that are only specific to each category and keeping the ones that are shared with attributes, which adds more semantics to the learned feature space. This semantic guidance not only makes the category recognition model more robust, but also leads to more semantically meaningful predictions. The 57

75 (a) DAP (b) Ours (c) DSLDA Figure 3.9: Conceptual graphical representations of direct attribute prediction (DAP) [65], our feature sharing method, and doubly-supervised latent Dirichlet allocation (DSLDA) [1]. Dark gray nodes denote observed nodes, light gray nodes denote nodes observed only during training and inferred in test, and white nodes denote latent nodes that are never observed. Further, M is the number of object classes, A is the number of attributes, and K is the dimensionality of the shared latent features. introduction of attributes here could be viewed as introducing a layer of flat higher-level semantic concepts that groups the categories as either having or not having the desired semantic property. While our method shows impressive results outperforming state-of-the art methods, there still remains further room for improvement. In our model, we treated the attributes as additional supervision to class labels in the output layer, and the features associated with each attribute were indirectly learned through the feature sharing. However, this indirect attribute-guided latent shared feature learning does not guarantee that the features learned on the latent space directly correspond to each attribute, especially when the attribute describes high-level semantic properties such as fast or domestic. Consequently, our model might result in learning less semantic features compared to explicit attributes modeling as in [65], while achieving more discrimination 58

76 power (Figure 3.9 (a), (b)). Obtaining better discrimination power with a possible sacrifice of semantics is perfectly fine for the object categorization task we are aiming at, but might be less optimal if the objective is to learn strictly semantic models (or features). Doubly-supervised latent Dirichlet allocation (DSLDA) [1], a recently proposed hybrid supervised-latent topic model, suggests a way to take advantage of both explicit attribute modeling and latent shared feature learning. DSLDA has both supervised attributes and latent shared features in the intermediate layer, where the former accounts for attributes while the latter accounts for high-level shared topics not included in the set of attributes (Figure 3.9 (c)). Still, DSLDA has its limitation that it cannot benefit from additional supervision from attributes when learning the shared latent features, as our method does, due to the separate training of attributes and latent features. The limitation common to all these models is that they only have a single intermediate layer to represent attributes, while the attributes come in diverse semantic granularities. Attributes such as longleg and lean can be directly inferred from visual features, while fast might require an inference based on the previous lower-level attributes. This observation suggests a possible multi-layer semantic model which improves upon our model, where the category classifiers are essentially learned on latent shared features guided with attributes as in our original problem formulation, where they have multiple layers of transformations instead of a single layer. In this multi-layer model, different levels of attributes can be associated with feature learning at 59

77 each layer. Section will discuss more on high-level ideas for this deeper semantic model. The limitation of having a global, binary single intermediate layer, and ignoring the difference in abstraction level between the semantic concepts and groups can be viewed more as a limitation inherent to attributes themselves. Some semantic concepts have more explicit subset relationships among themselves. For example, consider canine and carnivore. We can group animals into canine and non-canine groups, and carnivore and non-carnivore groups as with the attributes, but these have more obvious relation that the former is a subset of the latter. A hierarchical model is more suitable to such cases where we can define a clear subset relation between semantic concepts. In the next chapter, I show how a taxonomy could be exploited to help learn object category recognition. 60

78 Chapter 4 Learning Disjoint Features on a Taxonomy The binary attributes we explored in the previous chapter divide the categories into two groups: those that have the attribute, and those that do not. However, this introduction of a single layer of meta-categories is not the only way to categorize basic level categories into larger groups. Instead, we could merge categories into superclasses by their similarities, and split a category into subcategories by the observed difference, or certain humandesigned criteria, in a hierarchical way. Such a semantic hierarchy is called a taxonomy, and is the second type of external semantic knowledge I explore in this thesis. Well-known taxonomies employed for categorization include WordNet, which groups words into sets of cognitive synonyms and their super-subordinate relations [35], and the phylogenetic tree of life, which groups biological species based on their physical or genetic properties. Critically, such trees implicitly embed cues about human perception of categories, how they relate to one another, and how those relationships vary at different granularities. Thus, in the context of visual object recognition, such a structure has the potential to guide the selection of meaningful low-level features, essentially augmenting the 61

79 Tree of Metrics (ToM) Captures the hierarchical structure of a taxonomy Vehicle Wheeled vehicle Disjoint Sparsity Discovers features at the right semantic granularity Vehicle Wheeled vehicle Bicycle vessel aircraft Motor vehicle Bicycle disjoint Bicycle for two Mountain bike Figure 4.1: Main Idea: Leveraging parent-child relationship in a given semantic taxonomy, we learn a tree of metrics (ToM) that captures compact, discriminative visual features for each level. Left: we learn a local metric at each node of a taxonomy, that discriminates between its subclasses. Right: For the metrics that are associated in an ancestor-descendants relationship, we want each metric to select a set of features different from others, to identify exclusively informative features at each semantic granularity. standard supervision provided by image labels. Some initial steps have been made based on this intuition, typically by leveraging the WordNet hierarchy as a prior on inter-class visual similarity [124, 72, 98, 27, 37, 26, 105]. I propose a metric learning approach 1 to learn discriminative visual representations while also exploiting external knowledge about the target objects semantic similarity. 2 We assume the external knowledge itself is available in the form of a hierarchical taxonomy over the objects (e.g., from WordNet or 1 The work introduced in this chapter is published in [55]. 2 learned representation and learned metric are used interchangeably, since we deal with sparse Mahalanobis metrics, which are equivalent to selecting a subset of features and applying a linear feature space transformation. 62

80 some other knowledge base). My approach exploits these semantics in two novel ways. First, we construct a tree of metrics (ToM) to directly capture the hierarchical structure. In this tree, each metric is responsible for discriminating among its immediate object subcategories. Specifically, we learn one metric for each non-leaf node, and require it to satisfy (dis)similarity constraints generated among its subtree members training instances. We use a variant of the large-margin nearest neighbor objective [112], and augment it with a regularizer for sparsity in order to unify Mahalanobis parameter learning with a simple means of feature selection. Second, rather than learn the metrics at each node independently, I introduce a novel regularizer for disjoint sparsity that couples each metric with those of its ancestors. This regularizer specifies that a disjoint set of features should be selected for a given node and its ancestors, respectively. Intuitively, this represents that the visual features most useful to distinguish the coarse-grained classes (e.g., motor vehicle vs. bicycle. See Figure 4.1) should often be different than those cues most useful to distinguish their fine-grained subclasses (e.g., bicycle for two vs. mountain bike). The resulting optimization problem is convex, and can be optimized with a projected subgradient approach. Figure 4.1 shows the overview of these two main ideas. The ideas of exploiting label hierarchy and model sparsity are not completely new to computer vision and machine learning researchers. Hierarchical classifiers are used to speed up classification time and alleviate data sparsity 63

81 problems [72, 50, 62, 73, 16]. Parameter sparsity is increasingly used to derive parsimonious models with informative features [67, 60, 117]. My novel contribution lies in the idea of ToM and disjoint sparsity together as a new strategy for visual feature learning. My idea reaps the benefits of both schools of thought. Rather than relying on learners to discover both sparse features and a visual hierarchy fully automatically, we use external real-world knowledge expressed in hierarchical structures to bias which sparsity patterns we want the learned discriminative feature representations to exhibit. Thus, our end-goal is not any sparsity pattern returned by learners, but the patterns that are in concert with rich high-level semantics. I validate my approach with the Animals with Attributes [65] and ImageNet [27] datasets using the WordNet taxonomy. We demonstrate that the proposed ToM outperforms both global and multiple-metric metric learning baselines that have similar objectives but lack the hierarchical structure and proposed disjoint sparsity regularizer. In addition, we show that when the dimensions of the original feature space are interpretable (nameable) visual attributes, the disjoint features selected for super- and sub-classes by my method can be quite intuitive. 4.1 Approach I review briefly the techniques for learning distance metrics. I then describe an l 1 -norm based regularization for selecting a sparse set of features in the context of metric learning. Building on that, I proceed to describe our main 64

82 algorithmic contribution, that is, the design of a metric learning algorithm that prefers not only sparse but also disjoint features for discriminating different categories Distance metric learning Many learning algorithms depend on calculating distances between samples, notably k-nearest neighbor classifiers or clustering. While the default is to use the Euclidean distance, the more general Mahalanobis metric is often more suitable. For two data points x i, x j R D, their (squared) Mahalanobis distance is given by (4.1) d 2 M(x i, x j ) = (x i x j ) T M(x i x j ), where M is a positive semidefinite matrix M 0. Arguably, the Mahalanobis distance can better model complex data, as it considers correlations between feature dimensions. Learning the optimal M from labeled data has been an active research topic (e.g., [25, 47, 112]). Most methods follow an intuitively appealing strategy: a good metric M should pull data points belonging to the same class closer and push away data points belonging to different classes. As an illustrative example, we describe the technique used in constructing large margin nearest neighbor (LMNN) classifiers [112], to which our empirical studies extensively compare. In LMNN, each point x i in the training set is associated with two sets 65

83 of different data points in x i s nearest neighbors (identified in the Euclidean distance): the targets whose labels are the same as x i s and the impostors whose labels are different. Let x + i denote the target and x i denote the impostor sets, respectively. LMNN identifies the optimal M as the solution to, (4.2) min M 0 subject to l(m) = d 2 M (x i, x j ) + γ i ijl j x + i 1 + d 2 M (x i, x j ) d 2 M (x i, x l ) ξ ijl ; ξ ijl 0. j x + i, l x i ξ ijl where the objective function l(m) balances two forces: pulling the target towards x i and pushing the impostor away. The latter is characterized by the constraint composed of a triplet of data points: the distance to an impostor should be greater than the distance to a target by at least a margin of 1, possibly with the help of a slack variable ξ ijl. The minimization of equation 4.2 is a convex optimization problem with semidefinite constraints on M 0, and is tractable with standard techniques. My approach extends previous work on metric learning in two aspects: 1) We apply a sparsity-based regularization to identify informative features (Section 4.1.2); 2) at the same time, we seek metrics that rely on disjoint subsets of features for categories at different semantic granularities (Section 4.1.3) Sparse feature selection for metric learning How can we learn a metric such that only a sparse set of features are relevant? Examining the definition of the Mahalanobis distance in equation 4.1, 66

84 we deduce that if the d-th feature of x is not to be used, it is sufficient and necessary for the d-th diagonal element of M be zero. Therefore, analogous to the use of l 1 -norm by the popular LASSO technique [97],we add the l 1 -norm of M s diagonal elements to the large margin metric learning criterion l(m) in equation 4.2, min d (4.3) 2 M(x i, x j ) + γ ξ ijl + λtrace[m], M 0 i ijl j x + i where we have omitted the constraints as they are not changed. λ and γ are nonnegative (hyper)parameters trading off the sparsity of the model and the other parts in the objective. Note that since the matrix trace Trace[ ] is a linear function of its argument, this sparse feature metric learning problem remains a convex optimization Learning a tree of metrics (ToM) with disjoint visual features How can we learn a tree of metrics so each metric uses features disjoint from its ancestors? Using disjoint features To characterize the disjointness between two metrics M t and M t, we use the vectors of their nonnegative diagonal elements v t and v t as proxies to which features are (more heavily) used. This is a reasonable choice as we use the sparsity-inducing l 1 -norm in learning the metrics. We measure their degree of competition for common features, (4.4) C(M t, M t ) = v t + v t 2 2. Intuitively, if a feature dimension is not used by either metric, the competition 67

85 for that feature is low. If a feature dimension is used by both metrics heavily, then the competition is high. Consequently, minimizing eq. (4.4) as a regularization term will encourage different metrics to use disjoint features. Note that the measure is a convex function of v t and v t, hence also convex in M t and M t. Learning a tree of metrics Formally, assume we have a tree T where each node corresponds to a category. Let t index the T non-leaf or internal nodes. We learn a metric M t to differentiate its children categories c(t). For any node t, we use D(t) to denote those training samples whose labeled categories are offspring of t, and a(t) to denote the nodes on the path from the root to t. To learn our metrics {M t } T t=1, we apply similar strategies of learning metrics for large-margin nearest neighbor classifiers. We cast it as a convex optimization problem: (4.5) min {M t} 0 subject to d 2 M t (x i, x j ) + γ t + t c c(t) i,j D(c) a a(t) γ ta C(M t, M a ) t,c,r,ijlξ tcrijl + t λ t Trace[M t ] t, c c(t), r c(t) {c}, x i, x j D(c), x l D(r) 1 + d 2 M t (x i, x j ) d 2 M t (x i, x l ) ξ tcrijl ; ξ tcrijl 0. In short, there are T learning (sub)problems, one for each metric. Each metric learning problem is in the style of the sparse feature metric learning eq. (4.3). However, more importantly, these metric learning problems are coupled to- 68

86 gether through the disjoint regularization. Our disjoint regularization encourages a metric M t to use different sets of features from its super-categories categories on the tree path from the root. Numerical optimization The optimization problem in Equation (4.5) is convex, though nonsmooth due to the nonnegative slack variables. We use the subgradient method, previously used for similar problems [112]. For problems with a large taxonomy, learning all the regularization coefficients λ t and γ ta is prohibitive, as the number of coefficient combinations is O(k T ), where T is the number of nodes and k is the number of values a coefficient can take. Thus, for the large-scale problems we focus on, we use a simpler and computationally more efficient strategy of Sequential Optimization (SO) by sequentially optimizing one metric at a time. Specifically, we optimize the metric at the root node and then its children, assuming the metric at the root is fixed. We then recursively (in breadth-first-search) optimize the rest of the metrics, always treating the metrics at the higher level of the hierarchy as fixed. This strategy has a significantly reduced computational cost of O(kT). In addition, the SO procedure allows each metric to be optimized with different parameters and prevents a badly-learned low-level metric from influencing upper-level ones through the disjoint regularization terms. (This can also be achieved by adjusting all regularization coefficients in parallel through extensive cross-validation, but at a much higher computational expense.) Using a tree of metrics for classification Once the metrics at all nodes are learned, they can be used for several classification tasks (e.g., with 69

87 k-nn or as a kernel to a SVM). In this work, we study two tasks in particular: 1) We consider per-node classification, where the metric at each node is used to discriminate its sub-categories. Since decisions at higher-level nodes must span a variety of object sub-categories, these generic decisions are interesting to test the learned features in a broader context. 2) We consider hierarchical classification [33], a natural way to use the full ToM. In this case, we examine the recognition accuracy for the finest-level categories only. We classify an object from the root node down; the leaf node that terminates the path is the predicted label. I stress that our metric learning criterion of Equation (4.5) aims to minimize classification errors at each node. Thus, improvement in per-node accuracy is more directly indicative of whether the learning has resulted in useful metrics. Understanding the relation between per-node and full multiclass accuracy has been a challenging research problem in building hierarchical classifiers [16, 72]. Relationship to orthogonal transfer Our work shares a similar spirit to the orthogonal transfer idea explored in [121]. The authors there use non-overlapping features to construct multiple SVM classifiers for hierarchical classification of text documents. Concretely, they propose an orthogonal regularizer ij K ij wi Tw j where w i and w j are the SVM parameters. Minimizing it will reduce the similarity of the parameter vectors and make them orthogonal to each other. However, orthogonality does not necessarily imply disjoint features. This can be seen with a contrived two-dimensional counterex- 70

88 ample where w i = [1 1] T and w j = [ 1 1] T. Both features are used, yet the two parameter vectors are orthogonal. In contrast, our disjoint regularizer Equation (4.4) is more indicative of true disjointness. Specifically, when our regularizer attains its minimum value of zero, we are guaranteed that features are non-overlapping as our v i and v j are nonnegative diagonal elements of positive semidefinite matrices. Our regularizer is also guaranteed to be convex, whereas the convexity of the regularizer in [121] depends critically on tuning K ij. 4.2 Results baselines: We validate our ToM approach on several datasets, and consider three Euclidean: Euclidean distance in the original feature space Global LMNN: a single global metric for all classes learned with the LMNN algorithm [112] Multi-Metric LMNN: one metric learned per class using the multiple metric LMNN variant [112]. We chose these baselines to show the advantage of learning a tree of feature spaces over a global feature space, or a set of category-specific feature spaces. Note that our method learns features represented as metrics, instead of classifiers, and can be couple with any classifiers (e.g. SVM) other than 71

89 the k-nearest neighbor (knn) classifier we use for the experiments. Thus, our method is not directly comparable to other hierarchical methods tied to specific classifiers such as [62, 73, 16], since our focus is not on showing the advantage of using a knn classifier over other classifiers. We use the code provided by the authors. To evaluate the influence of each aspect of our method, we test it under three variants: ToM: ToM learning without any regularization terms ToM+Sparsity: ToM learning with the sparsity regularization term ToM+Disjoint: ToM learning with the disjoint regularization term. For all experiments, we test with five random data splits of 60%/20%/20% for train/validation/test. We use the validation data to set the regularization parameters λ and γ among candidate values {0, 1, 10, 100, 1000}, and we generate 500 (x i, x j, x l ) training triplets per class Proof of concept on synthetic dataset First we use synthetic data to clearly illustrate disjoint sparsity regularization. We generate data with precisely the property that coarser categories are distinguishable using feature dimensions distinct from those needed to discriminate their subclasses. Specifically, we sample 2000 points from each of four 4D Gaussians, giving four leaf classes {a, b, c, d}. They are grouped into two superclasses A = {a, b} and B = {c, d}. The first two dimensions of all 72

90 root:{a,b,c,d} Synthetic Features A:{a,b} B:{c,d} value a b c d (a) Class Hierarchy a b c d (b) Means of the features (c) TOM (d) TOM + Sparsity (e) TOM + Disjoint Figure 4.2: Synthetic dataset example. Our disjoint regularizer yields a sparse metric that only considers the feature dimension(s) necessary for discrimination at that given level. points are specific to the superclass decision (A vs. B), while the last two are specific to the subclasses. See Fig. 5.1 (a) and (b). We run hierarchical k-nearest neighbor classification (k = 3) on the test set. ToM+Sparsity increases the recognition rate by 0.90%, while ToM+Disjoint increases it by 4.05%. Thus, as expected, disjoint sparsity does best, since it selects different features for the super- and sub-classes. Accordingly, in the learned Mahalanobis matrices for each node (Fig. 5.1(c)-(e)), we see disjoint sparsity zeros out the unneeded features for the upper-level metric, showed as black squares in the figure (e). In contrast, the ToM+Sparsity features are sub-optimal and fit to some noise in the data (d). 73

91 4.2.2 Visual recognition experiments tasks. Next we demonstrate our approach on challenging visual recognition Datasets and implementation details We validate with three datasets drawn from two publicly available image collections: Animals with Attributes (AWA) [65] and ImageNet [27, 26]. Both are well-suited for our scenario, since they consist of fine-grained categories that can be grouped into more general object categories. From the AWA (Figure 3.2) that contains 30,475 images and 50 animal classes, and ImageNet image collections, we form three datasets for empirical validation. AWA-PCA, which uses the features provided from the dataset in [65] (SIFT, rgsift, PHOG, SURF, LSS, RGB), concatenated, standardized, and PCA-reduced to 50 dimensions. AWA-ATTR, which uses 85-dimensional attribute predictions as the original feature space, formed by concatenating the outputs of 85 linear SVMs trained to predict the presence/absence of the 85 nameable properties annotated by [65], e.g., furry, white, quadrupedal, etc. VEHICLE-20, which uses 20 vehicle classes and 26,624 images from ImageNet, and apply PCA to reduce the authors provided visual word features [26] to 50 dimensions per image 3. 3 This is the dimensionality that worked best for the Global LMNN baseline. 74

92 Figure 4.3: Examples images for VEHICLE-20 dataset. We use WordNet to generate the semantic hierarchies for all datasets. We retrieve all nodes in WordNet that contain any of the object class names on their word lists. In the case of homonyms, we manually disambiguate the word sense. Then, we build a compact partial hierarchy over those nodes by 1) pruning out any node that has only one child (i.e., removing superfluous nodes), and 2) resolving any instances of multiple parentship by choosing the path from the leaf to root having the most overlap with other classes. See Figures 4.4 and 4.5 for the resulting AWA and VEHICLE trees. Throughout, we evaluate classification accuracy using k-nearest neighbors (k-nn). For ToM, at node n we use k = 2 ln 1 +1, where l n is the level of the node, and l n = 1 for leaf nodes. This means we use a larger k at the higher nodes in the tree where there is larger intra-class variation, in an effort to be more robust to outliers. For the Euclidean and LMNN baselines, which lack a hierarchy, we simply use k=3. Note that ToM s setting at the final decision nodes (just above a leaf) is also k = 3, comparable to the baselines. 75

93 Per-node accuracy and analysis of the learned representations Since our algorithm optimizes the metrics at every node, we first examine the resulting per-node decisions. That is, how accurately can we predict the correct subcategory at any given node? The bar charts in Figures 4.4 and 4.5 show the results, in terms of raw k-nn accuracy improvements over the Euclidean baseline. For reference, we also show the Global LMNN baseline. Multi-Metric LMNN is omitted here, since its metrics are only learned for the leaf node classes. We observe a good increase for most classes, as well as a clear advantage relative to LMNN. Furthermore, our results are usually strongest when including the novel disjoint sparsity regularizer. This result supports our basic claim about the potential advantage of exploiting external semantics in ToM. We find that absolute gains are similar in either the PCA or ATTR feature spaces for AWA, though exact gains per class differ. While the ATTR variant exposes the semantic features directly to the learner, the PCA variant encapsulates an array of low-level descriptors into its dimensions. Thus, while we can better interpret the meaning of disjoint sparsity on the attributes, our positive result on raw image features assures that disjoint feature selection is also amenable in the more general case. 76

94 placental ungulate carnivore even toed ungulate aquatic mammal canine feline ruminant primate whale dog cat odd toed ungulate bovid g.ape pinniped dolphin bear procyonid sheperd rodent domestic equine baleen musteline big cat deer bovine mole bat rabbit elephant beaver rat squirrel hamster mouse seal chimpanzee gorilla spider monkey walrus killer whale blue whale skunk grizzly bear humpback common dolphin polar bear fox raccoon giant panda otter weasel wolf dalmatian collie tiger leopard German shepherd Chihuahua lion bobcat Persian cat horse rhinoceros Siamese cat zebra giraffe moose pig hippopotamus sheep antelope deer ox buffalo cow Accuracy improvement Accuracy improvement equine musteline big cat dolphin deer procyonid bovid sheperd dog bear bovine pinniped pinniped domestic bovid big cat procyonid g.ape deer dolphin AWA PCA musteline odd toed cat primate AWA ATTR placental canine even toed carnivore aquatic ruminant whale rodent domestic feline ungulate baleen Global LMNN: 1.33 TOM: 1.44 TOM+Sparsity: 1.93 TOM+Disjoint: 2.15 g.ape Global LMNN: 1.01 TOM: 1.53 TOM+Sparsity: 1.94 TOM+Disjoint: 2.45 sheperd primate rodent whale equine dog bear ruminant canine cat ungulate even toed aquatic odd toed carnivore baleen feline placental bovine Figure 4.4: Semantic hierarchy for AWA (top row) and the per-node accuracy improvements relative to Euclidean distance, for the AWA-PCA (middle row) and AWA-ATTR (bottom row) datasets. Numbers in legends denote average improvement over all nodes. We generally achieve a sizable accuracy gain relative to the Global LMNN baseline (dark left bar for each class), showing the advantage of exploiting external semantics with our ToM approach. 77

95 vehicle wheeled vehicle liner containership craft balloon airship airliner warplane speedboat canoe gondola vessel aircraft motor vehicle mountainbike bicyclefortwo motorscooter steam locomo. electric locomo. pickup garbagetruck racer convertible cab self propelled vehicle ship boat h. air l. air bicycle locomotive car truck trailertruck Accuracy improvement VEHICLE 20 Global LMNN: 0.86 TOM: 2.42 TOM+Sparsity: 2.79 TOM+Disjoint: lighter air vehicle ship aircraft truck locomotive craft boat wheeled car bicycle vessel self prop. motor vehicle heavier air Figure 4.5: Semantic hierarchy for VEHICLE-20 and the per-node accuracy gains, plotted as above. To look more closely at this, Table 4.1 displays representative superclasses from AWA-ATTR together with the attributes that ToM+Disjoint selects as discriminative for their subclasses. The attributes shown are those with nonzero weights in the learned metrics. Intuitively, we see that often the selected attributes are indeed useful for discriminating the child classes. For example, tusks and plankton attributes help distinguish common dolphins 78

96 Superclass Subclasses Attributes selected whale dolphin black, white, blue, gray, toughskin, chewteeth, strainteeth, smelly, slow, muscle, active, fish, hunter, skimmer, oldworld, arctic... tusks, plankton, blue, gray, red, patches, slow, muscle, active, insects oddtoed ungulate equine dolphin, baleen whale common dolphin, killer whale equine, rhinoceros horse, zebra fast, longneck, hairless, black, white, yellow, patches, spots, bulbous, longleg, buckteeth, horns, tusks, smelly... stripes, domestic, orange, red, yellow, toughskin, newworld, arctic, bush Table 4.1: Attributes selected by ToM+Disjoint for various superclass objects in AWA. See text. from killer whales, whereas stripes and domestic help distinguish zebras from horses. At the same time, as desired, we see that the attributes useful for coarser-level categories are distinct from those employed to discriminate the finer-level objects. For example, fast, longneck, or hairless are used to differentiate equine from rhino, but are excluded when differentiating horses from zebras (equine s subclasses) Hierarchical multi-class classification accuracy Next we evaluate the complete multi-class classification accuracy, where we use all the learned ToM metrics together to predict the leaf-node label of the test points. This is a 50-way task for AWA, and a 20-way task for VEHICLES. 79

97 AWA-ATTR Method Correct label Semantic similarity Euclidean ± ± 0.26 Global LMNN ± ± 0.88 Multi-metric LMNN ± ± 0.71 ToM ± ± 0.09 ToM + Sparsity ± ± 0.58 ToM + Disjoint ± ± 0.62 AWA-PCA Method Correct label Semantic similarity Euclidean ± ± 0.58 Global LMNN ± ± 0.32 Multi-metric LMNN ± ± 0.31 ToM ± ± 0.43 ToM + Sparsity ± ± 0.34 ToM + Disjoint ± ± 0.19 Table 4.2: Multi-class hierarchical classification accuracy and semantic similarity on the AWA-ATTR and AWA-PCA datasets. Numbers are averages over 5 splits, and standard errors for 95% confidence interval. Our method outperforms the baselines in almost all cases, and notably provides more semantically close predictions. See text. VEHICLE-20 Method Correct label Semantic similarity Euclidean ± ± 0.41 Global LMNN ± ± 0.45 Multi-metric LMNN ± ± 0.54 ToM ± ± 0.54 ToM + Sparsity ± ± 0.26 ToM + Disjoint ± ± 0.21 Table 4.3: Multi-class hierarchical classification accuracy and semantic similarity on the VEHICLE-20 dataset. 80

98 Table 4.2 and 4.3 shows the results. We score accuracy in two ways: Correct label records the percentage of examples assigned the correct (leaf) label, while Semantic similarity records the semantic similarity between the predicted and true labels. For both, higher is better. The former is standard recognition accuracy, while the latter gives a more nuanced view of the semantic magnitude of the classifiers errors. Specifically, we calculate the semantic similarity between classes (nodes) i and j using the metric defined in [37], which counts the number of nodes shared by their two parent branches, divided by the length of the longest of the two branches. In the spirit of other recent evaluations [9, 26, 37], this metric reflects that some errors are worse than others; for example, calling a Persian cat a Siamese cat is a less glaring error than calling a Persian cat a horse. This is especially relevant in our case, since our key motivation is to instill external semantics into the feature learning process. In terms of pure label correctness, ToM improves over the strong LMNN baselines for both AWA-ATTR and VEHICLE-20. Further, in all cases, we see that disjoint sparsity is an important addition to ToM. However, in AWA-PCA, Global LMNN produces the best results by a statistically insignificant margin. We did not find a clear rationale for this one case. For AWA-ATTR, however, our method is substantially better than Global LMNN, perhaps due to our method s strength in exploiting semantic features. While we initially expected Multi-Metric LMNN to outperform Global LMNN, we suspect it struggles with clusters that are too close together. For all cases when ToM+Disjoint 81

99 outperforms the LMNN or Euclidean baselines, the improvement is statistically significant. In terms of semantic similarity, ToM is better than all baselines on all datasets. This is a very encouraging result, since it suggests our approach is in fact leveraging semantics in a useful way. In practice, the ability to make such reasonable errors is likely to be increasingly important as the community tackles larger and larger multi-class recognition problems. 4.3 Discussion I presented a new metric learning approach for visual recognition that integrates external semantics about object hierarchy. Experiments with challenging datasets indicate its promise, and support our hypothesis that outside knowledge about how objects relate is valuable for feature learning. Instead of learning a discriminative metric that considers each category as a separate, independent entity, the proposed ToM learns metrics that preserve the distances between each group of categories at different semantic levels. Further, the added disjoint regularizer forces feature spaces that form ancestor-descendants relationships to compete for the features, which is shown to be effective in isolating features for each semantic granularity. The true selection of features and the convexity is what makes our method superior to the existing exclusive regularization methods based on competition [122, 121]. Both the hierarchical modeling and the isolation of the feature spaces were shown to be useful for hierarchical classification. However, it could still suf- 82

100 fer from the problem known as the semantic gap the discrepancy between the semantic and the visual space, which could limit the classification performance at the abstract high-level. This in turn could limit the performance of the whole model, due to the error propagating nature of the hierarchical classification model. There could be multiple possible solutions to this problem. The first is to construct a hierarchy that can account for both semantics and visual distributions. This could be done by either collapsing or splitting the nodes of the existing semantic taxonomy such that the taxonomy aligns better with the visual distribution, or constructing a hierarchy from the scratch while accounting for both semantic and visual similarities between categories. However, doing so might result in less semantic information being exploited, since our main idea was to exploit human criteria in grouping or splitting of the categories, where a large amount of useful semantic information comes from higher-level nodes representing abstract classes such as vehicle or carnivore. These highlevel nodes usually contain visually diverse subcategories but are nonetheless informative. Empirical results from [18] show that even an evaluation scheme that considers the whole path and holds off from making a hard decision at each node might not cope well with such abstract high-level semantic nodes. In addition to this semantic gap problem, there exists another problem: no single semantic taxonomy is perfect, and learning an optimal one is infeasible since different applications and views would prefer different groupings. How can we then overcome this inevitable limitation with a single semantic 83

101 taxonomy? The next chapter will explore this question. 84

102 Chapter 5 Combining Complementary Information in Multiple Taxonomies In the previous chapter, we have seen how a semantic taxonomy can be used to help category recognition by providing information to isolate granularityspecific features, and to hierarchically classify objects. Two fundamental issues, however, complicate its use. First, a given taxonomy may offer hints about visual relatedness, but its structure need not entirely align with useful splits for recognition. (For example, monkey and dog are fairly distant semantically according to WordNet, yet they share a number of visual features. An apple and applesauce are semantically close, yet are easily separable with basic visual features.) Thus, the hierarchical structure provided by a semantic taxonomy is often non-optimal for hierarchical classification. Second, given the complexity of visual objects, it is highly unlikely that some single optimal semantic taxonomy exists to lend insight for recognition. While previous work relies on a single taxonomy out of convenience, in reality objects can be organized along many semantic dimensions or views. (For example, a Dalmatian belongs to the same group as the wolf according to a biological taxonomy, as both are canines. However, in terms of visual attributes, it can be grouped with the leopard, as both are spotted; in terms of habitat, it can be grouped 85

103 canine vs. feline Biological spotted vs. pointy Appearance domestic vs. wild Habitat canine feline Spotted Pointy Ears Domestic Wild Dalmatian wolf Siamese cat leopard Dalmatian Leopard Siamese cat Wolf Dalmatian Siamese Cat Wolf leopard Figure 5.1: Main idea: For a given set of classes, we assume multiple semantic taxonomies exist, each one representing a different view of the inter-class semantic relationships. Rather than commit to a single taxonomy which may or may not align well with discriminative visual features we learn a tree of kernels for each taxonomy that captures the granularity-specific similarity at each node. Then we show how to exploit the inter-taxonomic structure when learning a combination of these kernels from multiple taxonomies (i.e., a kernel forest ) to best serve the object recognition tasks. with the Siamese cat, as both are domestic. See Figure 5.1.) Motivated by these issues, I next present a discriminative feature learning approach that leverages multiple taxonomies capturing different semantic views of the object categories 1. The key insight here is that some combination of the semantic views will be most informative to distinguish a given visual category. Continuing with the sketch in Figure 5.1, that might mean that the first taxonomy helps learn dog- and cat-like features, while the second taxonomy helps elucidate spots and pointy corner features, while the last reveals context cues such as proximity to humans or indoor scene features. While each view differs in its implicit human-designed splitting criterion, all separate 1 The work introduced in this chapter is published in [56]. 86

104 some classes from others, thereby lending (often complementary) discriminative cues. Thus, rather than commit to a single representation, we aim to inject pieces of the various taxonomies as needed. To this end, I propose semantic kernel forests. This novel kernel learning method takes as input training images labeled according to their object category, as well as a series of taxonomies, each of which hierarchically partitions those same labels (object classes) by a different semantic view. For each taxonomy, we first learn a tree of semantic kernels: each node in a tree has a Mahalanobis-based kernel optimized to distinguish between the classes in its children nodes. Following on ToM approach from the previous chapter, the kernels in one tree isolate image features useful at a range of category granularities. Then, using the resulting semantic kernel forest from all taxonomies, we apply a form of multiple kernel learning (MKL) to obtain class-specific kernel combinations, in order to select only those relationships relevant to recognize each object class. We introduce a novel hierarchical regularization term into the MKL objective that further exploits the taxonomies structure. The output of the method is one learned kernel per object class, which we can then deploy for one-versus-all multi-class classification on novel images. The main contribution of the work introduced in this chapter is to simultaneously exploit multiple semantic taxonomies for visual feature learning. Whereas past work focuses on building object hierarchies for scalable classification [113, 28] or using WordNet to gauge semantic distance [71, 98, 37, 26], we learn discriminative kernels that capitalize on the cues in diverse taxonomy 87

105 views, leading to better recognition accuracy. The primary technical contributions are i) an approach to generate semantic base kernels across taxonomies, ii) a method to integrate the complementary cues from multiple suboptimal taxonomies, and iii) a novel regularizer for multiple kernel learning that exploits hierarchical structure from the taxonomy, allowing kernel selection to benefit from semantic knowledge of the problem domain. I demonstrate my approach with challenging images from the Animals with Attributes and ImageNet datasets [65, 27] together with taxonomies spanning cognitive synsets, visual attributes, behavior, and habitats. The results show that the taxonomies can indeed boost feature learning, letting us benefit from humans perceived distinctions as implicitly embedded in the trees. Furthermore, I show that interleaving the forest of multiple taxonomic views leads to the best performance, particularly when coupled with the proposed novel regularization. 5.1 Approach I cast the problem of learning semantic features from multiple taxonomies as learning to combine kernels. The base kernels capture features specific to individual taxonomies and granularities within those taxonomies, and they are combined discriminatively to improve classification, weighing each taxonomy and granularity only to the extent useful for the target classification task. I describe the two main components of the approach in turn: construct- 88

106 ing the base kernels from the learned tree of metrics on each taxonomy which we call a semantic kernel forest (Sec ), and learning their combination across taxonomies (Sec ), where we devise a new hierarchical regularizer for MKL. In what follows, we assume that we are given a labeled dataset D = {(x i, y i )} N n=1 where (x i, y i ) stands for the ith instance (feature vector) and its class label is y i, as well as a set of tree-structured taxonomies {T t } T t=1. Each taxonomy T t is a collection of nodes. The leaf nodes correspond to class labels, and the inner nodes correspond to superclasses or, more generally, semantically meaningful groupings of categories. We index those nodes with double subscripts tn, where t refers to the tth taxonomy and n to the nth node in that taxonomy. Without loss of generality, we assign the leaf nodes (i.e., the class nodes) a number between 1 and C, where C is the number of class labels Learning a semantic kernel forest The first step is to learn a forest of base kernels. These kernels are granularity- and view-specific; that is, they are tuned to similarities implied by the given taxonomies. While base kernels are learned independently per taxonomy, they are learned jointly within each taxonomy, as we describe next. Formally, for each taxonomy T t, we learn a set of Gaussian kernels for the superclass at every internal node tn for which n C + 1. The Gaussian 89

107 kernels are parameterized as (5.1) K tn (x i, x j ) = exp{ γ tn d 2 M tn (x i, x j )} = exp{ γ tn (x i x j ) T M tn (x i x j )}, where the Mahalanobis distance metric M tn is used in lieu of the conventional Euclidean metric. Note that for leaf nodes where n C, we do not learn base kernels. We want the base kernels to encode similarity between examples using features that reflect their respective granularity in the taxonomy. Certainly, the kernel K tn should home in on features that are helpful to distinguish the node tn s subclasses. Beyond that, however, we specifically want it to use features that are as different as possible from the features used by its ancestors. Doing so ensures that the subsequent combination step can choose a sparse set of disconnected features. To that end, we apply our Tree of Metrics (ToM) technique introduced in the previous chapter to learn the Mahalanobis parameters M tn. To recap, In ToM, metrics are learned by balancing two forces: i) discriminative power and ii) a preference for different features to be chosen between parent and child nodes. The latter exploits the taxonomy semantics, based on the intuition that features used to distinguish more abstract classes (dog vs. cat) should differ from those used for finer-grained ones (Siamese vs. Persian cat). Briefly, for each node tn, the training data is reduced to D n = {(x i, y in )}, where y in is the label of n s child x i. If x i s class label y i is not a descendant 90

108 of the superclass at the node n, then x i is excluded from D n. The metrics are learned jointly, with each node mutually encouraging the others to use non-overlapping features. ToM achieves this by augmenting a large margin nearest neighbor [112] loss function n l(d n; M tn ) with the following disjoint sparsity regularizer: (5.2) Ω d (M) = λ Trace[M tn ]+µ diag(m tn )+diag(m tm ) 2 2, n C+1 n C+1 m n where m n denotes that node m is either an ancestor or descendant of n. The first part of the regularizer encourages sparsity in the diagonal elements of M tn, and the second part incurs a penalty when two different metrics compete for the same diagonal element, i.e., to use the same feature dimension. The resulting optimization problem is convex and can be solved efficiently [55]. After learning the metrics {M tn } in each taxonomy, we construct base kernels as in eq. (5.1). The bandwidths γ tn are set as the average distances on training data. We call the collection F = {K tn } of all base kernels the semantic kernel forest. Figure 5.1 shows an illustrative example. While ToM has shown promising results in learning metrics in a single taxonomy, its reliance on linear Mahalanobis metrics is inherently limited. A straightforward convex combination of ToMs would result in yet another linear mapping, incapable of capturing nonlinear inter-taxonomic interactions. In contrast, our kernel approach retains ToM s granularity-specific features but also enables nontrivial (nonlinear) combinations, especially when coupled with a novel hierarchical regularizer, which I will define next. 91

109 5.1.2 Learning class-specific kernels across taxonomies Base kernels in the semantic kernel forest are learned jointly within each taxonomy but independently across taxonomies. To leverage multiple taxonomies and to capture different semantic views of the object categories, we next combine them discriminatively to improve classification. In the following, I first describe a basic form of combining. I then describe our novel hierarchical regularization to incorporate semantic and structural knowledge in the combining process. Basic setting To learn class-specific features (or kernels), we compose a one-versus-rest supervised learning problem. Additionally, instead of combining all the base kernels in the forest F, we pre-select a subset of them based on the taxonomy structure. Specifically, from each taxonomy, we select base kernels that correspond to the nodes on the path from the root to the leaf node class. For example, in the Biological taxonomy of Figure 5.1, for the category Dalmatian, this path includes the nodes (superclasses) canine and animal. Thus, for class c, the linearly combined kernel is given by (5.3) F c (x i, x j ) = β ctn K tn (x i, x j ), t n c where n c indexes the nodes that are ancestors of c, which is a leaf node (recall that the first C nodes in every taxonomy are reserved for leaf class nodes). The combination coefficients β ctn are constrained to be nonnegative to ensure the positive semidefiniteness of the resulting kernel F c (, ). 92

110 We apply the kernel F c (, ) to construct the one-versus-rest binary classifier to distinguish instances from class c from all other classes. We then optimize β c = {β ctn } such that the classifier attains the lowest empirical misclassification risk. The resulting optimization (in its dual formulation) is analogous to standard multiple kernel learning [8]: (5.4) min max β c α c s.t. α ci 1 α ci α cj q ci q cj F c (x i, x j ) 2 i i j α ci q ci = 0, 0 α ci C, i, i where α c is the Lagrange multipliers for the binary SVM classifier, C is the regularizer for the SVM s hinge loss function, and q ci = ±1 is the indicator variable of whether or not x i s label is c. Hierarchical regularization Next, we extend the basic setting to incorporate richer modeling assumptions. We hypothesize that kernels at higherlevel nodes should be preferred to lower-level nodes. Intuitively, higher-level kernels relate to more classes, thus are likely essential to reduce loss. We leverage this intuition and knowledge about the relative priority of the kernels from each taxonomy s hierarchical structure. We design a novel structured regularization that prefers larger weights for a parent node compared to its children. Formally, the proposed MKL-H regularizer is given by: (5.5) Ω(β c ) = λ β ctn + µ max(0, β ctn β ctpn + 1). t,n c t,n c The first part prefers a sparse set of kernels. The second part (in the form of hinge loss) encodes our desire to have the weight assigned to a node n be less 93

111 than the weight assigned to the node s parent p n. We also introduce a margin of 1 to further increase the difference between the two weights. Hierarchical regularization was previously explored in [7], where a mixed (1, 2)-norm is used to regularize the relative sizes between the parent and the children. The main idea there is to discard children nodes if the parent is not selected. Our regularizer is somewhat similar in spirit, but we devise a simpler and more computationally efficient formulation. (Despite our complexity advantage, preliminary results do not indicate [7] has any empirical advantage over ours.) Numerical optimization The learning problem is cast as a convex optimization that balances the discriminative loss in equation 5.4 and the regularizer in equation 5.5: (5.6) min β c f(β c ) = g(β c ) + Ω(β c ), s.t. β c 0, where we use the function g(β) to encapsulate the inner maximization problem over α c in equation 5.4. We use the projected subgradient method to solve eq. (5.6), for its ease of implementation and practical effectiveness [13]. Specifically, at iteration t, let β t c be the current value of β. We compute f(β c) s subgradient s t, then perform the following update, (5.7) β t+1 c max ( 0, β t c α ts t ), 94

112 where the max( ) function implements the projection operation such that the update does not fall outside of the feasible region β c 0. For step size α t, we use the modified Polyak step size rule. Subgradient Update Rule g(β c ) encapsulates the inner maximization problem over α c in eq.(5.4), and is a differentiable function of β c where g is given as follows: (5.8) g β ctn = 1 2 α ci α cj q ci q cj F ctn (x i, x j ) ij The computation of g/ β ctn only depends on the the α, which is the solution of Eq.(5.4), that could be obtained using an off-the shelf SVM solver. We solve this using LIBSVM [19]. The second term of f(β c ), Ω(β c ) is nondifferentiable but convex. Thus, its subgradients with respect to β c exist, and defined as, (5.9) Ω(β ctn ) = λ + µ r ctnp(n) k C(tn) r ctkn,where C(tn) is the set of children node of tn. r ctij = 1 if β cti β ctj 1 and 0 otherwise. From the subgradient rule f = g + Ω is a subgradient for f. After obtaining the subgradient f, we could use the following update rule using the modified Polyak s stepsize rule to minimize f to its direction. 95

113 (5.10) βc t+1 max ( 0, βc t f(βt c ) ˆf ) t + δ f(β f(β c ) t 2 c ) t 2 where the max( ) function implements the projection operation such that the update does not fall outside of the feasible region β c 0. ˆft is an estimate of the optimal value of the objective function and is defined as (5.11) ˆft = min 0 j t f(βj c ) The variable δ is a constant controlling how close the update rule converges to the optimum. We set it such that in about 500 iterations, the update converges. 5.2 Experiments We validate our approach on multiple image datasets, and compare to several informative baselines Image datasets We use three datasets taken from two publicly available image collections: Animals with Attributes (AWA) [65] and ImageNet [27] 2. We form two datasets from AWA (Figure 3.2). The first consists of the four classes shown in 2 attributes.kyb.tuebingen.mpg.de/ and image-net.org/challenges/lsvrc/2011/ 96

114 bridge feather boa strawberry acorn bonsai daisy sunflower basketball bathtub comb police van lamp pool table rule buckle Figure 5.2: Example images for ImageNet-20 dataset Fig. 5.1, and totals 2, 228 images; the second contains the ten classes in [65], and totals 6, 180 images. We refer to them as AWA-4 and AWA-10, respectively. The third dataset, ImageNet-20 (Figure 5.2), consists of 28, 957 total images spanning 20 classes from ILSVRC2010. We chose classes that are non-animals (to avoid overlap with AWA) and that have attribute labels [88] Taxonomies To obtain multiple taxonomies per dataset, we use attribute labels and WordNet. As discussed above, attributes are human understandable properties shared among object classes, e.g., furry, flat, carnivorous [65]. AWA and ImageNet have 85 and 25 attribute labels, respectively. To form semantic taxonomies based on attributes, we first manually divide the attribute labels into subsets according to their mutual semantic relevance (e.g., furry and 97

115 placental carnivore feline Persian cat procyonid leopard raccoon aquatic giant panda seal even toed humpback hippopotamus pig rat chimpanzee seal humpback hippopotamus pig rat raccoon Persian cat leopard chimpanzee giant panda (a) WordNet (b) Appearance seal humpback hippopotamus pig Persian cat giant panda rat raccoon chimpanzee leopard pig Persian cat leopard giant panda hippopotamus chimpanzee seal humpback rat raccoon (c) Behavior (d) Habitat device instrumentality fruit ride rollercoaster percussion ferriswheel marimba fastener drum button buckle furniture rule lamp pooltable policevan vascular plant comb bathtub flower basketball sunflower daisy bonsai acorn strawberry featherboa bridge rule comb drum lamp buckle bathtub acorn button sunflower daisy featherboa basketball marimba bridge rollercoaster bonsai strawberry pooltable policevan ferriswheel (e) Wordnet (f) Appearance bonsai acorn strawberry daisy bathtub featherboa comb lamp bridge buckle button rollercoaster ferriswheel pooltable marimba rule drum basketball sunflower policevan (g) Attributes Figure 5.3: Taxonomies for the AWA (a-d) and ImageNet-20 (e-g) datasets.

116 shiny are attributes relevant for an Appearance taxonomy, while land-dwelling and aquatic are relevant for a Habitat taxonomy.). Then, for each subset of attributes, we perform agglomerative clustering using Euclidean distance on vectors of the training images real-valued attributes. We restrict the tree height (6 for ImageNet and 3 for AWA) to ensure that the branching factor at the root is not too high. To extract a WordNet taxonomy, we find all nodes in WordNet that contain the object class names on their word lists, and then build a hierarchy by pruning nodes with only one child and resolving multiple parentship. For AWA-10, we use 4 taxonomies: one from WordNet, and three based on attribute subsets reflecting Appearance, Behavior, and Habitat ties. For ImageNet-20, we use 3 taxonomies: one from WordNet, one reflecting Appearance as found by hierarchical clustering on the visual features, and one reflecting Attributes using annotations from [88]. For the AWA-4 taxonomies, we simply generate all 3 possible 2-level binary trees, which, based on manual observation, yield taxonomies reflecting Biological, Appearance, and Habitat ties between the animals. See Figures 5.1 and 5.3. I stress that these taxonomies are created externally with human knowledge, and thus they inject perceived object relationships into the feature learning problem. This is in stark contrast to prior work that focuses on optimizing hierarchies for efficiency, without requiring interpretability of the trees themselves [50, 113, 28, 41]. The two image datasets we employ are annotated with both object la- 99

117 Dataset Group Name Attributes Appearance black, white, blue, brown, gray, orange, red, AWA-10 yellow, patches, spots, stripes, furry, hairless, toughskin, big, small, bulbous, lean, flippers, hands, hooves, pads, paws, longleg, longneck, tail, chewteeth, meatteeth, buckteeth, strainteeth, horns, claws, tusks Behavior smelly, flys, hops, swims, tunnels, walks, fast, slow, strong, weak, muscle, bipedal, quadrupedal, active, inactive, nocturnal, hibernate, agility, fish, meat, plankton, vegetation, insects, forager, grazer, hunter, scavenger, skimmer, stalker Habitat newworld, oldworld, arctic, coastal, desert, bush, plains, forest, fields, jungle, ocean, ground, water, tree, cave ImageNet-20 - black, blue, brown, furry, gray, green, long, metallic, orange, pink, rectangular, red, rough, round, shiny, smooth, spotted, square, striped, vegetation, violet, wet, white, wooden, yellow Table 5.1: Attribute groups used to build each taxonomy for AWA-10 and ImageNet- 20. These groups are manually defined based on the available attribute labels and their semantic relationships. 100

118 bels and attribute labels. For every image, we have the real-valued attribute presence prediction for each attribute in the vocabulary. That is, for M total attributes, each image has an M-length vector recording the likelihood that each attribute is present in it. Because these attributes are semantically meaningful, we can use them to create a variety of semantic taxonomies. We do this by manually forming subsets of related attributes and then hierarchically clustering the data according to only those (fewer than M) selected attribute dimensions. Each group/subset generates one taxonomy. To generate the attribute-based taxonomies on the AWA-10 dataset, we manually group M=78 of the total attributes provided with the AWA datset as shown in Table 5.1, and perform agglomerative clustering as discussed in the main text to form the semantic hierarchies. For ImageNet-20, we perform agglomerative clustering on all 25 attributes shown in the bottom row of Table 5.1. As the attributes for ImageNet- 20 are binary, we use l 1 -distance when grouping them, and limit the tree height to 6 to avoid having too many branches at the root Baseline methods for comparison We compare our method to three key baselines: Raw feature kernel: an RBF kernel computed on the original image features, with the γ parameter set to the inverse of the mean Euclidean distance d among training instances. 101

119 Raw feature kernel + MKL: MKL combination of multiple such RBF kernels constructed by varying γ, which is a traditional approach to generate base kernels (e.g., [8]). For this baseline, we generate the same number N of base kernels as in the semantic kernel forest, with γ = σ, for σ = d {21 m,...,2 N m }, where m = N. 2 Perturbed semantic kernel tree: a semantic kernel tree trained with taxonomies that have randomly swapped leaves. The first two baselines will show the accuracy attainable using the same image features and basic classification tools (SVM, MKL) as our approach, but lacking the taxonomy insights. The last baseline will test if weakening the semantics in the taxonomy has a negative impact on accuracy. I evaluate several variants of my approach, in order to analyze the impact of each component: Semantic kernel tree + Avg: an equal-weight average of the semantic kernels from one taxonomy. Semantic kernel tree + MKL: the same kernels, but combined with MKL using sparsity regularization only (i.e., µ = 0 in eq. 5.5). Semantic kernel tree + MKL-H: the same as previous, but adding the proposed hierarchical regularization (eq. 5.5). Semantic kernel forest + MKL: semantic forest kernels from multiple taxonomies combined with MKL. 102

120 Semantic kernel forest + MKL-H: the same as previous, but adding our hierarchical regularizer Implementation details For all results, we use 30/30/30 images per class for training/validation/testing, and generate 5 such random splits. We report average multi-class recognition accuracy and standard errors for 95% confidence interval. For single taxonomy results, we report the average over all individual taxonomies. For all methods, the raw image features are bag-of-words histograms obtained on SIFT, provided with the datasets. We reduce their dimensionality to 100 with PCA to speed up the ToM training, following [55]. To train ToM, we sample 400 random constraints and cross-validate the regularization parameters λ, γ {0.1, 1, 10}. For MKL/MKL-H, we use C = 1000 for the C- SVM parameter, and cross-validate the sparsity and hierarchical parameters λ, µ {0, 0.1, 1, 10} Results Quantitative results Table 5.2 shows the multi-class classification accuracy on all three datasets. Our semantic kernel forests approach significantly outperforms all three baselines. It improves accuracy for 9 of the 10 AWA-10 classes, and 16 of the 20 classes in ImageNet-20 (see Figure 5.4). These gains clearly show the impact of injecting semantics into discriminative feature learning. The forests advantage over the individual trees supports our core claim 103

121 AWA-4 AWA-10 ImageNet-20 Raw feature kernel ± ± ± 1.45 Raw feature kernel + MKL ± ± ± 1.50 Perturbed semantic kernel tree N/A ± ± 2.02 Semantic kernel tree + Avg ± ± ± 1.61 Semantic kernel tree + MKL ± ± ± 1.26 Semantic kernel tree + MKL-H ± ± ± 0.70 Semantic kernel forest + MKL ± ± ± 1.14 Semantic kernel forest + MKL-H ± ± ± 1.00 Table 5.2: Multi-class classification accuracy on all datasets, across 5 train/test splits. (The perturbed semantic kernel tree baseline is not applicable for AWA-4, since all possible groupings are present in the taxonomies.) regarding the value of interleaving semantic cues from multiple taxonomies. Further, the proposed hierarchical regularization (MKL-H) outperforms the generic MKL, particularly for the multiple taxonomy forests. I stress that semantic kernel forests success is not simply due to having access to a variety of kernels, as we can see by comparing our method to both the raw feature MKL and perturbed tree results all of which use the same number of kernels. Instead, the advantage is leveraging the implicit discriminative criteria embedded in the external semantic groupings. Interestingly, the perturbed taxonomies show some improvement over the raw feature kernel on AWA-10, but not on ImageNet-20. We attribute this difference to the fact that for fine-grained classes like those in AWA-10 (all animals), almost any grouping of labels may have some semantic meaning, whereas for sparser classes like those in ImageNet-20 (from bridge to acorn), arbitrary perturbations are often meaningless. Thus, the baseline s semantics are weakened more noticeably in 104

122 the latter case. MKL-H has the most impact for the multiple taxonomy forests, and relatively little on the single kernel tree. This makes sense. For a single taxonomy, a single kernel is solely responsible for discriminating a class from the others, making all kernels similarly useful. In contrast, in the forest, two classes are related at multiple different nodes, making it necessary to select out useful views; here, the hierarchical regularizer plays the role of favoring kernels at higher levels, which might have more generalization power due to the training set size and number of classes involved. The per-class and per-taxonomy comparisons in Figure 5.4 further elucidate the advantage of using multiple complementary taxonomies. A single semantic kernel tree often improves accuracy on some classes, but at the expense of reduced accuracy on others. This illustrates that the structure of an individual taxonomy is often suboptimal. For example, the Habitat taxonomy on AWA-10 helps distinguish humpback whale well from the others it branches early from the other animals due to its distinctive oceanic background but it hurts accuracy for giant panda. The WordNet taxonomy does exactly the opposite, improving giant panda via the Biological taxonomy, but hurting humpback whale. The semantic kernel forest takes the best of both through its learned combination. The only cases in which it fails are when the majority of the taxonomies strongly degrade performance, as to be expected given the linear MKL combination (e.g., see the class marimba and rule). 105

123 AWA 10 Accuracy improvement Wordnet (1.73) Appearance (1.00) Behavior (2.53) Habitat (2.27) All (5.07) 2 4 seal h. whale Accuracy improvement P. cat hippo raccoon pig chimp panda leopard rat Imagenet Wordnet (0.73) Visual (1.97) Attributes (2.40) All (4.10) 10 ferriswheel strawberry bridge bathtub featherboa comb lamp basketball button drum sunflower policevan pooltable bonsai daisy r. coaster acorn buckle marimba rule Figure 5.4: Per-class accuracy improvements of each individual taxonomy and the semantic kernel forest ( All ) over the raw feature kernel baseline. Numbers in legends denote mean improvement. Best viewed in color. Further qualitative analysis Figure 5.5 (a-d) shows the confusion matrices for AWA-4 using only the root level kernels. We see how each taxonomy specializes the features, exactly in the manner sketched in the chapter introduction. The combination of all taxonomies achieves the highest accuracy (55.00), better than the maximally performing individual taxonomy (Appear- 106

124 Biological (38.33) Appearance (50.83) Habitat (43.33) All (55.00) dalmatian dalmatian dalmatian dalmatian S. cat S. cat S. cat S. cat leopard leopard leopard leopard wolf wolf wolf wolf dalmatian S. cat leopard wolf dalmatian S. cat leopard wolf dalmatian S. cat leopard wolf dalmatian S. cat leopard (a) Biological (b) Appearance (c) Habitat (d) All wolf Figure 5.5: (a-d): AWA-4 confusion matrices for individual taxonomies (a-c) and the combined taxonomies (d). Y-axis shows true classes; x-axis shows predicted classes. ance, 50.83). chimpanzee giant panda leopard Persian cat pig hippopotamus humpback whale raccoon rat seal l1 only (34.33) chimpanzee giant panda leopard Persian cat pig hippopotamus humpback whale raccoon rat seal l1 + Hierarchical (35.67) procyonid feline even toed aquatic carnivore placental cat/rat hairless ~panda appearance racoon/rat land aquatic predator/prey behavior jungle nonjungle aquatic land habitat procyonid feline even toed aquatic carnivore placental cat/rat hairless ~panda appearance racoon/rat land aquatic predator/prey behavior jungle nonjungle aquatic land habitat (a) MKL (b) MKL-H Figure 5.6: Example β k s to show the characteristics of the two regularizers. Each entry is a learned kernel weight (brighter=higher weight). Y-axis shows object classes; x-axis shows kernel node names. Figure 5.6 shows the learned kernel combination weights β k for each class k in AWA-10, using the two different regularizers. In Figure 5.6 (a), the l 1 regularizer selects a sparse set of useful kernels. For example, the humpback whale drops the kernels belonging to the whole Behavior taxonomy block, and gives the strongest weight to hairless, and habitat. However, by failing to 107

125 select some of the upper-level nodes, it focuses only on the most confusing finegrained problems. In contrast, with the proposed regularization Figure 5.6 (b), we see more emphasis on the upper nodes (e.g., the behavior and placental kernels), which helps accuracy. 5.3 Discussion In this chapter, I proposed a semantic kernel forest approach to learn discriminative visual features that leverage information from multiple semantic taxonomies. The results show that it improves object recognition accuracy, and they give good evidence that committing to a single external knowledge source is insufficient. The key novelty here is that the proposed method tackles the difficult problem of merging complementary information in different semantic views, by first isolating features at each granularity and then assembling the learned subfeature spaces in a single pool, with the sparsity and hierarchical regularization to enable interleaving and enforce a structure among the features. The remaining problems are on how to better combine the kernels, as the current additive kernel combination might not capture the strong similarity in a single view. While the proposed MKL method with the hierarchical regularizers is shown to already significantly improve the classification performance, it could still potentially benefit from using a non-additive, non-linear combination multiple kernel learning method. 108

126 Until now, we have focused on semantic knowledge from attributes and taxonomies, and I have shown how to leverage them in ways different from existing models. In the next chapter, I will show how to exploit analogies, a new type of semantic knowledge for visual recognition, to regularize a discriminative categorization model. 109

127 Chapter 6 Transferring Knowledge between Related Category Pairs with Analogies The attributes and taxonomies covered in earlier chapters provided ways to relate categories, which provided structures in the learned categorization models. However, information provided by both of these semantic knowledge types are limited to pairwise class similarities, defined by sharing or non-sharing of object properties. For example, two categories either share some attributes [58], or two categories in different semantic levels in parentchild relationship compete [55] for exclusive properties at each level. In other words, these pairwise similarity-driven models can only provide information on whether two categories are similar or dissimilar, and higher-order reasoning employed in the human recognition process is limited in these models. In the final component of my thesis, I aim to move beyond per-class semantic relatedness, and exploit higher-order relationships jointly involving multiple classes. Specifically, I propose to model analogies between classes in the form p is to q, as r is to s (or, in shorthand, p : q = r : s) 1. An analogy encodes the relational similarity between two pairs of semantic concepts. By 1 The work introduced in this chapter is published in [57]. 110

128 augmenting labeled data instances with a set of semantic analogies during training, we aim to enrich the learned representation and thereby improve generalization. Analogies can be defined with almost arbitrary abstraction, ranging from is-a relationships (dog : canine = cat : feline), to contextual dependencies (fish : water = bird : sky). To examine analogies most likely to benefit visual learning, we restrict our focus to analogical proportions [74] analogies between pairs of concrete objects in the same semantic universe and with similar abstraction level. Before sketching my approach, I want to first motivate why this form of analogy should offer new information to a learning algorithm. As any standardized test-taker knows, analogies are used to gauge both vocabulary skills and reasoning ability. Notably, the pairs of entities involved in an analogy need not share properties. For example, in the analogy planet : sun = electron : nucleus, the planet and electron do not have anything in common; rather, the relational similarity (orbiter and center) is what makes us recognize the two pairs as parallel in meaning [44]. Furthermore, the common difference exhibited by the two pairs in an analogy may encapsulate a combination of multiple properties and that combination need not have a succinct semantic name. For example, in the analogy leopard : cat = wolf : dog, the common difference relating the two pairs entails multiple low-level concepts; in both, the first class lives in the wild, has fangs, and is more aggressive, etc. Thus, to master analogies, one must not only estimate the similarity of words, but also infer the abstract relationships implied by their pairings. 111

129 Accordingly, we expect analogies to benefit a feature learning algorithm in ways that semantic distance constraints alone cannot. Whereas existing methods inject only vocabulary skills by requiring that semantically related instances be close and semantically unrelated ones be far, our method will also inject reasoning ability by requiring that the common differences implied by analogies be reflected in the learned semantic feature space. Often, the higherorder constraints may connect quite distant sets of categories. The analogies can thus facilitate a form of transfer from class pairs that are more easily discriminated in the original feature space to analogous class pairs that are not. For example, suppose leopard and cat are often confused in the visual space because the training set consists of only close-up images, whereas dog and wolf are easily separable due to their distinct backgrounds. Enforcing the analogy constraint leopard : cat = wolf : dog could make the separation in the first pair clearer, by aligning it with the same hypothetical semantic axis of differences (wild/fanged/aggressive) shared by the second (more distinctive) pair. I propose an Analogy-preserving Semantic Embedding (ASE), which embeds features discriminatively with analogies-based structural regularization. Given a set of analogies involving various object categories, we translate each one into a geometric constraint called an analogical parallelogram. This constraint states that the difference between the first pair of categories should be the same as the that between the second pair, where each category is represented by a (learned) prototype vector in some hypothetical semantic space. 112

130 Visual feature space Analogies leopard:cat = wolf:dog leopard:tiger = horse:zebra Regularization Semantic Embedding Space Figure 6.1: Concept of the analogy-preserving semantic embedding (ASE). I introduce analogical parallelogram constraints to regularize a semantic embedding. By learning from both labeled instances and analogies, the learned embedding space preserves structural similarities between category pairs. See Figure 6.1. We represent the constraints as a novel regularizer that augments a large-margin label embedding. Consequently, we obtain an embedding where examples with the same label are mutually close (and far from differently labeled points) and analogical parallelograms have nearly parallel sides. The learned embedding can be used for recognition, automatic analogy completion, visualization, and potentially other tasks. To use it for recognition, we project a novel image into the learned space, and predict its label based on the nearest category prototype. We further show how to automatically discover and prioritize useful analogies, which is valuable to concentrate 113

131 on constraints that are influential for recognition. Compared to traditional large-margin label embeddings [113, 11], our approach preserves a new form of relational similarity. While the prior methods also map to a space where semantic similarities are preserved, they risk learning spurious associations between features and labels. Our analogy-induced regularizer mitigates such adverse effects by constraining the hypothesis space with structural relations between category pairs, yielding robust models with better generalization. Even constraints not in the axes of visual properties can be helpful, as they shift the focus from brittle incidental correlations to higher-order semantic ties. 6.1 Analogy-preserving Semantic Embedding (ASE) In this section, I will present the analogy-preserving semantic embedding for categorization, which learns to place category embeddings (prototypes) in a low-dimensional semantic space, while also preserving the analogical structure between matched category pairs in the analogies Encoding analogies For each class c Y, u c R M denotes its coordinates in the M- dimensional semantic space. Each u c can be thought of as a prototype for the category; we will explain how the prototypes are optimized jointly with the data projection matrix W in Sec An analogy involves four categories, and we represent the relationship 114

132 with an ordered quadruplet (p, q, r, s) Y Y Y Y. As we focus on analogical proportions [74], the difference between p and q is equated with the difference between r and s. Moreover, the difference between p and r also is equated with the difference between q and s. Analogical proportions naturally induce geometric constraints among the embeddings of the four categories in the semantic space. In particular, the geometry is characterized by a parallelogram; we will show how to exploit this structure in our learning algorithm. Analogy parallelogram We use the vector shift (u q u p ) to represent the difference between the two categories q and p in the semantic space. Note that this difference is directed, that is, u q u p u p u q. The analogical proportion implied by (p, q, r, s) is thus encoded by the following pair of equalities: (6.1) u q u p = u s u r, and u r u p = u s u q. These constraints form a parallelogram in which each vertex is a category, as illustrated in Figure 6.2. Convex regularizer There are several ways of enforcing the analogical proportion constraints in equation 6.1. A natural choice is to exploit the parallel property of opposing sides. Specifically, the normalized inner products between opposing sides are the cosine of their intersection degree, which should be 1 if perfectly parallel. Concretely, for an analogy α = (p, q, r, s), the resulting parallelogram score would be defined as (6.2) S(α) = 1 ( (uq u p ) T (u r u s ) 2 u q u p u r u s + (u r u p ) T ) (u s u q ). u r u p u s u q 115

133 Input space Semantic space = = p:q = r:s Figure 6.2: Geometry of ASE. Analogy constraints for the semantic category embedding: The analogy quadruplet (p, q, r, s) forms a parallelogram in the semantic embedding space, cf. eq. (6.1). Data embedding W: At the same time, when projected onto the semantic space by W, the data point x i from class q should be closer to its semantic category embedding u q, compared to any other category embedding, by a large margin (see dotted circles). While intuitive, maximizing the parallelogram score (or equivalently, minimizing its negative) is computationally inconvenient, since it is not convex in the embeddings u. Thus, we use a relaxed version and compare the sides only in their lengths. Specifically, our regularizer is defined as (6.3) R(α) =1/σ 1 (u q u p ) (u r u s ) /σ 2 (u r u p ) (u s u q ) 2 2, where σ 1 and σ 2 are two scaling constants used to prevent either pair of sides from dominating the other. We simply estimate them as the mean distances between data instances from different classes. R(α) is convex in the embedding coordinates. Moreover, it is straight- 116

134 forward to kernelize as it depends only on the distances (and thus inner products) Automatic discovery of analogies Human knowledge is a natural source for harvesting analogy relationships among categories. However, it is likely expensive to completely rely on human assessment to acquire a sufficient number of analogies for training. To address this issue, we use auxiliary semantic knowledge to identify candidate analogies. In the context of visual object recognition, visual attributes are an appealing form of auxiliary semantic knowledge [65]. Attributes are binary predicates shared among certain visual categories for example, the category panda has the true value for the spotted attribute and the false value for the orange attribute. Supposing we have access to attribute descriptions stating the typical attribute values for each category, we can automatically discover plausible analogies. I next define two strategies to do so. The first is independent of the data instances, while the second exploits the instances to emphasize analogies more likely to lend discriminative information. Attribute-based analogy discovery Our first strategy is to view attributes as a proxy to the embedding coordinates of the visual categories in the semantic space we are trying to learn. In the attribute space, each category is encoded with a binary vector, with bits set to one for attributes the class does possess, 117

135 and bits set to zero for attributes the class does not possess. Note that this is a class-level description we have one binary vector per object class. Imagine that we enumerate all quadruplets of visual categories. For each quadruplet α, we compute its parallelogram score according to equation 6.2, using the categories attribute vectors as coordinates. We then select top-scoring quadruplets as our candidate analogies. Pragmatically, we can only score a subset of all possible analogies for a large number of visual categories. Thus, to ensure good coverage, for each randomly selected pivot category p, we select at most K triplets of other categories, where K is far fewer than the total number of possible ones. We also remove equivalent analogies. For example, (p, q, r, s) is equivalent to (p, r, q, s) or other shift-invariant forms. We will use the highest-scoring analogies to augment the class-labeled data when learning the embedding. We stress that while we discover analogies based on parallelogram scores computed in the space of attribute descriptions, we regularize the learned embedding according to parallelogram scores computed in the learned embedding coordinates (cf. Sec ). Thus, external semantics drive the training analogies, which in turn mold our learned semantic space. Discriminative analogy discovery The process described thus far has two possible issues. First, it does not take the data instances into consideration. While our goal is to find a joint embedding space for both data instances 118

136 and category labels, analogies inferred purely from attributes do not necessarily align the data and mid-level representations they might even lead to conflicting embedding preferences! Secondly, being fully unsupervised, this procedure need not discover analogies directly useful to our classification task. In particular, the extracted candidate analogies are not indicative of whether two categories are easily distinguishable or confused. I address both issues with an intuitive and empirically very effective heuristic. Mindful of our goal (described in the introduction) of improving discrimination for confusable categories by leveraging analogy relationships connecting those confusing categories to easily distinguishable categories, we first use baseline classifiers to estimate the pairwise confusability between categories. This step can be achieved easily with any off-the-shelf multi-way classifier and visual features computed from the training instances. The confusability between two categories p and q is defined in terms of the resulting misclassification error: C pq = 0.5 [ǫ p q + ǫ q p ], where ǫ p q is the rate of misclassifying instances from the category p as the category q, and likewise for ǫ q p. Our next step is to refine the candidate analogies generated above by finding those with unbalanced confusability. Specifically, for each analogy α = (p, q, r, s), we compute its discrimination potential: (6.4) P(α) = log(1 + C pq ) log(1 + C rs ). 119

137 Algorithm 3 Discriminative analogy generation Require: R c Ns, S Ensure: A set of analogies: A 1: Initialize A = φ. 2: while A M do 3: Select random category p {1,...,c} 4: Generate K quadruplets a k = (p,q k,r k,s k ),1 k K 5: Compute P(a k ) according to 6.4, for all k K. 6: Sort {a 1,...,a K } by P(a k ) {a s(1),...,a s(k) }. 7: while A A = φ do 8: Find a k = arg max{c(s,a s(1)),...,c(s,a s(κ) )} (κ K). 9: Set A as the all possible rotations of a k. 10: end while 11: A = A {a k } 12: end while 13: return A This score attains its maximum when C pq and C rs are drastically different that is, if one is 0 and the other is 1. We use this score to re-rank the K candidate analogies generated for each category p. Intuitively, we seek the quadruplet where one pair of categories is easily distinguishable (based on the image data) while the other pair is difficult to differentiate. Precisely by enforcing their analogy relationship, we expect the easy pair to assist discrimination for the difficult one. Algorithm 3 shows the details of this analogy generation considering confusability. To summarize, our automatic discovery of analogies is a two-phase strategy. We first use an auxiliary semantic space to identify a set of candidate analogies where the four categories are highly likely to form a parallelogram. 120

138 Then, we analyze misclassification error patterns of these categories and use the scoring function in equation 6.4 to determine the potential of each analogy in improving classification performance. We describe next how to use the highest-scoring analogies to learn the joint embedding of both features and categories Discriminative learning of the ASE Next I explain how we regularize a discriminative embedding to account for the analogies. Large margin-based discrimination We aim to learn a projection matrix W R M D to map each data instance (image example) x i into the semantic space, giving its M-dimensional coordinates z i = Wx i. 2 The ideal projection matrix W should make z i close to its corresponding label s embedding u yi and distant to all other labels embeddings [113] 3. Specifically, we enforce the large margin constraint for every training instance, (6.5) Wx i u yi Wx i u c ξ ic, c y i where ξ ic 0 is a slack variable for satisfying the separation by the margin of 1. Regularization To jointly embed both features and class labels, we regularize 2 Nonlinear embeddings are possible via kernelization. 3 We use 1 instead of the inter-class dissimilarity as the large margin to maximize class separation. 121

139 so that the class labels in the analogy set A form parallelograms as much as possible. The regularizer is given by (6.6) R total (A) = a ω a R(α a ), which is the weighted sum of the regularization defined in eq. (6.3) for each analogy α a. If using the raw attribute-based analogies, the weight ω a = S(α a ), thus enforcing stricter regularization for category quadruplets whose structure is closer to a perfect analogy. If using discriminatively discovered analogies, the weight is instead ω a = P(α a ), thus prioritizing those that are more discriminative. Additionally, we also constrain the parameters W and all u c with their Frobenius norms: W 2 F and R(u) = c u c u prior c 2 2. In particular, for the class label embeddings, we constrain them to be close to our prior knowledge on their locations u prior c. The prior knowledge could be null such that we set u prior c to zeroes. Or, the class label embeddings could be computed from auxiliary information, for example, the multi-dimensional embedding of class labels where the dissimilarities between labels are measured with tree distances from a taxonomy [113] or attributes. We consider both in the results Numerical optimization Our learning problem is thus cast as the following optimization problem: (6.7) min W,{u c} ξ ic + λr total (A) + µ W F + τr(u) ic 122

140 subject to both the large margin constraints in equation 6.5 and non-negativity constraints on the slack variables ξ ic. The regularization coefficients λ, µ, and τ are determined via cross-validation. The optimization is nonconvex due to the quadratically-formed large margin constraints. We have developed two methods for solving it. Our first method uses stochastic (sub)gradient descent, where we update W and u c according to their sub-gradients computed on a subset of instances. Despite its simplicity, this method works well in practice and scales better to problems with many categories. We also consider a convex relaxation analogous to the procedure in [113]. Briefly, in equation 6.7, we hold {u c } fixed first and solve W in closed-form, W = UQ where the matrix U is composed of {u c } as column vectors. The matrix Q depends only on x i and is constant with respect to U or W. Substituting the solution of W into both the objective function equation 6.7 and the large margin constraints equation 6.5, we can reformulate the optimization in terms of U T U. In particular, the original non-convex large margin constraints in U can be relaxed into convex if we reparameterize U T U as a positive semidefinite matrix V. We then solve V and recover the solutions U and W, respectively. For cases where D is much larger than the number of categories, we expect this variant to optimize faster. The details of the numerical optimization for the semidefinite programming relaxation problem are provided in Algorithm 4. At each step, we update the gradient by ηs t, where η is a general learning rate and s t is a step size 123

141 Algorithm 4 ASE (Convex) Require: training data (x n,y n ), analogical quadruplets A Ensure: Q, w 1: Initialize Q = I, and w by setting each element of w with 1/M 2: U = JX T (XX T + λi) 1 3: X = UX 4: while t < T and Q > ǫ do 5: G ξ t = 1/N N i=1 ( x i x j )( x i x j ) T 6: G mds t = 2 (Q Q mds ) 7: Compute gradient G a t from equation : G t = G ξ t + µgmds t + γg a t 9: Q t+1 = Q t ηs t G t with stepsize s t. 10: end while 11: return V = decomp(q) 12: return W = V U specified by some step size rule. We learn η on the validation set, and set s t according to Polyak s step size rule. 6.2 Results We validate three aspects: 1) the effectiveness of our analogy discovery approach; 2) recognition accuracy when incorporating discovered analogies in learning embeddings; and 3) fill in the blank a Graduate Record Examination (GRE)-style prediction task of filling in the category that would form a valid analogy. Datasets and implementation details We use three datasets created from two public image datasets: Animals with Attributes (AWA), which contains 50 animal classes [65] and ImageNet, which contains general object categories [27]. They were chosen due to their available attribute descriptions and their chal- 124

142 lenging diverse content. From AWA, we create two datasets: AWA-10 of 6, 180 images from 10 classes [65], and the complete 50-class AWA-50 of 30, 475 images. From ImageNet, we use the 50-class ImageNet-50 with annotated attributes [88], totaling 70, 380 images. We use the features provided by the authors, which consist of SIFT and other texture and color descriptors. We use PCA to reduce the feature dimensionality to D = 150 for efficient computation. Additionally, we augment ImageNet-50 with attribute labels for colors, material, habitat, and behaviors (e.g., big, round, feline), yielding 39 and 85 binary attributes for ImageNet and AWA, respectively. We fix K = 10, 000. We use the convex relaxation, since the dimensionality is much greater than the number of classes; accordingly, the semantic space dimensionality M equals the number of categories (10 or 50) Automatic discovery of analogies In real-world settings, acquiring all analogies from manual input may be costly and impractical. Thus, we first examine the analogies discovered by our method (Sec ), which assumes only that attribute-labeled object classes are available. Figure 6.3 displays several examples for AWA-50 and ImageNet-50. Most analogies are intuitive to understand. For example, in the second row of collie:dalmatian = lion:leopard, the categories collie and lion are both furry and brown, while the categories dalmatian and leopard are both spotted and 125

143 AWA-50 Imagenet-50 : = : : = : antelope lion zebra tiger comb button bridge ferriswheel : = : : = : collie dalmatian lion leopard comb marimba macaque gorilla Figure 6.3: Example analogies discovered from attributes. lean. We also see that the analogies can be largely visual (e.g., the third row), an upshot of the many visually relevant attributes offered with the datasets Visual recognition with ASE We compare the classification performance of our Analogy-preserving Semantic Embedding (ASE) to the following baselines, all of which lack analogies: (1) SVM-RBF: Multiclass SVM with RBF kernel. (2) Large margin embedding (LME): The existing technique of [113] without the taxonomy prior regularizer, which is also a special case of our approach where we disable both the attributes prior and analogy regularizers by setting τ = 0 and λ = 0 in eq. (6.7). For this baseline, the class label embeddings are constrained only to satisfy the large margin separation criterion of eq. (6.5); (3) Large margin embedding with attributes prior (LME prior ): This baseline adds the prior regularizer to LME, where we adjust τ for eq. (6.7) via cross-validation. In particular, we use the multi-dimensional scaling (MDS) 126

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur Module 12 Machine Learning 12.1 Instructional Objective The students should understand the concept of learning systems Students should learn about different aspects of a learning system Students should

More information

Lecture 1: Machine Learning Basics

Lecture 1: Machine Learning Basics 1/69 Lecture 1: Machine Learning Basics Ali Harakeh University of Waterloo WAVE Lab ali.harakeh@uwaterloo.ca May 1, 2017 2/69 Overview 1 Learning Algorithms 2 Capacity, Overfitting, and Underfitting 3

More information

Python Machine Learning

Python Machine Learning Python Machine Learning Unlock deeper insights into machine learning with this vital guide to cuttingedge predictive analytics Sebastian Raschka [ PUBLISHING 1 open source I community experience distilled

More information

Assignment 1: Predicting Amazon Review Ratings

Assignment 1: Predicting Amazon Review Ratings Assignment 1: Predicting Amazon Review Ratings 1 Dataset Analysis Richard Park r2park@acsmail.ucsd.edu February 23, 2015 The dataset selected for this assignment comes from the set of Amazon reviews for

More information

Rule Learning With Negation: Issues Regarding Effectiveness

Rule Learning With Negation: Issues Regarding Effectiveness Rule Learning With Negation: Issues Regarding Effectiveness S. Chua, F. Coenen, G. Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX Liverpool, United

More information

(Sub)Gradient Descent

(Sub)Gradient Descent (Sub)Gradient Descent CMSC 422 MARINE CARPUAT marine@cs.umd.edu Figures credit: Piyush Rai Logistics Midterm is on Thursday 3/24 during class time closed book/internet/etc, one page of notes. will include

More information

Lecture 1: Basic Concepts of Machine Learning

Lecture 1: Basic Concepts of Machine Learning Lecture 1: Basic Concepts of Machine Learning Cognitive Systems - Machine Learning Ute Schmid (lecture) Johannes Rabold (practice) Based on slides prepared March 2005 by Maximilian Röglinger, updated 2010

More information

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks

Taxonomy-Regularized Semantic Deep Convolutional Neural Networks Taxonomy-Regularized Semantic Deep Convolutional Neural Networks Wonjoon Goo 1, Juyong Kim 1, Gunhee Kim 1, Sung Ju Hwang 2 1 Computer Science and Engineering, Seoul National University, Seoul, Korea 2

More information

Rule Learning with Negation: Issues Regarding Effectiveness

Rule Learning with Negation: Issues Regarding Effectiveness Rule Learning with Negation: Issues Regarding Effectiveness Stephanie Chua, Frans Coenen, and Grant Malcolm University of Liverpool Department of Computer Science, Ashton Building, Ashton Street, L69 3BX

More information

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS

AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS AGS THE GREAT REVIEW GAME FOR PRE-ALGEBRA (CD) CORRELATED TO CALIFORNIA CONTENT STANDARDS 1 CALIFORNIA CONTENT STANDARDS: Chapter 1 ALGEBRA AND WHOLE NUMBERS Algebra and Functions 1.4 Students use algebraic

More information

Probabilistic Latent Semantic Analysis

Probabilistic Latent Semantic Analysis Probabilistic Latent Semantic Analysis Thomas Hofmann Presentation by Ioannis Pavlopoulos & Andreas Damianou for the course of Data Mining & Exploration 1 Outline Latent Semantic Analysis o Need o Overview

More information

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models

Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Learning Structural Correspondences Across Different Linguistic Domains with Synchronous Neural Language Models Stephan Gouws and GJ van Rooyen MIH Medialab, Stellenbosch University SOUTH AFRICA {stephan,gvrooyen}@ml.sun.ac.za

More information

On-Line Data Analytics

On-Line Data Analytics International Journal of Computer Applications in Engineering Sciences [VOL I, ISSUE III, SEPTEMBER 2011] [ISSN: 2231-4946] On-Line Data Analytics Yugandhar Vemulapalli #, Devarapalli Raghu *, Raja Jacob

More information

Word Segmentation of Off-line Handwritten Documents

Word Segmentation of Off-line Handwritten Documents Word Segmentation of Off-line Handwritten Documents Chen Huang and Sargur N. Srihari {chuang5, srihari}@cedar.buffalo.edu Center of Excellence for Document Analysis and Recognition (CEDAR), Department

More information

arxiv: v1 [cs.cl] 2 Apr 2017

arxiv: v1 [cs.cl] 2 Apr 2017 Word-Alignment-Based Segment-Level Machine Translation Evaluation using Word Embeddings Junki Matsuo and Mamoru Komachi Graduate School of System Design, Tokyo Metropolitan University, Japan matsuo-junki@ed.tmu.ac.jp,

More information

The Good Judgment Project: A large scale test of different methods of combining expert predictions

The Good Judgment Project: A large scale test of different methods of combining expert predictions The Good Judgment Project: A large scale test of different methods of combining expert predictions Lyle Ungar, Barb Mellors, Jon Baron, Phil Tetlock, Jaime Ramos, Sam Swift The University of Pennsylvania

More information

arxiv: v2 [cs.cv] 30 Mar 2017

arxiv: v2 [cs.cv] 30 Mar 2017 Domain Adaptation for Visual Applications: A Comprehensive Survey Gabriela Csurka arxiv:1702.05374v2 [cs.cv] 30 Mar 2017 Abstract The aim of this paper 1 is to give an overview of domain adaptation and

More information

Using Web Searches on Important Words to Create Background Sets for LSI Classification

Using Web Searches on Important Words to Create Background Sets for LSI Classification Using Web Searches on Important Words to Create Background Sets for LSI Classification Sarah Zelikovitz and Marina Kogan College of Staten Island of CUNY 2800 Victory Blvd Staten Island, NY 11314 Abstract

More information

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1

Notes on The Sciences of the Artificial Adapted from a shorter document written for course (Deciding What to Design) 1 Notes on The Sciences of the Artificial Adapted from a shorter document written for course 17-652 (Deciding What to Design) 1 Ali Almossawi December 29, 2005 1 Introduction The Sciences of the Artificial

More information

OCR for Arabic using SIFT Descriptors With Online Failure Prediction

OCR for Arabic using SIFT Descriptors With Online Failure Prediction OCR for Arabic using SIFT Descriptors With Online Failure Prediction Andrey Stolyarenko, Nachum Dershowitz The Blavatnik School of Computer Science Tel Aviv University Tel Aviv, Israel Email: stloyare@tau.ac.il,

More information

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks

Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Predicting Student Attrition in MOOCs using Sentiment Analysis and Neural Networks Devendra Singh Chaplot, Eunhee Rhim, and Jihie Kim Samsung Electronics Co., Ltd. Seoul, South Korea {dev.chaplot,eunhee.rhim,jihie.kim}@samsung.com

More information

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS

OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS OPTIMIZATINON OF TRAINING SETS FOR HEBBIAN-LEARNING- BASED CLASSIFIERS Václav Kocian, Eva Volná, Michal Janošek, Martin Kotyrba University of Ostrava Department of Informatics and Computers Dvořákova 7,

More information

Learning From the Past with Experiment Databases

Learning From the Past with Experiment Databases Learning From the Past with Experiment Databases Joaquin Vanschoren 1, Bernhard Pfahringer 2, and Geoff Holmes 2 1 Computer Science Dept., K.U.Leuven, Leuven, Belgium 2 Computer Science Dept., University

More information

Generative models and adversarial training

Generative models and adversarial training Day 4 Lecture 1 Generative models and adversarial training Kevin McGuinness kevin.mcguinness@dcu.ie Research Fellow Insight Centre for Data Analytics Dublin City University What is a generative model?

More information

Seminar - Organic Computing

Seminar - Organic Computing Seminar - Organic Computing Self-Organisation of OC-Systems Markus Franke 25.01.2006 Typeset by FoilTEX Timetable 1. Overview 2. Characteristics of SO-Systems 3. Concern with Nature 4. Design-Concepts

More information

Diverse Concept-Level Features for Multi-Object Classification

Diverse Concept-Level Features for Multi-Object Classification Diverse Concept-Level Features for Multi-Object Classification Youssef Tamaazousti 12 Hervé Le Borgne 1 Céline Hudelot 2 1 CEA, LIST, Laboratory of Vision and Content Engineering, F-91191 Gif-sur-Yvette,

More information

Abstractions and the Brain

Abstractions and the Brain Abstractions and the Brain Brian D. Josephson Department of Physics, University of Cambridge Cavendish Lab. Madingley Road Cambridge, UK. CB3 OHE bdj10@cam.ac.uk http://www.tcm.phy.cam.ac.uk/~bdj10 ABSTRACT

More information

Australian Journal of Basic and Applied Sciences

Australian Journal of Basic and Applied Sciences AENSI Journals Australian Journal of Basic and Applied Sciences ISSN:1991-8178 Journal home page: www.ajbasweb.com Feature Selection Technique Using Principal Component Analysis For Improving Fuzzy C-Mean

More information

Twitter Sentiment Classification on Sanders Data using Hybrid Approach

Twitter Sentiment Classification on Sanders Data using Hybrid Approach IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 4, Ver. I (July Aug. 2015), PP 118-123 www.iosrjournals.org Twitter Sentiment Classification on Sanders

More information

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System

QuickStroke: An Incremental On-line Chinese Handwriting Recognition System QuickStroke: An Incremental On-line Chinese Handwriting Recognition System Nada P. Matić John C. Platt Λ Tony Wang y Synaptics, Inc. 2381 Bering Drive San Jose, CA 95131, USA Abstract This paper presents

More information

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT

WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT WE GAVE A LAWYER BASIC MATH SKILLS, AND YOU WON T BELIEVE WHAT HAPPENED NEXT PRACTICAL APPLICATIONS OF RANDOM SAMPLING IN ediscovery By Matthew Verga, J.D. INTRODUCTION Anyone who spends ample time working

More information

Comment-based Multi-View Clustering of Web 2.0 Items

Comment-based Multi-View Clustering of Web 2.0 Items Comment-based Multi-View Clustering of Web 2.0 Items Xiangnan He 1 Min-Yen Kan 1 Peichu Xie 2 Xiao Chen 3 1 School of Computing, National University of Singapore 2 Department of Mathematics, National University

More information

Truth Inference in Crowdsourcing: Is the Problem Solved?

Truth Inference in Crowdsourcing: Is the Problem Solved? Truth Inference in Crowdsourcing: Is the Problem Solved? Yudian Zheng, Guoliang Li #, Yuanbing Li #, Caihua Shan, Reynold Cheng # Department of Computer Science, Tsinghua University Department of Computer

More information

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus

Language Acquisition Fall 2010/Winter Lexical Categories. Afra Alishahi, Heiner Drenhaus Language Acquisition Fall 2010/Winter 2011 Lexical Categories Afra Alishahi, Heiner Drenhaus Computational Linguistics and Phonetics Saarland University Children s Sensitivity to Lexical Categories Look,

More information

Visual CP Representation of Knowledge

Visual CP Representation of Knowledge Visual CP Representation of Knowledge Heather D. Pfeiffer and Roger T. Hartley Department of Computer Science New Mexico State University Las Cruces, NM 88003-8001, USA email: hdp@cs.nmsu.edu and rth@cs.nmsu.edu

More information

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17.

Semi-supervised methods of text processing, and an application to medical concept extraction. Yacine Jernite Text-as-Data series September 17. Semi-supervised methods of text processing, and an application to medical concept extraction Yacine Jernite Text-as-Data series September 17. 2015 What do we want from text? 1. Extract information 2. Link

More information

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler

Machine Learning and Data Mining. Ensembles of Learners. Prof. Alexander Ihler Machine Learning and Data Mining Ensembles of Learners Prof. Alexander Ihler Ensemble methods Why learn one classifier when you can learn many? Ensemble: combine many predictors (Weighted) combina

More information

University of Groningen. Systemen, planning, netwerken Bosman, Aart

University of Groningen. Systemen, planning, netwerken Bosman, Aart University of Groningen Systemen, planning, netwerken Bosman, Aart IMPORTANT NOTE: You are advised to consult the publisher's version (publisher's PDF) if you wish to cite from it. Please check the document

More information

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition

Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Introduction to Ensemble Learning Featuring Successes in the Netflix Prize Competition Todd Holloway Two Lecture Series for B551 November 20 & 27, 2007 Indiana University Outline Introduction Bias and

More information

Attributed Social Network Embedding

Attributed Social Network Embedding JOURNAL OF LATEX CLASS FILES, VOL. 14, NO. 8, MAY 2017 1 Attributed Social Network Embedding arxiv:1705.04969v1 [cs.si] 14 May 2017 Lizi Liao, Xiangnan He, Hanwang Zhang, and Tat-Seng Chua Abstract Embedding

More information

The Strong Minimalist Thesis and Bounded Optimality

The Strong Minimalist Thesis and Bounded Optimality The Strong Minimalist Thesis and Bounded Optimality DRAFT-IN-PROGRESS; SEND COMMENTS TO RICKL@UMICH.EDU Richard L. Lewis Department of Psychology University of Michigan 27 March 2010 1 Purpose of this

More information

Ontologies vs. classification systems

Ontologies vs. classification systems Ontologies vs. classification systems Bodil Nistrup Madsen Copenhagen Business School Copenhagen, Denmark bnm.isv@cbs.dk Hanne Erdman Thomsen Copenhagen Business School Copenhagen, Denmark het.isv@cbs.dk

More information

AQUA: An Ontology-Driven Question Answering System

AQUA: An Ontology-Driven Question Answering System AQUA: An Ontology-Driven Question Answering System Maria Vargas-Vera, Enrico Motta and John Domingue Knowledge Media Institute (KMI) The Open University, Walton Hall, Milton Keynes, MK7 6AA, United Kingdom.

More information

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks

System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks System Implementation for SemEval-2017 Task 4 Subtask A Based on Interpolated Deep Neural Networks 1 Tzu-Hsuan Yang, 2 Tzu-Hsuan Tseng, and 3 Chia-Ping Chen Department of Computer Science and Engineering

More information

Modeling user preferences and norms in context-aware systems

Modeling user preferences and norms in context-aware systems Modeling user preferences and norms in context-aware systems Jonas Nilsson, Cecilia Lindmark Jonas Nilsson, Cecilia Lindmark VT 2016 Bachelor's thesis for Computer Science, 15 hp Supervisor: Juan Carlos

More information

How to Judge the Quality of an Objective Classroom Test

How to Judge the Quality of an Objective Classroom Test How to Judge the Quality of an Objective Classroom Test Technical Bulletin #6 Evaluation and Examination Service The University of Iowa (319) 335-0356 HOW TO JUDGE THE QUALITY OF AN OBJECTIVE CLASSROOM

More information

An Introduction to Simio for Beginners

An Introduction to Simio for Beginners An Introduction to Simio for Beginners C. Dennis Pegden, Ph.D. This white paper is intended to introduce Simio to a user new to simulation. It is intended for the manufacturing engineer, hospital quality

More information

Extending Place Value with Whole Numbers to 1,000,000

Extending Place Value with Whole Numbers to 1,000,000 Grade 4 Mathematics, Quarter 1, Unit 1.1 Extending Place Value with Whole Numbers to 1,000,000 Overview Number of Instructional Days: 10 (1 day = 45 minutes) Content to Be Learned Recognize that a digit

More information

SARDNET: A Self-Organizing Feature Map for Sequences

SARDNET: A Self-Organizing Feature Map for Sequences SARDNET: A Self-Organizing Feature Map for Sequences Daniel L. James and Risto Miikkulainen Department of Computer Sciences The University of Texas at Austin Austin, TX 78712 dljames,risto~cs.utexas.edu

More information

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown

Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology. Michael L. Connell University of Houston - Downtown Digital Fabrication and Aunt Sarah: Enabling Quadratic Explorations via Technology Michael L. Connell University of Houston - Downtown Sergei Abramovich State University of New York at Potsdam Introduction

More information

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS

THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS THE PENNSYLVANIA STATE UNIVERSITY SCHREYER HONORS COLLEGE DEPARTMENT OF MATHEMATICS ASSESSING THE EFFECTIVENESS OF MULTIPLE CHOICE MATH TESTS ELIZABETH ANNE SOMERS Spring 2011 A thesis submitted in partial

More information

Multi-Lingual Text Leveling

Multi-Lingual Text Leveling Multi-Lingual Text Leveling Salim Roukos, Jerome Quin, and Todd Ward IBM T. J. Watson Research Center, Yorktown Heights, NY 10598 {roukos,jlquinn,tward}@us.ibm.com Abstract. Determining the language proficiency

More information

A Case Study: News Classification Based on Term Frequency

A Case Study: News Classification Based on Term Frequency A Case Study: News Classification Based on Term Frequency Petr Kroha Faculty of Computer Science University of Technology 09107 Chemnitz Germany kroha@informatik.tu-chemnitz.de Ricardo Baeza-Yates Center

More information

1 3-5 = Subtraction - a binary operation

1 3-5 = Subtraction - a binary operation High School StuDEnts ConcEPtions of the Minus Sign Lisa L. Lamb, Jessica Pierson Bishop, and Randolph A. Philipp, Bonnie P Schappelle, Ian Whitacre, and Mindy Lewis - describe their research with students

More information

Speech Recognition at ICSI: Broadcast News and beyond

Speech Recognition at ICSI: Broadcast News and beyond Speech Recognition at ICSI: Broadcast News and beyond Dan Ellis International Computer Science Institute, Berkeley CA Outline 1 2 3 The DARPA Broadcast News task Aspects of ICSI

More information

Radius STEM Readiness TM

Radius STEM Readiness TM Curriculum Guide Radius STEM Readiness TM While today s teens are surrounded by technology, we face a stark and imminent shortage of graduates pursuing careers in Science, Technology, Engineering, and

More information

The University of Amsterdam s Concept Detection System at ImageCLEF 2011

The University of Amsterdam s Concept Detection System at ImageCLEF 2011 The University of Amsterdam s Concept Detection System at ImageCLEF 2011 Koen E. A. van de Sande and Cees G. M. Snoek Intelligent Systems Lab Amsterdam, University of Amsterdam Software available from:

More information

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge

Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Innov High Educ (2009) 34:93 103 DOI 10.1007/s10755-009-9095-2 Maximizing Learning Through Course Alignment and Experience with Different Types of Knowledge Phyllis Blumberg Published online: 3 February

More information

Software Maintenance

Software Maintenance 1 What is Software Maintenance? Software Maintenance is a very broad activity that includes error corrections, enhancements of capabilities, deletion of obsolete capabilities, and optimization. 2 Categories

More information

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda

Content Language Objectives (CLOs) August 2012, H. Butts & G. De Anda Content Language Objectives (CLOs) Outcomes Identify the evolution of the CLO Identify the components of the CLO Understand how the CLO helps provide all students the opportunity to access the rigor of

More information

Full text of O L O W Science As Inquiry conference. Science as Inquiry

Full text of O L O W Science As Inquiry conference. Science as Inquiry Page 1 of 5 Full text of O L O W Science As Inquiry conference Reception Meeting Room Resources Oceanside Unifying Concepts and Processes Science As Inquiry Physical Science Life Science Earth & Space

More information

Word learning as Bayesian inference

Word learning as Bayesian inference Word learning as Bayesian inference Joshua B. Tenenbaum Department of Psychology Stanford University jbt@psych.stanford.edu Fei Xu Department of Psychology Northeastern University fxu@neu.edu Abstract

More information

Learning Methods for Fuzzy Systems

Learning Methods for Fuzzy Systems Learning Methods for Fuzzy Systems Rudolf Kruse and Andreas Nürnberger Department of Computer Science, University of Magdeburg Universitätsplatz, D-396 Magdeburg, Germany Phone : +49.39.67.876, Fax : +49.39.67.8

More information

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics

Machine Learning from Garden Path Sentences: The Application of Computational Linguistics Machine Learning from Garden Path Sentences: The Application of Computational Linguistics http://dx.doi.org/10.3991/ijet.v9i6.4109 J.L. Du 1, P.F. Yu 1 and M.L. Li 2 1 Guangdong University of Foreign Studies,

More information

A Neural Network GUI Tested on Text-To-Phoneme Mapping

A Neural Network GUI Tested on Text-To-Phoneme Mapping A Neural Network GUI Tested on Text-To-Phoneme Mapping MAARTEN TROMPPER Universiteit Utrecht m.f.a.trompper@students.uu.nl Abstract Text-to-phoneme (T2P) mapping is a necessary step in any speech synthesis

More information

Copyright Corwin 2015

Copyright Corwin 2015 2 Defining Essential Learnings How do I find clarity in a sea of standards? For students truly to be able to take responsibility for their learning, both teacher and students need to be very clear about

More information

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL

UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL UNIVERSITY OF CALIFORNIA SANTA CRUZ TOWARDS A UNIVERSAL PARAMETRIC PLAYER MODEL A thesis submitted in partial satisfaction of the requirements for the degree of DOCTOR OF PHILOSOPHY in COMPUTER SCIENCE

More information

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT

Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the SAT The Journal of Technology, Learning, and Assessment Volume 6, Number 6 February 2008 Using the Attribute Hierarchy Method to Make Diagnostic Inferences about Examinees Cognitive Skills in Algebra on the

More information

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology

Essentials of Ability Testing. Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology Essentials of Ability Testing Joni Lakin Assistant Professor Educational Foundations, Leadership, and Technology Basic Topics Why do we administer ability tests? What do ability tests measure? How are

More information

Knowledge Transfer in Deep Convolutional Neural Nets

Knowledge Transfer in Deep Convolutional Neural Nets Knowledge Transfer in Deep Convolutional Neural Nets Steven Gutstein, Olac Fuentes and Eric Freudenthal Computer Science Department University of Texas at El Paso El Paso, Texas, 79968, U.S.A. Abstract

More information

Diagnostic Test. Middle School Mathematics

Diagnostic Test. Middle School Mathematics Diagnostic Test Middle School Mathematics Copyright 2010 XAMonline, Inc. All rights reserved. No part of the material protected by this copyright notice may be reproduced or utilized in any form or by

More information

Unpacking a Standard: Making Dinner with Student Differences in Mind

Unpacking a Standard: Making Dinner with Student Differences in Mind Unpacking a Standard: Making Dinner with Student Differences in Mind Analyze how particular elements of a story or drama interact (e.g., how setting shapes the characters or plot). Grade 7 Reading Standards

More information

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model

Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Unsupervised Learning of Word Semantic Embedding using the Deep Structured Semantic Model Xinying Song, Xiaodong He, Jianfeng Gao, Li Deng Microsoft Research, One Microsoft Way, Redmond, WA 98052, U.S.A.

More information

Axiom 2013 Team Description Paper

Axiom 2013 Team Description Paper Axiom 2013 Team Description Paper Mohammad Ghazanfari, S Omid Shirkhorshidi, Farbod Samsamipour, Hossein Rahmatizadeh Zagheli, Mohammad Mahdavi, Payam Mohajeri, S Abbas Alamolhoda Robotics Scientific Association

More information

Learning Disability Functional Capacity Evaluation. Dear Doctor,

Learning Disability Functional Capacity Evaluation. Dear Doctor, Dear Doctor, I have been asked to formulate a vocational opinion regarding NAME s employability in light of his/her learning disability. To assist me with this evaluation I would appreciate if you can

More information

CS Machine Learning

CS Machine Learning CS 478 - Machine Learning Projects Data Representation Basic testing and evaluation schemes CS 478 Data and Testing 1 Programming Issues l Program in any platform you want l Realize that you will be doing

More information

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach

Deep search. Enhancing a search bar using machine learning. Ilgün Ilgün & Cedric Reichenbach #BaselOne7 Deep search Enhancing a search bar using machine learning Ilgün Ilgün & Cedric Reichenbach We are not researchers Outline I. Periscope: A search tool II. Goals III. Deep learning IV. Applying

More information

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1

Activities, Exercises, Assignments Copyright 2009 Cem Kaner 1 Patterns of activities, iti exercises and assignments Workshop on Teaching Software Testing January 31, 2009 Cem Kaner, J.D., Ph.D. kaner@kaner.com Professor of Software Engineering Florida Institute of

More information

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski

Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski Training a Neural Network to Answer 8th Grade Science Questions Steven Hewitt, An Ju, Katherine Stasaski Problem Statement and Background Given a collection of 8th grade science questions, possible answer

More information

Computerized Adaptive Psychological Testing A Personalisation Perspective

Computerized Adaptive Psychological Testing A Personalisation Perspective Psychology and the internet: An European Perspective Computerized Adaptive Psychological Testing A Personalisation Perspective Mykola Pechenizkiy mpechen@cc.jyu.fi Introduction Mixed Model of IRT and ES

More information

Guide to Teaching Computer Science

Guide to Teaching Computer Science Guide to Teaching Computer Science Orit Hazzan Tami Lapidot Noa Ragonis Guide to Teaching Computer Science An Activity-Based Approach Dr. Orit Hazzan Associate Professor Technion - Israel Institute of

More information

learning collegiate assessment]

learning collegiate assessment] [ collegiate learning assessment] INSTITUTIONAL REPORT 2005 2006 Kalamazoo College council for aid to education 215 lexington avenue floor 21 new york new york 10016-6023 p 212.217.0700 f 212.661.9766

More information

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION

THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION THE ROLE OF TOOL AND TEACHER MEDIATIONS IN THE CONSTRUCTION OF MEANINGS FOR REFLECTION Lulu Healy Programa de Estudos Pós-Graduados em Educação Matemática, PUC, São Paulo ABSTRACT This article reports

More information

Innovative Methods for Teaching Engineering Courses

Innovative Methods for Teaching Engineering Courses Innovative Methods for Teaching Engineering Courses KR Chowdhary Former Professor & Head Department of Computer Science and Engineering MBM Engineering College, Jodhpur Present: Director, JIETSETG Email:

More information

CSL465/603 - Machine Learning

CSL465/603 - Machine Learning CSL465/603 - Machine Learning Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Introduction CSL465/603 - Machine Learning 1 Administrative Trivia Course Structure 3-0-2 Lecture Timings Monday 9.55-10.45am

More information

A Case-Based Approach To Imitation Learning in Robotic Agents

A Case-Based Approach To Imitation Learning in Robotic Agents A Case-Based Approach To Imitation Learning in Robotic Agents Tesca Fitzgerald, Ashok Goel School of Interactive Computing Georgia Institute of Technology, Atlanta, GA 30332, USA {tesca.fitzgerald,goel}@cc.gatech.edu

More information

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT

SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT SETTING STANDARDS FOR CRITERION- REFERENCED MEASUREMENT By: Dr. MAHMOUD M. GHANDOUR QATAR UNIVERSITY Improving human resources is the responsibility of the educational system in many societies. The outputs

More information

A student diagnosing and evaluation system for laboratory-based academic exercises

A student diagnosing and evaluation system for laboratory-based academic exercises A student diagnosing and evaluation system for laboratory-based academic exercises Maria Samarakou, Emmanouil Fylladitakis and Pantelis Prentakis Technological Educational Institute (T.E.I.) of Athens

More information

Using dialogue context to improve parsing performance in dialogue systems

Using dialogue context to improve parsing performance in dialogue systems Using dialogue context to improve parsing performance in dialogue systems Ivan Meza-Ruiz and Oliver Lemon School of Informatics, Edinburgh University 2 Buccleuch Place, Edinburgh I.V.Meza-Ruiz@sms.ed.ac.uk,

More information

Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving

Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving Inquiry Learning Methodologies and the Disposition to Energy Systems Problem Solving Minha R. Ha York University minhareo@yorku.ca Shinya Nagasaki McMaster University nagasas@mcmaster.ca Justin Riddoch

More information

Reinforcement Learning by Comparing Immediate Reward

Reinforcement Learning by Comparing Immediate Reward Reinforcement Learning by Comparing Immediate Reward Punit Pandey DeepshikhaPandey Dr. Shishir Kumar Abstract This paper introduces an approach to Reinforcement Learning Algorithm by comparing their immediate

More information

Transfer Learning Action Models by Measuring the Similarity of Different Domains

Transfer Learning Action Models by Measuring the Similarity of Different Domains Transfer Learning Action Models by Measuring the Similarity of Different Domains Hankui Zhuo 1, Qiang Yang 2, and Lei Li 1 1 Software Research Institute, Sun Yat-sen University, Guangzhou, China. zhuohank@gmail.com,lnslilei@mail.sysu.edu.cn

More information

Team Formation for Generalized Tasks in Expertise Social Networks

Team Formation for Generalized Tasks in Expertise Social Networks IEEE International Conference on Social Computing / IEEE International Conference on Privacy, Security, Risk and Trust Team Formation for Generalized Tasks in Expertise Social Networks Cheng-Te Li Graduate

More information

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard

Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA. 1. Introduction. Alta de Waal, Jacobus Venter and Etienne Barnard Chapter 10 APPLYING TOPIC MODELING TO FORENSIC DATA Alta de Waal, Jacobus Venter and Etienne Barnard Abstract Most actionable evidence is identified during the analysis phase of digital forensic investigations.

More information

Chinese Language Parsing with Maximum-Entropy-Inspired Parser

Chinese Language Parsing with Maximum-Entropy-Inspired Parser Chinese Language Parsing with Maximum-Entropy-Inspired Parser Heng Lian Brown University Abstract The Chinese language has many special characteristics that make parsing difficult. The performance of state-of-the-art

More information

Mandarin Lexical Tone Recognition: The Gating Paradigm

Mandarin Lexical Tone Recognition: The Gating Paradigm Kansas Working Papers in Linguistics, Vol. 0 (008), p. 8 Abstract Mandarin Lexical Tone Recognition: The Gating Paradigm Yuwen Lai and Jie Zhang University of Kansas Research on spoken word recognition

More information

1. READING ENGAGEMENT 2. ORAL READING FLUENCY

1. READING ENGAGEMENT 2. ORAL READING FLUENCY Teacher Observation Guide Animals Can Help Level 28, Page 1 Name/Date Teacher/Grade Scores: Reading Engagement /8 Oral Reading Fluency /16 Comprehension /28 Independent Range: 6 7 11 14 19 25 Book Selection

More information

Statewide Framework Document for:

Statewide Framework Document for: Statewide Framework Document for: 270301 Standards may be added to this document prior to submission, but may not be removed from the framework to meet state credit equivalency requirements. Performance

More information

A Pipelined Approach for Iterative Software Process Model

A Pipelined Approach for Iterative Software Process Model A Pipelined Approach for Iterative Software Process Model Ms.Prasanthi E R, Ms.Aparna Rathi, Ms.Vardhani J P, Mr.Vivek Krishna Electronics and Radar Development Establishment C V Raman Nagar, Bangalore-560093,

More information

Knowledge-Based - Systems

Knowledge-Based - Systems Knowledge-Based - Systems ; Rajendra Arvind Akerkar Chairman, Technomathematics Research Foundation and Senior Researcher, Western Norway Research institute Priti Srinivas Sajja Sardar Patel University

More information