[ICML23] On the Complexity of Bayesian Generalization

Abstract

We examine concept generalization at a large scale in the natural visual spectrum. Established computational modes (i.e., rule-based or similarity-based) are primarily studied isolated, focusing on confined and abstract problem spaces. In this work, we study these two modes when the problem space scales up and when the complexity of concepts becomes diverse. At the representational level, we investigate how the complexity varies when a visual concept is mapped to the representation space. Prior literature has shown that two types of complexities build an inverted-U relation. Leveraging Representativeness of Attribute (RoA), we computationally confirm: Models use attributes with high RoA to describe visual concepts, and the description length falls in an inverted-U relation with the increment in visual complexity. At the computational level, we examine how the complexity of representation affects the shift between the rule- and similarity-based generalization. We hypothesize that category-conditioned visual modeling estimates the co-occurrence frequency between visual and categorical attributes, thus potentially serving as the prior for the natural visual world. Experimental results show that representations with relatively high subjective complexity outperform those with low subjective complexity in rule-based generalization, while the trend is the opposite in similarity-based generalization.

Publication
In Proceedings of the International Conference on Machine Learning
Manjie Xu
Manjie Xu
MSc. '22, co-advised with Prof. Wei Liang

I focus on making my agents not too silly.

Song-Chun Zhu
Song-Chun Zhu
Chair Professor
Ying Nian Wu
Ying Nian Wu
Professor
Yixin Zhu
Yixin Zhu
Assistant Professor

I build humanlike AI.

Related