Skip to main content
Log in

On the principles of Parsimony and Self-consistency for the emergence of intelligence

论智能起源中的简约与自洽原则

  • Position Paper
  • Published:
Frontiers of Information Technology & Electronic Engineering Aims and scope Submit manuscript

Abstract

Ten years into the revival of deep networks and artificial intelligence, we propose a theoretical framework that sheds light on understanding deep networks within a bigger picture of intelligence in general. We introduce two fundamental principles, Parsimony and Self-consistency, which address two fundamental questions regarding intelligence: what to learn and how to learn, respectively. We believe the two principles serve as the cornerstone for the emergence of intelligence, artificial or natural. While they have rich classical roots, we argue that they can be stated anew in entirely measurable and computable ways. More specifically, the two principles lead to an effective and efficient computational framework, compressive closed-loop transcription, which unifies and explains the evolution of modern deep networks and most practices of artificial intelligence. While we use mainly visual data modeling as an example, we believe the two principles will unify understanding of broad families of autonomous intelligent systems and provide a framework for understanding the brain.

摘要

深度学习重振人工智能十年后的今天, 我们提出一个理论框架来帮助理解深度神经网络在整个智能系统里面扮演的角色. 我们引入两个基本原则: 简约与自洽; 分别解释智能系统要学习什么以及如何学习. 我们认为这两个原则是人工智能和自然智能之所以产生和发展的基石. 虽然这两个原则的雏形早已出现在前人的经典工作里, 但是我们对这些原则的重新表述使得它们变得可以精准度量与计算. 确切地说, 简约与自洽这两个原则能自然地演绎出一个高效计算框架: 压缩闭环转录. 这个框架统一并解释了现代深度神经网络以及众多人工智能实践的演变和进化. 尽管本文主要用视觉数据建模作为例子, 我们相信这两个原则将会有助于统一对各种自动智能系统的理解, 并且提供一个帮助理解大脑工作机理的框架.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

Download references

Afterword and acknowledgements

Although the research of Yi and Harry focuses more on computer vision and computer graphics, they both happened to major in control and automation as undergraduate students. They started their collaboration many years ago at Microsoft Research Asia (MSRA) with a compression-based approach to data clustering and classification (Ma et al., 2007; Wright et al., 2007). In the past two years, they have had frequent discussions and debates about understanding (deep) learning and (artificial) intelligence. Their shared interests in intelligence have brought all these fundamental ideas together and led to the recent collaboration on closed-loop transcription (Dai et al., 2022), and eventually to many of the views shared in this paper. Doris is deeply interested in whether and how the brain implements generative models for visual perception, and her group has been having intense discussions with Yi on this topic since her moving to UC Berkeley a year ago.

The idea of writing this position paper is partly motivated by recent stimulating discussions among a group of researchers with very diverse backgrounds in artificial intelligence, applied mathematics, optimization, and neuroscience: Professors John Wright and Stefano Fusi of Columbia University, Professors Yann LeCun and Rob Fergus of New York University, Dr. Xin Tong of MSRA. We realize that these perspectives might be interesting to broader scientific and engineering communities.

Some of the thoughts presented about integrating pieces of the puzzle for intelligent systems can be traced back to an advanced robotics course that Yi had led and organized jointly with Professors Jitendra Malik, Shankar Sastry, and Claire Tomlin as Berkeley EECS290-005: the Integration of Perception, Learning, and Control in Spring 2021. The need for an integrated view or a “unite and build” approach seems to be a topic that is drawing increasing interest and importance for the study of artificial intelligence.

We would also like to thank many of our former and current students who, against extraordinary odds, have worked on projects under this new framework in the past several years when some of the ideas were still in their infancy and seemed not in accordance with the mainstream, including Xili Dai, Yaodong Yu, Peter Tong, Ryan Chan, Chong You, Ziyang Wu, Christina Baek, Druv Pai, Brent Yi, Michael Psenska, and others. Many of the technical evidence and figures used in this position paper are conveniently borrowed from their recent research results.

Author information

Authors and Affiliations

Authors

Contributions

The technical evidence of this paper is mainly based on research results from Yi MA’s group in recent years, some of which are in collaboration with Heung-Yeung SHUM. Yi MA drafted the original manuscript. Heung-Yeung SHUM and Doris TSAO helped reorganize and revise the paper significantly. Particularly, Doris TSAO helped establish good connection of the proposed framework with studies and implications in neuroscience.

Corresponding author

Correspondence to Yi Ma  (马毅).

Ethics declarations

Yi MA, Doris TSAO, and Heung-Yeung SHUM declare that they have no conflict of interest.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ma, Y., Tsao, D. & Shum, HY. On the principles of Parsimony and Self-consistency for the emergence of intelligence. Front Inform Technol Electron Eng 23, 1298–1323 (2022). https://doi.org/10.1631/FITEE.2200297

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1631/FITEE.2200297

Key words

CLC number

关键词

Navigation