Skip to main content

Minsky, Chomsky and Deep Nets

  • Conference paper
  • First Online:
Text, Speech, and Dialogue (TSD 2018)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 11107))

Included in the following conference series:

  • 1533 Accesses

Abstract

When Minsky and Chomsky were at Harvard in the 1950s, they started out their careers questioning a number of machine learning methods that have since regained popularity. Minsky’s Perceptrons was a reaction to neural nets and Chomsky’s Syntactic Structures was a reaction to ngram language models. Many of their objections are being ignored and forgotten (perhaps for good reasons, and perhaps not). While their arguments may sound negative, I believe there is a more constructive way to think about their efforts; they were both attempting to organize computational tasks into larger frameworks such as what is now known as the Chomsky Hierarchy and algorithmic complexity. Section 5 will propose an organizing framework for deep nets. Deep nets are probably not the solution to all the world’s problems. They don’t do the impossible (solve the halting problem), and they probably aren’t great at many tasks such as sorting large vectors and multiplying large matrices. In practice, deep nets have produced extremely exciting results in vision and speech, though other tasks may be more challenging for deep nets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    http://money.cnn.com/2017/05/31/technology/mary-meeker-internet-trends/index.html.

  2. 2.

    http://www.dailymail.co.uk/news/article-4918342/China-installs-20-million-AI-equipped-street-cameras.html.

  3. 3.

    http://www.thedrum.com/news/2016/07/03/marketing-moments-11-american-express-dont-leave-home-without-it.

  4. 4.

    https://www.brainyquote.com/quotes/sam_levenson_100238.

  5. 5.

    http://www.earningmyturns.org/2017/06/a-computational-linguistic-farce-in.html.

  6. 6.

    https://en.wikiquote.org/wiki/Yogi_Berra.

  7. 7.

    https://quoteinvestigator.com/2013/10/20/no-predict/.

  8. 8.

    http://neuralnetworksanddeeplearning.com/chap4.html.

  9. 9.

    https://en.wikipedia.org/wiki/Universal_approximation_theorem.

  10. 10.

    https://en.wikipedia.org/wiki/AI_winter.

  11. 11.

    https://stevenpdennis.com/2015/07/10/a-bunch-of-little-kids-running-toward-a-soccer-ball/.

  12. 12.

    https://www.nytimes.com/2011/12/06/science/george-dyson-looking-backward-to-put-new-technology-in-focus.html.

  13. 13.

    https://www.wired.com/2014/02/halting-problem/.

  14. 14.

    http://www.lel.ed.ac.uk/~gpullum/loopsnoop.html.

  15. 15.

    https://en.wikipedia.org/wiki/Asymptotic_equipartition_property.

  16. 16.

    http://www.cs.toronto.edu/~rgrosse/csc321/lec9.pdf.

  17. 17.

    https://blog.openai.com/requests-for-research-2/.

  18. 18.

    https://en.wikipedia.org/wiki/Thinking_Machines_Corporation.

  19. 19.

    https://developer.nvidia.com/gpugems/GPUGems3/gpugems3_ch39.html.

  20. 20.

    http://www.personal.psu.edu/ago109/giles-ororbia-rnn-icml2016.pdf.

  21. 21.

    http://www.lrec-conf.org/lrec2004/doc/jelinek.pdf.

  22. 22.

    http://techtalks.tv/talks/closing-session/60532/ (at 6:07 min).

  23. 23.

    https://en.wikipedia.org/wiki/Geoffrey_West.

  24. 24.

    http://cs229.stanford.edu/notes/cs229-notes4.pdf.

References

  1. Church, K.: Emerging trends: artificial intelligence, China and my new job at Baidu. J. Nat. Lang. Eng. (to appear). University Press, Cambridge

    Google Scholar 

  2. Minsky, M., Papert, S.: Perceptrons. MIT Press, Cambridge (1969)

    MATH  Google Scholar 

  3. Chomsky, N.: Syntactic Structures. Mouton & Co. (1957). https://archive.org/details/NoamChomskySyntcaticStructures

  4. Shannon, C.: A mathematical theory of communication. Bell Syst. Tech. J. 27, 379–423, 623–656 (1948). http://math.harvard.edu/~ctm/home/text/others/shannon/entropy/entropy.pdf

    Article  MathSciNet  Google Scholar 

  5. Shannon, C.: Prediction and entropy of printed English. Bell Syst. Tech. J. 30(1), 50–64 (1951). https://www.princeton.edu/~wbialek/rome/refs/shannon51.pdf

    Article  Google Scholar 

  6. Zipf, G.: Human Behavior and the Principle of Least Effort: An Introduction to Human Ecology. Addison-Wesley, Boston (1949)

    Google Scholar 

  7. Harris, Z.: Distributional structure. Word 10(2–3), 146–162 (1954)

    Article  Google Scholar 

  8. Firth, J.: A synopsis of linguistic theory, 1930–1955. Stud. Linguist. Anal. Basil Blackwell (1957). http://annabellelukin.edublogs.org/files/2013/08/Firth-JR-1962-A-Synopsis-of-Linguistic-Theory-wfihi5.pdf

  9. Church, K.: A pendulum swung too far. Linguist. Issues Lang. Technol. 6(6), 1–27 (2011)

    Google Scholar 

  10. Turing, A.: On computable numbers, with an application to the Entscheidungsproblem. In: Proceedings of the London Mathematical Society, vol. 2, no. 1, pp. 230–265. Wiley Online Library (1937). http://www.turingarchive.org/browse.php/b/12

    Article  MathSciNet  Google Scholar 

  11. Hillis, W.: The Connection Machine. MIT Press, Cambridge (1989)

    Google Scholar 

  12. Blelloch, G., Leiserson, C., Maggs, B., Plaxton, C., Smith, S., Marco, C.: A comparison of sorting algorithms for the connection machine CM-2. In: Proceedings of the Third Annual ACM Symposium on Parallel Algorithms and Architectures, SPAA, pp. 3–16 (1991). https://courses.cs.washington.edu/courses/cse548/06wi/files/benchmarks/radix.pdf

  13. Church, K.: On memory limitations in natural language processing, unpublished Master’s thesis (1980). http://publications.csail.mit.edu/lcs/pubs/pdf/MIT-LCS-TR-245.pdf

  14. Koskenniemi, K., Church, K.: Complexity, two-level morphology and Finnish. In: Coling (1988). https://aclanthology.info/pdf/C/C88/C88-1069.pdf

  15. Graves, A., Wayne, G., Danihelka, I.: Neural Turing Machines. arXiv (2014). https://arxiv.org/abs/1410.5401

  16. Sun, G., Giles, C., Chen, H., Lee, Y: The Neural Network Pushdown Automaton: Model, Stack and Learning Simulations. arXiv (2017). https://arxiv.org/abs/1711.05738

  17. Banko, M., Brill, E.: Scaling to very very large corpora for natural language disambiguation, pp. 26–33. ACL (2001). http://www.aclweb.org/anthology/P01-1005

  18. Church, K., Mercer, R.: Introduction to the special issue on computational linguistics using large corpora. Comput. Linguist. 19(1), 1–24 (1993). http://www.aclweb.org/anthology/J93-1001

    Google Scholar 

  19. West, G.: Scale. Penguin Books, New York (2017)

    Google Scholar 

  20. Hestness, J., Narang, S., Ardalani, N., Diamos, G., Jun, H.: Deep Learning Scaling is Predictable, Empirically. arXiv (2017). https://arxiv.org/abs/1712.00409

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kenneth Ward Church .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Church, K.W. (2018). Minsky, Chomsky and Deep Nets. In: Sojka, P., Horák, A., Kopeček, I., Pala, K. (eds) Text, Speech, and Dialogue. TSD 2018. Lecture Notes in Computer Science(), vol 11107. Springer, Cham. https://doi.org/10.1007/978-3-030-00794-2_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-00794-2_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-00793-5

  • Online ISBN: 978-3-030-00794-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics