Abstract
A full account of the structure and rules of natural languages remains an elusive goal for linguists. One way to gain insights into the mechanics of human language is to create an artificial language. Such languages are often constructed by speakers who are motivated by the desire to design a language with a “perfect” structure that lacks the idiosyncrasies of natural languages (e.g. Lojban). Others are invented by people interested in creating a utopia (as in the case of Esperanto) or a dystopia, as portrayed in certain works of science fiction (e.g. Star Trek and its Klingon language). In all cases, creators of the artificial languages strove to make them functional languages that would or could be accepted and usable by speakers. It seems therefore reasonable to assume that the inventors of these languages drew on their native linguistic knowledge and intuitions. They deliberately designed a lexicon appropriate to their purposes and probably reflected on the morphological and syntactic properties of their languages. By contrast, the statistical properties of natural languages are opaque to everyday speakers, although they have been shown to play a role in language acquisition (Safran et al. 1996) and linguistic behavior (Fine et al. 2013). Just as phonological and syntactic features of natural languages arguably draw from a universal inventory, statistical properties may set natural languages apart from other forms of superficially similar information-encoding systems. Rao et al. (2009) undertook a statistical analysis of several natural languages and non-linguistic systems including human chromosome sequences and the programming language Fortran. Rao et al. showed that the two kinds of data can be clearly differentiated in particular with respect to entropy, which measures the unpredictability of elements (such as words) in a sequence (such as a phrase or sentence). We extend this approach by comparing the statistical properties, including entropy, of two different artificial languages, Klingon and Lojban, to Rao et al.’s data. The results reveal both similarities with, and differences from patterns that characterize natural languages and non-linguistic sequences.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Bird, S., Loper, E., & Klein, E. (2009). Natural language processing with Python. Sebastopol: O’Reilly Media Inc.
Brown, J. C. (1960). Loglan. Scientific American, 202, 43–63.
Curnow, R. (2001). Jbofihe. Retrieved from http://www.rc0.org.uk/jbofihe/ and http://www.rpcurnow.force9.co.uk/jbofihe/.
Farmer, S., Sproat, R., & Witzel, M. (2009). A refutation of the claimed refutation of the nonlinguistic nature of Indus symbols. Retrieved from http://www.safarmer.com/Refutation3.pdf.
Fine, A. B., Jaeger, T. F., Farmer, T. A., & Qian, T. (2013). Rapid expectation adaptation during syntactic comprehension. PLoS ONE, 8(10), e77661.
Francis, W. N., & Kučera, H. (1982). Frequency analysis of English usage: Lexicon and grammar. Boston: Houghton Mifflin.
Lee, R., Jonathan, P., & Ziman, P. (2010). Pictish symbols revealed as a written language through application of Shannon entropy. Proceedings of the Royal Society A, 466, 2545–2560.
Lujvo expander version 0.2. (2001). Retrieved from the Lojban electronic mailing list http://tech.groups.yahoo.com/group/lojban/message/11387.
Nicholas, N., & Strader, A. (2000). Hamlet Prince of Denmark: The restored Klingon version. New York: Pocket Books.
Okrand, M. (1985). The Klingon dictionary. New York: Pocket Books.
Rao, R., Yadav, N., Vahia, M. N., Joglekar, H., Adhikari, R., & Mahadevan, I. (2009). Entropic evidence for linguistic structure in the Indus script. Science, 324, 1165. doi:10.1126/science.1170391.
Rao, R., Yadav, N., Vahia, M. N., Joglekar, H., Adhikari, R., & Mahadevan, I. (2010). Entropy, the Indus script, and language: A reply to R. Sproat. Computational Linguistics, 36(4), 795–805.
Rao, R. (2010). Probabilistic analysis of an ancient undeciphered script. IEEE Computer, 43(4), 76–80.
Safran, E., Aslin, R., & Newport, E. (1996). Statistical learning by 8-months old infants. Science, 274(5294), 1926–1928.
Shannon, C. E. (1948). A Mathematical theory of communication. Bell System Technical Journal, 27(3), 379–423.
Yadav, N., Joglekar, H., Rao, R., Vahia, M. N., Adhikari, R., & Mahadevan, I. (2010). Statistical analysis of the Indus script using n-grams. PLoS ONE, 5(3), e9506. doi:10.1371/journal.pone.0009506.
Zamenhof, L. L. (1887). Unua Libro. Warsaw.
Zipf, G. K. (1935). The psychobiology of language. Boston: Houghton-Mifflin.
Acknowledgments
Funding for this work, which was performed as part of R. S.’s graduation requirement, was generously provided by the Princeton University Program in Linguistics. R. S. is grateful to Constantine Nakos and Vyas Ramasubramani for technical help.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Smaha, R., Fellbaum, C. (2015). How Natural Are Artificial Languages?. In: Gala, N., Rapp, R., Bel-Enguix, G. (eds) Language Production, Cognition, and the Lexicon. Text, Speech and Language Technology, vol 48. Springer, Cham. https://doi.org/10.1007/978-3-319-08043-7_17
Download citation
DOI: https://doi.org/10.1007/978-3-319-08043-7_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-08042-0
Online ISBN: 978-3-319-08043-7
eBook Packages: Computer ScienceComputer Science (R0)