I am surprised we don't see more "Dr. Doolittle" projects like this. I assumed rats or corvids would be good candidates for animal language translation projects since you can keep them in a confined space and record video of them. I am sure that body language plays a huge role in animal communication.
I recently read a paper[0] that claimed to have decoded the basic building blocks of Sperm whale language. I went and took a look at the github for the project CETI and found that most of the code was for the whale trackers and hydrophones. It seems like there are a lot of pre-requisite problems that you have to solve to even get good whale recordings.
On the other hand though, whales probably don't rely on body language since they are communicating way out of line of sight. So it may be easier in that regard.
Anyways, I am convinced that we will figure out how to teach some basic human concepts (like self) to animals and the intelligence of even "stupid" animals like chickens will make people more reluctant to eat meat.
[0] https://www.nature.com/articles/s41467-024-47221-8#:~:text=S....
A major reason was that there was a dogma that language is uniquely human phenomenon. It was already part of the hugely Decartes' philosophy of mind and got a strong revival with Chomskian linguistics and the "cognitive revolution" in the late 1950's.
There was sporadaric research into animal linguistics, e.g. the Koko study, but those were dismissed on the grounds that animals can't have language by definition.
The ethics of enslaving and torturing animals is definitely part of the motivation for the dogma.
To be clear - as of today, many researchers would agree that language is still a uniquely human phenomenon. They discuss this pretty explicitly in the article linked, how it is important to draw a distinction between language and communication. There are no non-human species that have been found to use language for the Chomskian definition of language (using a finite set of symbols to represent an infinite number of communicable meanings).
However, this "dogma" as you call it is beginning to be weakened as researchers document more nuance and complexity in non-human communication than ever before, and so some researchers begin to say, "maybe we shouldn't have this all-or-nothing view of language". But it is simply not true that researchers are suppressing evidence of language in animals out of a desire to enslave and torture them.
> There are no non-human species that have been found to use language for the Chomskian definition of language (using a finite set of symbols to represent an infinite number of communicable meanings)
It's far from clear whether humans are capable of the Chomskian criteria of language. And Chomskian linguistics have more or less collapsed with the huge success of statistical methods.
Chomsky's poverty of stimulus argument is, if anything, strengthened by LLMs. You need to read the entire internet to make statistical methods work at producing grammatical texts. Children don't read the entire internet but do produce grammatical texts. Therefore &c. QED.
I think this is greatly complicated by the fact that the human brain has been "pre-trained" (in the deep learning sense) by hundreds of millions of years of evolution.
A pre-trained LLM also can also learn new concepts from extremely few examples. Humans may still be much smarter but I think there's a lot of reason to believe that the mechanics are similar.
The poverty of the stimulus (POS) argument is that "evolutionary pre-training" in the form (recursive) grammar is fundamentally required and can not be inferred from the stimulus.
The argument is based on multiple questionable assumptions of Chomskian linguistics:
- Humans actually learn grammar in the Chomskian way - Syntax is separate from semantics, so only language (utterances) can be learned from uttrances, and not e.g. what is seen in the environment - At least in the Gold's formalization of the argument language is learned only from "positive examples", so e.g. the learner can't observe that some does not understand some utterance
One could argue for a (very) weak form of POS that there has to be some kind of "inductive bias" in the learning system, but this applies to all learning as shown by Kant. The inductive bias can be very generic.
>At least in the Gold's formalization
It seems to be a persistent myth (possibly revived more recently due to Norvig?) that Chomsky's POS argument has some interesting connection to Gold's theorem. The two things have only a very loose logical connection (Gold's theorem is in no sense a formalization of any claim of Chomsky's), and Chomsky himself never based any of his arguments for innateness on Gold's theorem. Here is a secondary source making the same point (search for 'Gold'): https://stevenpinker.com/files/pinker/files/jcl_macwhinney_c...
The assumption that syntax is 'separate from semantics' also does not figure in any of Chomsky's POS arguments. Chomsky argued that syntax was separate from semantics only in the fairly uncontroversial sense that there are properly syntactic primitives (e.g. 'noun', 'chain', 'c-command') that do not reduce entirely to semantic or phonological notions. But even if that were untrue, it would not undermine POS arguments, which for the most part can be run without any specific assumptions about the syntax/semantics boundary. Indeed, semantic and conceptual knowledge provides an equally fertile source of POS problems.
Yeah, I don't necessarily buy the whole Chomskian program. I'm willing to be persuaded that the reason kids learn to speak despite their individual poverty of stimulus is that there was sufficient empirically experience stimulus over evolutionary time. The Chomskian grammar stuff seems way too Platonic to be a description of human neuroanatomy. But be that as it may, it's clear the stimulus it takes to train an LLM is orders of magnitude greater than the stimulus necessary to train an individual child, so children must have a different process for language acquisition.
Children do get ~6000 hours a year of stimulus. Spoken, unspoken, written, and body language. Even then they aren't able to form language proficiently until 5 or 6 years old. Does the internet contain 30,000 hours of stimulus?
30,000 hours is about the amount of new video uploaded to YouTube every hour.
That's astonishing. If you watched all of them, how much new information would you learn? I suspect a large portion of them are the same information presented differently; for example a news story duplicated by hundreds of different channels.
It's a huge amount of video-game footage included in those "hours of video uploaded per-hour".
So very, very little new info will be conveyed by the vast majority of the content.
Yeah, I imagine every moment of communication a child receives is new information not just baby talk about getting the spoon in their mouth and asking them if they have pooped.
> Even then they aren't able to form language proficiently until 5 or 6 years old.
Not even close; they're already proficient at two.
> Does the internet contain 30,000 hours of stimulus?
Is this a joke?
I'm sure someone else could calculate the informational density of all of the text on the internet vs. 30,000 hours of sight, smell, touch, sound, etc density. My intuition tells me it's not even close.
Does the information contained in smell and touch contribute to the acquisition of language? Keep in mind you'd be arguing that people born without a sense of smell take longer to develop language, or are otherwise deficient in it in some way. I'm doubtful. It's certainly tricky to measure full sight / sound vs. text, but luckily we don't have to, because we also have video online, which, surprise surprise, utterly dwarfs 30,000 hours of sight and sound in terms of total information.
One qualitative difference is that the child's 30,000 hours is realtime, interactive, and often bespoke to the individual and context. All the videos on youtube are static and impersonal.
I agree its not even close! A single day of YouTube uploads alone is 720,000 hours!
I think what he's saying is that "real world" interaction is so high bandwidth it dwarfs internet (screen based) stimulation. Not saying I agree just that he's not comparing hours being alive to hours of youtube
> And Chomskian linguistics have more or less collapsed with the huge success of statistical methods.
People have been saying this for decades. But the hype around large language models is finally starting to wane and I wouldn't be surprised if in another 10 years we hear again that we "finally disproved generative linguistics" (again?)
Also, how many R's are in "racecar"?
Counterpoint: What progress has generative linguistics made in the same amount of time that deep learning has been around? It sure doesn't seem to be working well.
Also, the racecar example is because of tokenization in LLMs - they don't actually see the raw letters of the text they read. It would be like me asking you to read this sentence in your head and then tell me which syllable would have the lowest pitch when spoken aloud. Maybe you could do it, but it would take effort because it doesn't align with the way you're interpreting the input.
>What progress has generative linguistics made in the same amount of time that deep learning has been around? It sure doesn't seem to be working well.
Working well for what? Generative linguistics has certainly made progress in the past couple of decades, but it's not trying to solve engineering problems. If you think that generative linguistics and deep learning models are somehow competitors, you've probably misunderstood the former.
Also being able to count number of letters of a word is not required for language capability in the Chomskian sense at least.
> using a finite set of symbols to represent an infinite number of communicable meanings
This always seemed wildly implausible to me. A very large number of communicable meanings, sure, but infinite?
> This always seemed wildly implausible to me. A very large number of communicable meanings, sure, but infinite?
This is "trivial" in the boring kind of way. With just digits, we can communicate an infinite set of distinct numbers simply by counting.
We can't really communicate an infinite amount of numbers. People just can't read or remember too many digits.
We can. Scientific notation with 1 significant figure can be meaningful because we can use it to figure out order relations. It’s an infinite language.
David Deutsch claims in “The Beginning of Infinity” this is a property called universality, and that we have it. A short excerpt:
https://www.lesswrong.com/posts/HDyePg6oySYQ9hY4i/david-deut...
The whole book is worth reading, though, as it lays it out in more detail.
Seems trivially demonstrable because you can just chain things forever?
Mary ran after the dog and the dog was brown and a cat came along and…
> you can just chain thing forever
I think you're going to find out that no, you can't, and this impossibility is going to trivially demonstrate itself.
Recite 99 bottles of beer on the wall, but start from 1 and change so the number increases? Stop when there are no remaining numbers or when you reach infinity, whichever comes first.
So, is this a proposal to test how long it takes for you to lose your count?
They are talking as if language was some platonic construct like a Turing machine with an infinite tape and you are talking about the concrete reality where there are no such things as an infinite tape.
Both viewpoints are useful, they can prove general properties that hold for arbitrary long sequence of words and you put a practical bound on that length.
The question is if human are capable of infinitely extensible language.
That's clearly false. It's not about some platonic mathematical simplification. Humans patently do not fit the Chomsky criterium for intelligence.
In fact, I'm pretty sure it's physically impossible for any real being to fit it.
Can you say more? English doesn't have any cap on sentence length I think i'm missing your point
> English doesn't have any cap on sentence length
Well, yes and no. Constructing this "infinite" sentence will run into some serious problems once the last star burns out, possibly sooner.
"I have a truly marvelous demonstration of this proposition which this margin is too narrow to contain."
Since English has several possible sentences that are infinite in length, made up of only one word even https://medium.com/luminasticity/grammatical-infinities-what... I have to agree with all the this is trivial comments.
Whatever "finite set of symbols" humans use to communicate is not the finite set of symbols that form letters or words. Communication isn't discrete in practical sense, it's continuous - any symbol can take not just different meanings, but different shades and superposition of meanings, based on the differences in way it's articulated (tone, style of writing - including colors), context in which it shows, and context of the whole situation.
The only way you can represent this symbolically is in the trivial sense like you can represent everything, because you can use few symbols to build up natural numbers, and then you can use those numbers to approximate everything else. But I doubt it's what Chomsky had in mind.
> "maybe we shouldn't have this all-or-nothing view of language"
That idea seems like a strawperson, not something anyone seriously thinking about it would say. Everyone sees animals communicate.
It's not a question of communication, it's a question of language.
This is not really the point though. Why is "uses Chomskian language" the criteria for whether it not it's okay to to change and slaughter a living being?
There is and remains a desire to explain exactly how it is that humans are different than other animals. Language or the language faculty has been touted by some as this thing.
I do not know anyone did this. I don't think biologist care about this.
Those researchers are just making noises. It doesn't mean anything.
Your usage of "language" here is akin to laymen usage of "hypothesis" and "theory" and then trying to apply it in an academic context. Same sequence of letters but different meaning. In linguistics, "language" has a specific definition that only humans have been shown to have. Some trained individuals like Koko do seem to demostrate an very limited ability to use "language" in the linguistics sense.
You might argue that the definition itself is arbitrary and coming from the same place that geocentrism, creationism and flat-Earth views come from. I can't argue for or against that.
I suspect things as more nuanced than the current definition that we have though, especially after the recent study from the Scientific American that heated up Hacker News in a way that only "Is CS a science" articles can.
There's no consensus on the definition of what language is.
Chomskian linguistics does posit that human language is based on (innate) recursive grammars (narrow language faculty hypothesis), but this has always been a contentious question. And per that definition humans too have demonstrated only very limited ability in e.g. infinite embedding.
> There's no consensus on the definition of what language is.
That's according to you. Just because a few apples disagree doesn't mean there is no consensus.
My dog can push buttons to let me know what he wants. Those buttons speak in English. Is that language?
"Language" in the sense of "the thing only humans have been shown to do" requires a bit more than just one to one correlations between signifiers and objects (or a "sentence" of signifiers with the same meaning as all of the words added together independently). For a system of symbols to be "language" there must be a difference between "what the cat ate" and "what ate the cat". No animal communication has been shown to have a grammar to it, and thus the ability to express exponentially many unique ideas with each additional word.
I feel like there are human languages where the symbolic distinction between "what the cat ate" and "what ate the cat" are nil and the understanding is achieved contextually.
Is there a grammar? ie. a set of rules to form valid sentences in the language
Is there linguistic creativity? ie. you can generate new words to describe things never encountered
Is there metalinguistic reflexion? ie. can you use the language to talk about the language itself?
Can the language allow displacement? i.e. talking about things that are not in the current spacetime point
These are some of the core requirements for a system of communication to be considered a language.
1. Buttons don't "speak"
2. You don't need to understand the words to push the buttons. You could replace the English words with gibberish and it would still work as long as you always give the same thing when the same button is pressed. Many animals can do this. It is called positive reinforcement. Nothing to do with language
I think it only counts if he can express that he wants you to urinate on the same fire hydrant after he does.
That’s the minimum level of complexity science will accept.
I think most dog owners would tell you that their adult dogs can communicate things like this, but that the language is unfortunately siloed into a very personal relationship that is difficult for even the human part of the pair to demonstrate, making it difficult to do science about
Sometimes at bedtime my cat will go to the door and scream nonstop. I don't know why he does it. Maybe it is for food or attention. But the only way I have found to get him to stop is to pick him up, put him on his special pillow, squish him, and have my partner join me in telling him "we are going to bed, it's bedtime".
I'd say about 80% of the time he listens. So he is capable of understanding what we want him to do, and capable of supressing his own personal desires in order to maintain harmony in our group. Funny enough, he won't go to bed unless both me and my partner tell him it is bedtime, so maybe he is only obeying because there is some majority consensus?
Because of this, I find it easy to believe that a cat or dog could be taught something as abstract as "self" if they can understand commands and intent and group dynamics. It's just difficult to tell what is "understood" and what is just conditioned behavior. Hell, I can't even answer that question for myself as a human.
If your cat could do most of the below:
- use arbitrary links between signifier and signified
- generate new linguistic tokens (new signifier and signifieds as well as links between them)
- refer to events and times beyond the current ones
- talk about the system of communication itself using the system itself
Then, you would have grounds to think that your cat uses a language in the linguistic sense. But until, then it is just a communication system no matter how sophisticated.
Your cat being a extremely well oiled operant conditioning system does not mean that it is able to think the way you can even if they are likely more intelligent than what we give them credit for because, as much as we would like to believe we do, we don't know what they think and any patterns that we see are just good old human pareidolia. Like hearing voices in the wind, faces on rocks, etc. Your feeling is real but the belief that feeling induces does not have grounds in reality as far as we know so far.
It is possible to over-select on skepticism about other species. Imho, the simplest explanation is that it is far more likely that there is nothing special about us, and a mere quirk of say, the combination of tool-use, foresight, and social cohesion that makes humanity special.
Or is it special? We are just a well oiled operant conditioning system, it does not mean that we are able to think the way cats can.
Laymen think of language the same way they do about theory and then try to apply that to an academic context. Different meaning. A system of communication is not necessarily a language even if all languages are systems of communication. If your dog could use arbitrary linguistic tokens, generate new ones, describe things that it has not seen before, talk about the past, the future or places other than the current one, then I would be more willing to entertain the idea that your dog has a language
There still is, as far as I can tell. Whenever my curiosity drives me to take a psychology or philosophy class I end up with the feeling that they think part of their job is to reassure the rest of the humans that we are in fact special. It feels like some kind of leftover from when that kind of work was done by monks.
We are objectively special in creating technological civilization with all sorts of cultural artifacts like philosophy that we have no evidence for in other species that have existed on this planet, other than possibly a few of our close hominid relatives. Hominids are a very special evolutionary branch in that sense.
When we think about ETs, we're wondering about technological civilizations on other planets with space craft and radio telescopes, not the equivalent of birds or dolphins.
Really? I distinctly remember a lot of pissed off kids in my college philosophy and psychology classes trying to defend their religious beliefs and that we are more than just monkeys with fancier tools. Most of the religious folks (at least vocally) dropped out of Philosophy 101 after 2 weeks. It was incredibly entertaining. I guess this was 20 years ago, but assuming we are a more secular society I guess I thought that would still be the case.
It's a bit different at the intro level. I'm talking about the professors and the grad students. It's not that they're directly religious, but I get a status-quo-preserving kind of feeling from them. Like maybe they're influenced by a tradition of not calling your patron an ape--or somesuch.
Weird, I would think it would be much less hand-holdy at that level.
It feels more like gatekeeping: let nobody who thinks animals talk to each other call themselves a psychologist.
Hum... So, you are in full agreement with the GP?
However, another major reason is that people have repeatedly gone seeking for language-like or human-language-level behaviors in animals, and repeatedly and consistently failed.
It is also worth pointing out that detecting language is a great deal easier than understanding language. Something like https://www.youtube.com/watch?v=vvr9AMWEU-c is reasonably recognizable as clearly some sort of language even if we have no (unassisted) human idea what it is saying. We can tell with quite high confidence that most animal sounds are not hiding some deeper layer of information content.
Such exceptions as there are, like whalesong, take you back to my first paragraph, though.
The idea that language is a uniquely human phenomenon may be "dogma", but it is also fairly well-founded in fact. It should also not be that surprising; had another species developed language first, they'd be the ones looking around at their surroundings being surprised they are the only ones with proper language, because they'd probably be the dominant species on the planet. It isn't a "humanist" bias, in some sense that humans are super special because they're humans, it's a "first species to high language" bias, which happens on this planet to be humans.
> those were dismissed on the grounds that animals can't have language by definition.
What modern person said that? I've read a bit about it and didn't encounter that.
> The ethics of enslaving and torturing animals is definitely part of the motivation for the dogma.
What does that have to do with language?
Dr Doolittle on corvids would be easy.
1. Train crows to push a touchscreen for reward of food.
2. Next set up two touchscreens back to back. Make it so touching one screen only dispenses food on the other side.
3. Next make it so food is dispensed on the other side only one crow is perched at each terminal.
4. Next make it so food is only dispensed after a crow says something to the other crow on the other side.
5. Next display a picture on one terminal and give the other crow the choice of four quadrants. The food is dispensed if the picture on the far side matches the displayed picture.
6. Start decoding words.
Not exactly your protocol, but I'm reminded of https://www.smithsonianmag.com/smart-news/scientists-taught-...
This could be done cheaply with some rooted Kindle Fire tablets. I don't follow 100% but it sounds cool.
Hey, that makes a lot of sense. There's a lot of crows who come to the bird feeder in my backyard. I would just have to figure out how to easily make a food dispenser, and what sort of touch screen a crow can activate....
I would do it myself but where I am there are almost no corvids, they (and city pigeons) are basically displaced by the highly aggressive local bird. Which seems to swarm around from time to time but doesn't stay anywhere permanent locally except for grocery store parking lots. That's probably enough to leak location info on this anon account so I'll shut up now
> I am surprised we don't see more "Dr. Doolittle" projects like this
Robots asking millions of birds to attack any hooman at sight in its own bird language, AKA the Tippi Hedren project. Seems pretty fly.
More likely, instructing birds to fly in Nike Swoosh formation (or some other logo) on command for cheap low altitude sky adverts.
And this is why capitalism will always win over communism :)
I think the thing about animal language is there is no objective truth to it. Even with well defined words and educated and experienced communicators, miscommunication and ambiguity is everywhere in human language.
For animals I think there is less language as we know it. Communication is an instantaneous and unfiltered expression of mental state, and there isn’t much guarantee that it’s received as intended. My dog is the friendliest little dude. He will play bow other dogs immediately, and then jump around with excitement. But it’s extremely common that other dogs misread this and respond with aggression.
I suppose I think animal language is less something we could ever learn academically and more something you just feel.
I think one problem with the idea of learning the language of captive animals is that it runs on the assumption that animals have an inherent language. There was some research centuries ago that involved raising kids without exposure to the outside world of languages in hopes of learning what our "natural" language was. Turns out there isn't one.
And I think the same applies to animals. Bird songs have already been noted to have accents by region. We'll probably end up finding out that if animals do have complex language, it's something they develop through community exposure just like humans do, and it'll vary by area. Crows in Texas will talk very different from crows in England, I'm sure.
With pets, most of them are kept isolated from the world at large. It wouldn't surprise me if dogs raised by people have very stilted or nonexistent language compared to feral dogs. Maybe 200 years from now, we'll see the modern concept of raising dogs completely alone as inhumane and sending them off to doggy daycare a couple times a week to learn the local dog language and "culture" will be normalized.