Can humans learn from how AI learns languages?

AI outperforms humans not only at coding and playing chess, but also at language learning. What “strategies” does AI apply to master languages so successfully? Are we humans capable of performing the same strategies? And is it any good asking these questions?

What will AI become in the next 20 years?

I have just finished listening to one of the most chilling podcasts episode in recent history with Geoffrey Hinton about our future with AI. Hinton, one of the founding fathers of AI, wants to warn us about AI becoming so knowledgeable and intelligent in the years to come that it just might decide to eliminate its inferior creator.

Scientist and Frankenstein

To be honest, I had not worried about that until now. I have a growing sense that the usage of AI could make us lazy and less creative, that it will replace much ‘low-end’ intellectual workers (already happening), however, so far, I’ve been mostly amazed by the rapid development of AI. It has improved so much since it was first released that it would be difficult to put the toy back in the box. From quickly formulating an email, doing research to practicing languages. It seems that AI can do more every day and it is “just too good for too many things”.

AI knows my mother tongue better than I do

Coincidentally, I had another interesting experience today. It turns out that you can also chat in Frisian with ChatGPT. That is a language with approximately 400,000 speakers, has a small written tradition and is most certainly not a world language. It happens to be my first language, but I had not yet thought of testing it in conversation with the popular AI-assistant. Very much to my surprise, it turned out that ChatGPT also mastered this minority language and can even teach it to you.

A conversation in Frisian with ChatGPT
I asked ChatGPT how it learned Frisian. Answer: the training material included Frisian texts.

What can we learn from AI?

This made me wonder how AI models are trained in various languages ​​so effectively and what we humans can learn from that. Do they use strategies that work well for us humans too?

By the way, I see the irony of first creating human-like computer-based intelligence that imitates the way our brains work and, once it has surpassed human cognitive capabilities, it’s us looking for ways to resemble AI.

Key ‘learning strategies’ in AI language models

The answer I got from ChatGPT was (of course) very helpful and not in the least bit derogatory towards us humans. It stated that although the human brain and AI function very differently, there are key learning strategies in AI language models that human learners can meaningfully adopt – especially when it comes to acquiring new languages efficiently and intuitively. Let’s see if ChatGPT is just being kind or we can actually follow the same path.


Unorganized bookstore, bookshelves, books piled up to the ceiling.
A bookstore could be considered a dataset

1. Massive exposure + repetition (input-based learning)

In AI

Language models like GPT are trained on millions of text examples. So for “massive exposure” massive data-sets are needed. Theoretically, the model only needs to see the training material once and store it somewhere. It doesn’t forget and it can be trained 24/7. Imagine reading all the world’s great literature on Gutenberg.org or entire volumes of the Encyclopædia Britannica – and be able to remember every fact and tiny detail for the rest of time.

What’s more, AI models learn language purely through exposure to usage patterns, not abstract rules. Unlike the first AI researcher expected, building artificial intelligence from logic and abstract reasoning never really worked out.

For humans

This whole procedure is somewhat similar to comprehensible input theory, where consuming lots of content – like what I do in my vocabulary notes – that you’re mostly able to understand is key. The brain picks up patterns naturally from rich, frequent exposure on your language level.

The main difference: we humans are obviously more limited in our choice of input. We can’t just upload all classical Chinese literature to our brain nor can we absorb Chinese Wikipedia and start seeing the relevant patterns. Not to mention learning being too slow and frustrating if we would attempt to do so.

Human takeaway

If there’s anything that we can learn from all this then probably this:

  • Read and listen a lot – daily, in meaningful contexts.
  • Use narrow reading/listening: immerse yourself in one topic at a time for repetition.
  • Don’t stress about full understanding (easier said for AI than for humans); regular exposure helps to build fluency over time.

Wizard gazing in his crystal ball
Predicting what comes next…

2. Learning through prediction

In AI

AI learns by predicting the next word in a sentence – this forces it to internalize grammar, word order, and meaning. It reminds me of listening to a boring teacher for so long that you know his next words before he actually says them.

For humans

Neurolinguistics suggests our brains also work predictively when processing language. This is most obvious in conversations, we’re able to complete unfinished sentences or fill in information gaps, but also in texts. When you read “once upon a …”, your brain is likely going to add “time” for example, because it knows the pattern.

Human takeaway

What follows from this is actually surprisingly useful. You can:

  • Pause while reading or listening and try to guess what comes next.
  • Use fill-in-the-blank exercises (cloze tests).
  • Practice active listening by anticipating sentence completions. I also find repeating sentences from hearing an activity that forces you to reproduce certain patterns. The faster you’re able to do it, the better you’ve mastered the patterns.

3. Learning in context, not in isolation

In AI

AI doesn’t learn abstract grammar rules – it learns usage patterns from rich context. It understands context extremely well, far better than Google search used to do – if that is a fair comparison.

For humans

We all know that grammar makes more sense when learned through real examples. It doesn’t make much sense to internalize abstract grammar patterns like a mathematical formula. Although it would be very nice both for learners and teachers, unfortunately, memorizing rules alone rarely leads to fluent usage. Your brain needs practice (in authentic contexts) to reproduce patterns without conscious effort.

Human takeaway

So what do we learn from this? That we should:

  • Learn through chunks and patterns (e.g. “Would you mind…”, “I’m planning to…”). I think Chunking Chinese does a good job at this.
  • Focus on how phrases are actually used, not just what they “mean.”
  • Surround grammar with context – stories, conversations, etc. Make it real and authentic.

Example from a LinkedIn learning course on Large Language Model
This is an example from a LinkedIn learning course on Large Language Models. The ChatGPT 3.0 davinci model fails to follow the instruction to write a shopping list.

4. Transfer learning: build on what you know

In AI

AI is often pretrained on general language data, then fine-tuned on specific tasks. Without this fine-tuning, AI wouldn’t be able to produce meaningful output. ChatGPT 3.0, trained on massive amounts of data, was still struggling to produce a shopping list for example. You had to remind the model what a typical shopping list looks like, give hints and clues for it be able to produce the desired results. So without task-specific training it wasn’t much good. I think this also explains the acceleration of progress, when OpenAI opened ChatGPT to millions of users. It then started using that knowledge on such a scale that it started creating all kinds of new connections.

For humans

You can transfer knowledge from one language to another — especially between related languages. Chinese is not the perfect candidate for this – at least for many learners. But the richer your frame of reference, the easier to spot similarities. When determining the severeness of an insult in Chinese, you can link it to a similar word in English for example, or maybe you’ll find that it can be used in similar contexts and settings.

Human takeaway

So again, when trying to imitate AI learning strategies, you could:

  • Leverage similarities between languages you know.
  • Build on shared grammar, vocabulary, or sentence structure.
  • Use contrastive analysis (to some extent) to accelerate learning. This can be useful for pronunciation. To give an example for people with a German language background: the consonant in ‘ich’ [ç] is quite similar to the Chinese consonant in ‘xi’ [ɕ]. Not identical, but close. Knowing this as a German speaker can be helpful.

5. Focus on frequency: learn the core first

In AI

Models learn high-frequency words and patterns first – because they appear the most in training data. In other words, AI becomes fluent in “high-frequency areas” first. And although by now, I something get the impression ChatGPT knows almost everything, it still might not be able to answer questions about a hardly known medical treatment for some obscure disease entirely correctly. It might use extremely low-frequency medical terms in the wrong way, precisely because they’re not commonly used.

For humans

It’s not so different for us humans. In fact, most learners of Chinese tend to focus on the 1000 most frequently used words or HSK vocabulary.

Human takeaway

So – quite obviously – we shouldn’t waste time on rare or academic words and instead:

  • Focus on the top 1,000–2,000 most common words.
  • Learn high-frequency expressions and sentence structures.
  • Use frequency lists or tools like the CEFR vocabulary framework.

What AI does that humans can emulate (even partially)

TraitIn AITakeaway for humans
Massive repetitionProcesses millions of sentences→ Learn in small but consistent daily chunks
Zero ego or fearMakes endless mistakes→ See mistakes as data, not failure
Unforgiving memoryRetains all frequent input→ Use spaced repetition tools (e.g., Anki)

Summary — what can humans learn from AI language learning?

AI strategyHuman learning tactic
Massive exposureRead/listen constantly in the target language
Predictive processingPractice guessing what comes next
Context-based learningLearn through example-rich input, not rules alone
Transfer from other knowledgeUse related languages to your advantage
Frequency-first focusMaster high-frequency vocabulary and patterns

Steam train roaring down the track - symbol of the industrial revolution

Conclusion

So much for this little excursion into the world of AI. Of course, one might ask how meaningful it really is to try to learn from machines. When the steam train was invented, people didn’t start wondering whether they could learn to run faster by closely examining the locomotive’s mechanism either.

It’s also rather intimidating to compare your own brainpower to that of thousands of processors humming day and night. Just as the cabinetmaker of old couldn’t keep up with the pace of a furniture factory, the human brain is simply outperformed by AI.

Nevertheless, I believe we can retain the principles discussed here:

  • Lots of comprehensible input
  • Learn by predicting patterns
  • Embedded learning, focus on ready-made chunks with context
  • Build on what you already know, activate your frame of reference
  • Focus on high-frequency vocabulary first

Thanks for reading! Any thoughts? Feel free let me know in the comments!

Affiliate links

HSK Standard Course 1 SET - Textbook +Workbook (Chinese and English Edition)
HSK Standard Course 2 SET - Textbook +Workbook (Chinese and English Edition)
HSK Standard Course 3 SET - Textbook +Workbook (Chinese and English Edition)
HSK Standard Course 4a SET - Textbook +Workbook (Chinese and English Edition)
HSK Standard Course 5a SET - Textbook +Workbook (Chinese and English Edition)
HSK Standard Course 6a SET - Textbook +Workbook (Chinese and English Edition)
HSK 1-6 Full Vocabulary Guide: All 5000 HSK Vocabularies with Pinyin and Translation
LELEYU Hieroglyphic Pictograph Symbols Chinese Learning Flash Memory Cards for Children Mandarin Simplified Edition, 252 Characters with Pinyin Stroke Illustrations Kids Educational Toy Gift

Disclosure: These are affiliate links. They help me to support this blog, meaning, at no additional cost to you, I will earn a small commission if you click through and make a purchase.

Leave a Reply

Your email address will not be published. Required fields are marked *