What's up in

Large language models

Latest Articles

How ‘Embeddings’ Encode What Words Mean — Sort Of

September 18, 2024

Machines work with words by embedding their relationships with other words in a string of numbers.

Will AI Ever Have Common Sense?

July 18, 2024

Common sense has been viewed as one of the hardest challenges in AI. That said, ChatGPT4 has acquired what some believe is an impressive sense of humanity. How is this possible? Listen to this week’s “The Joy of Why” with co-host Steven Strogatz.

Q&A

Does AI Know What an Apple Is? She Aims to Find Out.

April 25, 2024

The computer scientist Ellie Pavlick is translating philosophical concepts such as “meaning” into concrete, testable ideas.

How Chain-of-Thought Reasoning Helps Neural Networks Compute

March 21, 2024

Large language models do better at solving problems when they show their work. Researchers are beginning to understand why.

How Quickly Do Large Language Models Learn Unexpected Skills?

February 13, 2024

A new study suggests that so-called emergent abilities actually develop gradually and predictably, depending on how you measure them.

New Theory Suggests Chatbots Can Understand Text

January 22, 2024

Far from being “stochastic parrots,” the biggest large language models seem to learn enough skills to understand the words they’re processing.

Tiny Language Models Come of Age

October 5, 2023

To better understand how neural networks learn to simulate writing, researchers trained simpler versions on synthetic children’s stories.

Some Neural Networks Learn Language Like Humans

May 22, 2023

Researchers uncover striking parallels in the ways that humans and machine learning models acquire language skills.

Chatbots Don’t Know What Stuff Isn’t

May 12, 2023

Today’s language models are more sophisticated than ever, but they still struggle with the concept of negation. That’s unlikely to change anytime soon.