Market Cap: $3.1418T -3.170%
Volume(24h): $101.498B 3.590%
  • Market Cap: $3.1418T -3.170%
  • Volume(24h): $101.498B 3.590%
  • Fear & Greed Index:
  • Market Cap: $3.1418T -3.170%
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
Top News
Cryptos
Topics
Cryptospedia
News
CryptosTopics
Videos
bitcoin
bitcoin

$95869.388821 USD

-2.58%

ethereum
ethereum

$2601.967006 USD

-4.30%

tether
tether

$0.999844 USD

-0.03%

xrp
xrp

$2.398502 USD

-4.77%

solana
solana

$194.031240 USD

-5.54%

bnb
bnb

$631.627084 USD

-0.98%

usd-coin
usd-coin

$1.000012 USD

0.00%

dogecoin
dogecoin

$0.251912 USD

-6.20%

cardano
cardano

$0.759874 USD

-4.61%

tron
tron

$0.238115 USD

-3.18%

chainlink
chainlink

$18.478536 USD

-6.37%

avalanche
avalanche

$24.975027 USD

-6.62%

sui
sui

$3.245364 USD

-8.04%

stellar
stellar

$0.313605 USD

-6.52%

toncoin
toncoin

$3.712774 USD

-4.18%

Cryptocurrency News Articles

Autoregressive Language Models Are Not Doomed: How Chain-of-Thought Prompting and Attentive Reasoning Queries Prevent Incoherence

Feb 12, 2025 at 08:15 am

Yann LeCun, Chief AI Scientist at Meta and one of the pioneers of modern AI, recently argued that autoregressive Large Language Models (LLMs) are fundamentally flawed.

Autoregressive Language Models Are Not Doomed: How Chain-of-Thought Prompting and Attentive Reasoning Queries Prevent Incoherence

Yann LeCun, Chief AI Scientist at Meta and one of the pioneers of modern AI, recently shared his thoughts on a fundamental limitation he sees in autoregressive Large Language Models (LLMs). According to LeCun, the probability of generating a correct response decreases exponentially with each token, making them impractical for long-form, reliable AI interactions.

While I deeply respect LeCun’s work and approach to AI development and resonate with many of his insights, I believe this particular claim overlooks some key aspects of how LLMs function in practice. In this post, I’ll explain why autoregressive models are not inherently divergent and doomed, and how techniques like Chain-of-Thought (CoT) and Attentive Reasoning Queries (ARQs)—a method we’ve developed to achieve high-accuracy customer interactions with Parlant—effectively prove otherwise.

What is Autoregression?

At its core, an LLM is a probabilistic model trained to generate text one token at a time. Given an input context, the model predicts the most likely next token, feeds it back into the original sequence, and repeats the process iteratively until a stop condition is met. This allows the model to generate anything from short responses to entire articles.

For a deeper dive into autoregression, check out our recent technical blog post.

Do Generation Errors Compound Exponentially?

LeCun’s argument can be unpacked as follows:

Let P be the probability of making a generation error at each token.

For an output of length n, the probability of maintaining coherence is (1-E)^n.

This leads to LeCun’s conclusion that for sufficiently long responses, the likelihood of maintaining coherence exponentially approaches zero, suggesting that autoregressive LLMs are inherently flawed.

But here’s the problem: E is not constant.

To put it simply, LeCun’s argument assumes that the probability of making a mistake in each new token is independent. However, LLMs don’t work that way.

As an analogy to what allows LLMs to overcome this problem, imagine you’re telling a story: if you make a mistake in one sentence, you can still correct it in the next one to keep the narrative coherent. The same applies to LLMs, especially when techniques like Chain-of-Thought (CoT) prompting guide them toward better reasoning by helping them reassess their own outputs along the way.

Why This Assumption is Flawed

LLMs exhibit self-correction properties that prevent them from spiraling into incoherence.

Take Chain-of-Thought (CoT) prompting, which encourages the model to generate intermediate reasoning steps. CoT allows the model to consider multiple perspectives, improving its ability to converge to an acceptable answer. Similarly, Chain-of-Verification (CoV) and structured feedback mechanisms like ARQs guide the model in reinforcing valid outputs and discarding erroneous ones.

A small mistake early on in the generation process doesn’t necessarily doom the final answer. Figuratively speaking, an LLM can double-check its work, backtrack, and correct errors on the go.

Attentive Reasoning Queries (ARQs) are a Game-Changer

At Parlant, we’ve taken this principle further in our work on Attentive Reasoning Queries (a research paper describing our results is currently in the works, but the implementation pattern can be explored in our open-source codebase). ARQs introduce reasoning blueprints that help the model maintain coherence throughout long completions by dynamically refocusing attention on key instructions at strategic points in the completion process, continuously preventing LLMs from diverging into incoherence. Using them, we’ve been able to maintain a large test suite that exhibits close to 100% consistency in generating correct completions for complex tasks.

This technique allows us to achieve much higher accuracy in AI-driven reasoning and instruction-following, which has been critical for us in enabling reliable and aligned customer-facing applications.

Autoregressive Models Are Here to Stay

We think autoregressive LLMs are far from doomed. While long-form coherence is a challenge, assuming an exponentially compounding error rate ignores key mechanisms that mitigate divergence—from Chain-of-Thought reasoning to structured reasoning like ARQs.

If you’re interested in AI alignment and increasing the accuracy of chat agents using LLMs, feel free to explore Parlant’s open-source effort. Let’s continue refining how LLMs generate and structure knowledge.

Disclaimer:info@kdj.com

The information provided is not trading advice. kdj.com does not assume any responsibility for any investments made based on the information provided in this article. Cryptocurrencies are highly volatile and it is highly recommended that you invest with caution after thorough research!

If you believe that the content used on this website infringes your copyright, please contact us immediately (info@kdj.com) and we will delete it promptly.

Other articles published on Feb 12, 2025