
LLMs reshaped how teams think about automation, speed, and knowledge work. Yet a recent article by AI critic Gary Marcus has renewed a key conversation in the industry, raising questions about reliability, reasoning, and where AI may need to go next.
This issue shares a curated summary of the article’s viewpoints and our perspective on what today’s leaders should take away from it.
In this issue, you’ll get:
Below is a condensed view of the positions presented in the original Substack piece. These are the author’s views, summarized for readers who want fast clarity.
For those who want the full context, you can read the original article here:
https://garymarcus.substack.com/p/three-years-on-chatgpt-still-isnt
Although the viewpoints come from a critical perspective, they surface questions every enterprise should consider:
This moment calls for thoughtful adoption rather than blind enthusiasm. Leaders who approach AI with clarity and realistic expectations will be better prepared than those who hope for automatic transformation.
Alongside the Substack critique, recent comments from an OpenAI co-founder suggest that simply training ever-larger models may offer diminishing returns. Training approaches may need to shift toward more efficient methods, improved architectures, or systems designed specifically for reasoning.
Yocum Technology Group continues to watch these developments because they signal where enterprise-ready AI may ultimately land.
For business leaders: these shifts suggest that AI’s next phase could move from general-purpose “talking tools” toward more purpose-built, reliable systems — if developed carefully.
Q: Is the article suggesting that LLMs have no business value?
A: No. The author critiques their reasoning limitations, but he acknowledges that LLMs are useful for language-generation tasks.
Q: Does the article believe LLMs can improve with scaling?
A: The author expresses skepticism that scaling alone will solve core reasoning challenges.
Q: Are hallucinations expected to persist?
A: According to the article, yes. Because hallucinations stem from how LLMs operate, the author believes they will remain a persistent issue without architectural changes.
Q: Is this critique shared across the entire AI field?
A: Not universally. The article represents one perspective, though these questions are actively debated in the AI community.
Q: Are LLMs “good enough” for enterprise-level automation?
A: For many supportive tasks (drafting, summarization, ideation), yes. For tasks requiring reliable reasoning, decision-making, or compliance — not yet.
Q: Could LLMs evolve to the point where they can reason like humans?
A: The field is trying. Recent directions point to more reasoning-focused architectures and hybrid systems. That said, fundamental differences remain — pattern recognition vs. true reasoning.
Q: What are the risks of over-relying on LLMs internally or with clients?
A: Hallucinations can lead to misinformation, flawed output, compliance issues, or reputational damage. Without oversight, reliance on LLMs can be dangerous.
Q: If “bigger” LLMs aren’t the answer, where should we look?
A: Expect progress from smarter architectures, hybrid AI combining reasoning and pattern recognition, and models optimized for specific tasks — not general-purpose fluff.
Based on and inspired by the article “Three Years On, ChatGPT Still Isn’t Very Good At (Most) Things…” by Gary Marcus
Original article: https://garymarcus.substack.com/p/three-years-on-chatgpt-still-isnt
We are following conversations about:
If your team is evaluating where LLMs can safely add value or how to build guardrails around their use, we can help design workflows that balance speed with reliability.
Thanks for reading this condensed look at one of the more talked-about critiques in the AI world. If you’d like help applying these ideas inside your organization, we’re here to support.