Saturday, August 16, 2025
HomeBlock ChainAnthropic takes on OpenAI and Google with new Claude AI features designed...

Anthropic takes on OpenAI and Google with new Claude AI features designed for students and developers

Published on

spot_img

Want smarter insights in your inbox? Sign up for our weekly newsletters to get only what matters to enterprise AI, data, and security leaders. Subscribe Now


Anthropic is launching new “learning modes” for its Claude AI assistant that transform the chatbot from an answer-dispensing tool into a teaching companion, as major technology companies race to capture the rapidly growing artificial intelligence education market while addressing mounting concerns that AI undermines genuine learning.

The San Francisco-based AI startup will roll out the features starting today for both its general Claude.ai service and specialized Claude Code programming tool. The learning modes represent a fundamental shift in how AI companies are positioning their products for educational use — emphasizing guided discovery over immediate solutions as educators worry that students become overly dependent on AI-generated answers.

“We’re not building AI that replaces human capability—we’re building AI that enhances it thoughtfully for different users and use cases,” an Anthropic spokesperson told VentureBeat, highlighting the company’s philosophical approach as the industry grapples with balancing productivity gains against educational value.

The launch comes as competition in AI-powered education tools has reached fever pitch. OpenAI introduced its Study Mode for ChatGPT in late July, while Google unveiled Guided Learning for its Gemini assistant in early August and committed $1 billion over three years to AI education initiatives. The timing is no coincidence — the back-to-school season represents a critical window for capturing student and institutional adoption.


AI Scaling Hits Its Limits

Power caps, rising token costs, and inference delays are reshaping enterprise AI. Join our exclusive salon to discover how top teams are:

  • Turning energy into a strategic advantage
  • Architecting efficient inference for real throughput gains
  • Unlocking competitive ROI with sustainable AI systems

Secure your spot to stay ahead: https://bit.ly/4mwGngO


The education technology market, valued at approximately $340 billion globally, has become a key battleground for AI companies seeking to establish dominant positions before the technology matures. Educational institutions represent not just immediate revenue opportunities but also the chance to shape how an entire generation interacts with AI tools, potentially creating lasting competitive advantages.

“This showcases how we think about building AI—combining our incredible shipping velocity with thoughtful intention that serves different types of users,” the Anthropic spokesperson noted, pointing to the company’s recent product launches including Claude Opus 4.1 and automated security reviews as evidence of its aggressive development pace.

How Claude’s new socratic method tackles the instant answer problem

For Claude.ai users, the new learning mode employs a Socratic approach, guiding users through challenging concepts with probing questions rather than immediate answers. Originally launched in April for Claude for Education users, the feature is now available to all users through a simple style dropdown menu.

The more innovative application may be in Claude Code, where Anthropic has developed two distinct learning modes for software developers. The “Explanatory” mode provides detailed narration of coding decisions and trade-offs, while the “Learning” mode pauses mid-task to ask developers to complete sections marked with “#TODO” comments, creating collaborative problem-solving moments.

This developer-focused approach addresses a growing concern in the technology industry: junior programmers who can generate code using AI tools but struggle to understand or debug their own work. “The reality is that junior developers using traditional AI coding tools can end up spending significant time reviewing and debugging code they didn’t write and sometimes don’t understand,” according to the Anthropic spokesperson.

The business case for enterprise adoption of learning modes may seem counterintuitive — why would companies want tools that intentionally slow down their developers? But Anthropic argues this represents a more sophisticated understanding of productivity that considers long-term skill development alongside immediate output.

“Our approach helps them learn as they work, building skills to grow in their careers while still benefitting from the productivity boosts of a coding agent,” the company explained. This positioning runs counter to the industry’s broader trend toward fully autonomous AI agents, reflecting Anthropic’s commitment to human-in-the-loop design philosophy.

The learning modes are powered by modified system prompts rather than fine-tuned models, allowing Anthropic to iterate quickly based on user feedback. The company has been testing internally across engineers with varying levels of technical expertise and plans to track the impact now that the tools are available to a broader audience.

Universities scramble to balance AI adoption with academic integrity concerns

The simultaneous launch of similar features by Anthropic, OpenAI, and Google reflects growing pressure to address legitimate concerns about AI’s impact on education. Critics argue that easy access to AI-generated answers undermines the cognitive struggle that’s essential for deep learning and skill development.

A recent WIRED analysis noted that while these study modes represent progress, they don’t address the fundamental challenge: “the onus remains on users to engage with the software in a specific way, ensuring that they truly understand the material.” The temptation to simply toggle out of learning mode for quick answers remains just a click away.

Educational institutions are grappling with these trade-offs as they integrate AI tools into curricula. Northeastern University, the London School of Economics, and Champlain College have partnered with Anthropic for campus-wide Claude access, while Google has secured partnerships with over 100 universities for its AI education initiatives.

Behind the technology: how Anthropic built AI that teaches instead of tells

Anthropic’s learning modes work by modifying system prompts to exclude efficiency-focused instructions typically built into Claude Code, instead directing the AI to find strategic moments for educational insights and user interaction. The approach allows for rapid iteration but can result in some inconsistent behavior across conversations.

“We chose this approach because it lets us quickly learn from real student feedback and improve the experience Anthropic launches learning modes for Claude AI that guide users through step-by-step reasoning instead of providing direct answers, intensifying competition with OpenAI and Google in the booming AI education market.
— even if it results in some inconsistent behavior and mistakes across conversations,” the company explained. Future plans include training these behaviors directly into core models once optimal approaches are identified through user feedback.

The company is also exploring enhanced visualizations for complex concepts, goal setting and progress tracking across conversations, and deeper personalization based on individual skill levels—features that could further differentiate Claude from competitors in the educational AI space.

As students return to classrooms equipped with increasingly sophisticated AI tools, the ultimate test of learning modes won’t be measured in user engagement metrics or revenue growth. Instead, success will depend on whether a generation raised alongside artificial intelligence can maintain the intellectual curiosity and critical thinking skills that no algorithm can replicate. The question isn’t whether AI will transform education—it’s whether companies like Anthropic can ensure that transformation enhances rather than diminishes human potential.



Source link

Latest articles

More like this