News/Anthropic

Apr 18, 2025

DOGE speeds up AI with GSAi chatbot launch as government job cuts loom

Elon Musk's DOGE is accelerating the deployment of AI tools within federal agencies while simultaneously cutting government jobs, creating tension between automation and workforce reduction. The GSAi chatbot, now available to 1,500 General Services Administration employees, allows government workers to perform various tasks using models from Anthropic and Meta. This rollout comes amid significant staff reductions at GSA, raising questions about the relationship between AI implementation and federal workforce changes. The deployment details: DOGE has expanded the GSAi chatbot from 150 to 1,500 GSA employees after taking over and accelerating a project that was already in development. The chatbot uses...

read
Apr 17, 2025

Google must divest Chrome but can keep AI assets, DOJ rules

Google's ongoing antitrust battle has reached a pivotal moment as the Justice Department surprisingly drops its effort to force the tech giant to divest its AI investments while maintaining pressure on Google to sell Chrome. This shift highlights the DOJ's evolving strategy in addressing Google's market dominance while acknowledging the potential negative consequences of disrupting the rapidly developing AI landscape, especially as Google has invested heavily in AI startups like Anthropic. The big picture: The Department of Justice has abandoned its earlier position that Google should divest from AI companies, citing potential "unintended consequences in the evolving AI space." Google's...

read
Apr 15, 2025

Study reveals Claude 3.5 Haiku may have its own universal language of thought

New research into Claude 3.5 Haiku suggests AI models may develop their own internal language systems that transcend individual human languages, adding a fascinating dimension to our understanding of artificial intelligence cognition. This exploration into what researchers call "AI psychology" highlights both the growing sophistication of large language models and the significant challenges in fully understanding their internal processes—mirroring in some ways our incomplete understanding of human cognition. The big picture: Researchers examining Claude 3.5 Haiku have discovered evidence that the AI model may possess its own universal "language of thought" that combines elements from multiple world languages. Scientists traced...

read
Apr 15, 2025

Salesforce data: AI generates 20% of production code, not Anthropic’s predicted 90%

Salesforce's AI coding tool reveals the real-world impact of AI on software development, contradicting predictions of immediate developer displacement. While Anthropic's CEO predicted AI would write 90% of code within months, Salesforce's actual data shows Agentforce generating 20% of production-level APEX code. This gap between prediction and reality highlights how AI is transforming development roles without replacing humans, offering valuable insights into how AI coding assistants are actually being used at enterprise scale. The big picture: Salesforce's Agentforce coding assistant demonstrates significant but measured AI adoption in enterprise development, with 35,000 monthly users and 10 million lines of accepted code....

read
Apr 15, 2025

Wikipedia’s bandwidth costs surge 50% as AI crawlers strain free knowledge model

Wikipedia's bandwidth costs have spiked 50% since January 2024, a surge the Wikimedia Foundation directly attributes to AI crawlers harvesting its content. This growing tension highlights a fundamental conflict in the AI economy: large language models are consuming vast amounts of online information while potentially diverting traffic and revenue from the very sources that make their existence possible. The big picture: The Wikimedia Foundation reports a dramatic 50% increase in bandwidth costs since January 2024, explicitly blaming AI crawlers for the surge. Why this matters: This significant cost increase threatens Wikipedia's sustainability as a free knowledge resource while raising broader...

read
Apr 14, 2025

Anthropic’s new AI tutor guides students through thinking instead of giving answers

Anthropic's new AI assistant specifically designed for education transforms the traditional student-AI interaction by prioritizing critical thinking development over providing instant answers. This approach represents a significant shift in how AI might be integrated into education, potentially addressing educators' concerns that AI tools encourage shortcut thinking rather than deeper learning. As universities struggle to develop comprehensive AI policies, Anthropic's partnerships with major institutions create large-scale real-world tests of whether AI can enhance rather than undermine the educational process. The big picture: Anthropic has launched Claude for Education with a "Learning Mode" that uses Socratic questioning to guide students through their...

read
Apr 12, 2025

ULMFit, not GPT-1, was the first true LLM according to new analysis

The development of Large Language Models (LLMs) has fundamentally transformed AI capabilities, but understanding their origins helps contextualize today's rapid advancements. While GPT-4 and Claude dominate current discussions, identifying the first true LLM clarifies the evolutionary path of these increasingly sophisticated systems and provides valuable perspective on how quickly this technology has developed in just a few years. The big picture: According to Australian tech blogger Jonathon Belotti, ULMFit, published by Jeremy Howard in January 2018, represents the first true LLM, predating OpenAI's GPT-1 by several months. GPT-1, created by Alec Radford, was published on June 11, 2018, several months...

read
Apr 12, 2025

Study reveals AI models remember past images despite new conversations

Multimodal LLMs appear to leverage conversation memory in ways that affect their performance and reliability, particularly when interpreting ambiguous visual inputs. This research reveals important differences in how models like GPT-4o and Claude 3.7 handle contextual information across conversation threads, raising questions about model controllability and the nature of instruction following in advanced AI systems. The experiment setup: A researcher tested GPT-4o's and Claude 3.7's visual recognition capabilities using foveated blur on CAPTCHA images of cars. The test used 30 images with cars positioned in different regions, applying varying levels of blur that mimicked human peripheral vision. Initially asking "Do...

read
Apr 12, 2025

Microsoft embraces Model Context Protocol as AI agents gain web browsing abilities

The Model Context Protocol (MCP) marks a significant milestone in AI interoperability with its latest specification release, enabling AI agents to interact more effectively with digital tools and interfaces. Microsoft's simultaneous announcement of MCP support—including a new browser automation tool—signals growing industry momentum for a standard that could become the universal language for AI-tool interactions. This development addresses a critical challenge in the AI ecosystem: creating a common protocol for agents to seamlessly communicate with various digital services and applications. The big picture: The MCP project just released an updated specification with substantial upgrades to security, capability, and interoperability for...

read
Apr 11, 2025

Claude enters the web search chat, rivalling Perplexity and ChatGPT

Anthropic's Claude chatbot is catching up to competitors by finally adding web search capabilities, expanding its functionality beyond its native knowledge base. This development marks a significant shift in the AI chatbot landscape as these tools increasingly compete with traditional search engines, offering users conversational interfaces that process and deliver search results with cited sources. The big picture: Anthropic has announced that Claude can now search the web, joining competitors like ChatGPT Search and Perplexity in offering real-time internet access to supplement its built-in knowledge. How it works: Claude processes internet information and delivers relevant sources in a conversational format...

read
Apr 11, 2025

Anthropic finds AI models gaining college-kid-level cybersecurity and bioweapon skills

Anthropic's frontier AI red team reveals concerning advances in cybersecurity and biological capabilities, highlighting how AI models are rapidly acquiring skills that could pose national security risks. These early warning signs emerge from a year-long assessment across four model releases, providing crucial insights into both current limitations and future threats as AI continues to develop potentially dangerous dual-use capabilities. The big picture: Anthropic's assessment finds that while frontier AI models don't yet pose substantial national security risks, they're displaying alarming progress in dual-use capabilities that warrant close monitoring. Current models are approaching undergraduate-level skills in cybersecurity and demonstrate expert-level knowledge...

read
Apr 8, 2025

Outside the lines: 5 ways to unlock Claude’s hidden creative powers beyond text responses

Claude's latest model introduces powerful capabilities that extend far beyond basic text responses, enabling users to create interactive games, animations, and productivity tools directly in their browser. As AI assistants become increasingly versatile, strategic prompting techniques allow users to unlock Claude's full creative and functional potential without requiring technical expertise. 1. Design an immersive murder mystery game The prompt asks Claude to create a complete murder mystery experience set in a 1920s mansion, including eight unique characters, clues, and a surprise twist ending. This showcases Claude's ability to design complex narrative experiences that can be brought to life at real-world...

read
Apr 8, 2025

AI alignment researchers issue urgent call for practical solutions as AGI arrives

The AI alignment movement is sounding urgent alarms as artificial general intelligence (AGI) appears to have arrived much sooner than expected. This call-to-action from prominent alignment researchers emphasizes that theoretical debates must now give way to practical solutions, as several major AI labs are pushing capabilities forward at an accelerating pace that they believe threatens humanity's future. The big picture: The author claims AGI has already arrived in March 2025, with multiple companies including xAI, OpenAI, and Anthropic rapidly advancing capabilities while safety measures struggle to keep pace. Why this matters: The post frames AI alignment as no longer a...

read
Apr 7, 2025

Kai-Fu Lee predicts only 7 AI foundation models will survive globally

Prominent AI entrepreneur Kai-Fu Lee predicts a major consolidation in the AI industry, with just seven foundational models surviving globally across the U.S. and China markets. This forecast reflects the intensifying competition among AI powerhouses as technical barriers and capital requirements create natural selection pressures in the race to dominate large language model development. The big picture: Lee anticipates that only three Chinese AI models will ultimately survive amid a domestic market shakeout triggered by DeepSeek's emergence, while four U.S. companies will dominate the American market. Lee's prediction: The computer scientist and entrepreneur believes DeepSeek, Alibaba Group Holding Ltd., and...

read
Apr 6, 2025

Anthropic aligns with California’s AI transparency push as powerful models loom by 2026

Anthropic's commitment to AI transparency aligns with California's policy direction, offering a roadmap for responsible frontier model development. As Governor Newsom's Working Group on AI releases its draft report, Anthropic has positioned itself as a collaborative partner by highlighting how transparency requirements can create trust, improve security, and generate better evidence for policymaking without hindering innovation—particularly crucial as powerful AI systems may arrive as soon as late 2026. The big picture: Anthropic welcomes California's focus on transparency and evidence-based standards for frontier AI models while noting their current practices already align with many of the working group's recommendations. The company...

read
Apr 6, 2025

Google slashes Gemini 2.5 Pro prices, intensifying AI model pricing war

Google's aggressive pricing strategy for its new Gemini 2.5 Pro model represents a significant escalation in the AI pricing wars, challenging competitors like Anthropic and OpenAI with substantially lower rates. The move comes after Google observed strong developer interest following the model's quiet launch, prompting the company to not only increase rate limits but also position its "most intelligent model ever" at a price point that has surprised industry observers and potentially disrupted the reasoning model market. The big picture: Google has moved Gemini 2.5 Pro into public preview with remarkably competitive pricing that undercuts major competitors. At $1.24 per...

read
Apr 5, 2025

AI firms adopt responsible scaling policies to set safety guardrails for development

Responsible Scaling Policies have emerged as a framework for AI companies to define safety thresholds and capability limits, establishing guardrails for AI development while balancing innovation with risk management. These policies represent a significant evolution in how leading AI organizations approach the responsible advancement of increasingly powerful systems. The big picture: Major AI companies have established formalized policies that specify what AI capabilities they can safely handle and when development should pause until better safety measures are created. Anthropic pioneered this approach in September 2023 with their AI Safety Levels (ASL) system, categorizing AI systems from ASL-1 (posing no meaningful...

read
Apr 1, 2025

Study: Anthropic uncovers neural circuits behind AI hallucinations

Anthropic's new research illuminates crucial neural pathways that determine when AI models hallucinate versus when they admit uncertainty. By identifying specific neuron circuits that activate differently for familiar versus unfamiliar information, the study provides rare insight into the mechanisms behind AI confabulation—a persistent challenge in the development of reliable language models. This research marks an important step toward more transparent and truthful AI systems, though Anthropic acknowledges we're still far from a complete understanding of these complex decision-making processes. The big picture: Researchers at Anthropic have uncovered specific neural network "circuitry" that influences when large language models fabricate answers versus...

read
Apr 1, 2025

AI evaluation shifts back to human judgment and away from benchmarks as models outgrow traditional tests

Actually, human, stick around for a minute, could ya? The evolution of AI evaluation is shifting from automated benchmarks to human assessment, signaling a new era in how we measure AI capabilities. As traditional accuracy tests like GLUE, MMLU, and "Humanity's Last Exam" become increasingly inadequate for measuring the true value of generative AI, researchers and companies are turning to human judgment to evaluate AI systems in ways that better reflect real-world applications and needs. The big picture: Traditional AI benchmarks have become saturated as models routinely achieve near-perfect scores without necessarily demonstrating real-world usefulness. "We've saturated the benchmarks," acknowledged...

read
Apr 1, 2025

Anthropic researchers reveal how Claude “thinks” with neuroscience-inspired AI transparency

Anthropic's breakthrough AI transparency method delivers unprecedented insight into how large language models like Claude actually "think," revealing sophisticated planning capabilities, universal language representation, and complex reasoning patterns. This research milestone adopts neuroscience-inspired techniques to illuminate previously opaque AI systems, potentially enabling more effective safety monitoring and addressing core challenges in AI alignment and interpretability. The big picture: Anthropic researchers have developed a groundbreaking technique for examining the internal workings of large language models like Claude, publishing two papers that reveal these systems are far more sophisticated than previously understood. The research employs methods inspired by neuroscience to analyze how...

read
Mar 28, 2025

Precocious AI: Stanford’s open-source NNetNav agent rivals GPT-4 while learning like a child

Stanford researchers have developed NNetNav, an open-source AI agent that can perform tasks on websites by learning through exploration, similar to how children learn. This development comes as major tech companies like OpenAI, ByteDance, and Anthropic are releasing commercial AI agents that can take actions online on behalf of users. NNetNav addresses key concerns about proprietary AI systems by being fully transparent, more efficient, and equally capable while remaining completely open source. The big picture: Stanford graduate student Shikhar Murty and professor Chris Manning have created an AI system that can reduce the burden of repetitive computer tasks while addressing...

read
Mar 26, 2025

Observe.AI launches VoiceAI agents to automate routine customer service calls

Observe.AI's launch of VoiceAI agents represents a significant advancement in contact center automation, blending various AI technologies to handle routine customer interactions. This solution addresses growing enterprise interest in automating basic customer service tasks while allowing human agents to focus on more complex issues, potentially transforming the economics and experience of customer service operations. The big picture: Observe.AI has released VoiceAI agents to automate routine customer service interactions, positioning itself as the only complete AI-powered platform supporting the entire customer journey. The new solution is designed to handle everything from simple FAQs to multi-step conversations, eliminating long hold times and...

read
Mar 24, 2025

Publishers take legal stand against AI training on copyrighted books

It's East Coast legacy publishing vs. West Coast tech. The publishing industry finds itself locked in an escalating legal battle with tech companies over AI training on copyrighted books, with major implications for intellectual property rights in the digital age. As AI development races forward, publishers are fighting to establish precedents that protect authors' works while acknowledging the need for responsible innovation, creating tension between traditional copyright protections and technological advancement. The big picture: Publishing industry organizations including the AAP and AUP have escalated their AI copyright concerns to the White House, responding to the administration's request for input on...

read
Mar 24, 2025

Pokémon No-Go: Claude’s advanced AI struggles to navigate Pokémon Red despite 3.7 upgrade

Anthropic's advanced AI agent Claude 3.7 Sonnet is struggling to complete the decades-old children's game Pokémon Red, despite being one of the industry's most sophisticated AI models. This experiment highlights the significant gap between current AI capabilities and true autonomous agency, as Claude's difficulties with basic visual processing and navigation demonstrate that even advanced language models still face fundamental challenges when interacting with virtual environments. The big picture: Anthropic is livestreaming "Claude Plays Pokémon" as a demonstration of AI agent capabilities, but progress has been painfully slow and inconsistent. Claude has managed to obtain three Gym badges and reach Cerulean...

read
Load More