News/Computing
Firefox’s new AI features cause CPU spikes and drain batteries
Firefox users are experiencing unexpected CPU usage spikes and increased power consumption, with early evidence pointing to artificial intelligence features recently integrated into the popular web browser. Mozilla has acknowledged the performance issues and reversed the rollout of the problematic feature while working on a permanent fix. The performance problems appear linked to an "inference engine" - essentially a local AI system that runs directly on users' devices rather than in the cloud. This engine powers several new features in recent Firefox versions, most notably an AI-powered tab organization system that automatically suggests names for groups of browser tabs. What's...
read Aug 12, 2025Liquid AI’s new vision-language models run 2x faster on smartphones
Liquid AI has launched LFM2-VL, a new family of vision-language foundation models designed for efficient deployment on smartphones, laptops, wearables, and embedded systems. The models promise up to twice the GPU inference speed of comparable vision-language models while maintaining competitive accuracy, addressing the growing demand for on-device AI that can process both text and images without relying on cloud infrastructure. What you should know: LFM2-VL represents a significant step toward making multimodal AI accessible for resource-constrained devices through architectural innovations that prioritize efficiency. The models can process images at native resolutions up to 512×512 pixels without distortion, using smart patching...
read Aug 8, 2025Qualcomm plans AI server chips for 2028 amid competitive challenges
Qualcomm has announced plans to develop server CPUs and full rack systems for AI inferencing clusters in hyperscale data centers, marking a significant expansion beyond its mobile chip dominance. The company is currently in "advanced discussions" with a leading hyperscaler about custom silicon, though revenue from this initiative won't begin until fiscal 2028—potentially giving competitors a substantial head start. What you should know: Qualcomm is positioning itself to capitalize on the shift toward ARM-based processors in cloud computing as AI workloads demand better efficiency metrics. CEO Cristiano Amon explained the company is developing "a general purpose CPU" specifically aimed at...
read Aug 5, 2025Rose-Hulman launches computer science major with AI and cybersecurity tracks
Rose-Hulman Institute of Technology, a private engineering school in Indiana, has launched a redesigned computer science major featuring specialized tracks in artificial intelligence, cybersecurity, and data science. The restructured program offers two distinct pathways—one focused on industry-ready software development and another emphasizing research and theory—allowing students to tailor their education around emerging technologies and high-demand career fields. What you should know: The new unified computer science major replaces Rose-Hulman's previous program structure with a more flexible approach that addresses current industry needs. Students can choose between a Developer pathway for real-world software development and industry careers, or a Researcher pathway...
read Aug 5, 2025AMD beats revenue but stock drops 3% on $800M China export impact
AMD reported weaker-than-expected second-quarter earnings despite beating revenue estimates, with adjusted earnings per share of 48 cents versus the expected 49 cents, while revenue reached $7.69 billion against expectations of $7.42 billion. The chipmaker's stock dropped about 3% in extended trading as the company continues battling U.S. export controls that cost it $800 million in the quarter. Major AI customers like Meta and OpenAI are increasingly turning to AMD as an alternative to Nvidia's dominant graphics processing unit (GPU) offerings. Key financial results: AMD's mixed quarterly performance reflects both growth opportunities and regulatory headwinds in the AI chip market. Net...
read Aug 4, 2025Distance running: Broadcom’s Jericho4 chip connects AI data centers up to 60 miles apart
Broadcom has launched its next-generation Jericho4 networking chip designed to connect data centers up to 60 miles apart and accelerate artificial intelligence computation. The chip addresses growing demands from cloud computing giants like Microsoft and Amazon for faster, more sophisticated networking infrastructure as AI workloads become increasingly computationally intensive and require connecting thousands of graphics processors. What you should know: The Jericho4 introduces several performance enhancements specifically designed for large-scale AI data center operations. A single system can encompass roughly 4,500 chips, enabling massive deployment across extensive networks that operate inside and between data centers. The chip uses the same...
read Aug 4, 2025AMD’s $1.2K R9700 GPU challenges Nvidia in professional AI computing market
AMD's latest professional graphics card has arrived with serious firepower for AI workloads, and major manufacturers are racing to deliver their own versions. The Radeon AI Pro R9700 represents the chip giant's most powerful GPU to date, designed specifically for artificial intelligence training, scientific computing, and high-end rendering tasks that demand sustained performance under intensive workloads. Unlike consumer graphics cards optimized for gaming, professional GPUs like the R9700 prioritize reliability, memory capacity, and multi-card scalability over peak gaming performance. This matters because AI training and scientific simulation often require cards to run at maximum capacity for days or weeks without...
read Jul 30, 2025Foxconn and TECO team up to build modular AI data centers
Foxconn and TECO Electric & Machinery have announced a strategic share-swap alliance to jointly develop modular AI data center infrastructure, combining Foxconn's AI server manufacturing expertise with TECO's electromechanical systems and green energy capabilities. The partnership positions both Taiwanese companies to compete more effectively in the rapidly growing global AI data center market, targeting key regions including Taiwan, Asia, the Middle East, and the United States. The big picture: This alliance represents a significant consolidation in the AI infrastructure supply chain, as hyperscalers and cloud service providers increasingly demand comprehensive, turnkey data center solutions rather than piecemeal components. Key details:...
read Jul 29, 2025Arcee.ai releases AFM-4.5B enterprise AI model for free commercial use
Arcee.ai has opened up its AFM-4.5B enterprise AI model for limited free use, posting the weights on Hugging Face and allowing companies with less than $1.75 million in annual revenue to use it without charge under a custom license. The 4.5-billion-parameter model addresses key enterprise pain points around cost, customizability, and regulatory compliance while being trained exclusively on "clean, rigorously filtered data" to avoid intellectual property violations. What you should know: AFM-4.5B represents Arcee's attempt to bridge the gap between expensive proprietary models and open-weight alternatives that carry licensing risks. The model was developed after discussions with over 150 organizations,...
read Jul 28, 2025It’s such a “Betty”! Penn unveils supercomputer to quadruple AI research capacity
The University of Pennsylvania has unveiled "Betty," a new off-campus supercomputer that quadruples the university's computing capacity and is designed to run AI models analyzing videos, images, texts, and databanks. Located 30 miles from campus in a Collegeville data center, Betty positions Penn to compete in what officials describe as an "arms race in computing" among top research universities seeking to attract faculty and students with cutting-edge AI capabilities. What you should know: Betty represents a significant leap in Penn's computational infrastructure, built in record time to meet surging demand for AI research capabilities. The supercomputer is an Nvidia "SuperPOD"...
read Jul 25, 2025SoftBank deploys world’s largest AI system with 4,000 Nvidia GPUs
SoftBank has deployed what it claims is the world's largest artificial intelligence computing system, installing more than 4,000 of Nvidia's most advanced Blackwell graphics processing units (GPUs) in a massive infrastructure expansion that signals Japan's aggressive push into the global AI race. The Japanese telecommunications giant's new system delivers 13.7 exaflops of computing power—roughly equivalent to performing 13.7 quintillion calculations per second. To put that in perspective, this represents more raw computational horsepower than most national supercomputing centers, positioning SoftBank as a major player in the infrastructure arms race powering today's AI revolution. This deployment isn't just about impressive numbers....
read Jul 24, 2025Answer.AI enables 70B model training on consumer gaming GPUs
Answer.AI has released an open-source system that enables training 70-billion parameter language models on consumer gaming GPUs for the first time. The breakthrough combines FSDP (Fully Sharded Data Parallel) and QLoRA techniques, making it possible to train massive AI models on two 24GB RTX 3090 or 4090 graphics cards—hardware costing under $10,000 compared to hundreds of thousands for data center equipment. The big picture: This development democratizes large language model training by making it accessible to individual researchers, small labs, and the broader open-source community rather than limiting it to well-funded tech companies with expensive data center hardware. Why this...
read Jul 24, 2025Georgia Tech receives $20M to build AI-integrated Nexus supercomputer
Georgia Tech has received a $20 million federal grant to build Nexus, a new supercomputer designed to integrate high-performance computing, artificial intelligence, data analytics, and visualization capabilities into a single system. The supercomputer aims to make advanced computing more accessible for researchers nationwide and could lead to breakthroughs in quantum materials design and brain research. What makes Nexus unique: Unlike traditional supercomputers that require researchers to jump between different machines for different tasks, Nexus will provide multiple computing capabilities in one integrated system. "What is unique about Nexus is that it is going to be designed to provide high-performance computing,...
read Jul 23, 2025Great expectations: IBM’s AI mainframe refresh drives $17B revenue beat
IBM exceeded Wall Street expectations for second-quarter revenue and profit, driven by strong demand for its AI-enhanced mainframe systems and a recovery in its consulting business. The results reflect growing enterprise investment in AI infrastructure as businesses prioritize technology spending amid ongoing economic uncertainty. Key financial highlights: IBM reported revenue of $16.98 billion for the June quarter, up nearly 8% year-over-year and beating analyst estimates of $16.59 billion. The infrastructure segment, which includes mainframes, generated $4.14 billion in revenue, significantly exceeding estimates of $3.81 billion. Second-quarter adjusted earnings of $2.80 per share also surpassed expectations. IBM's "AI book of business"...
read Jul 22, 2025AI data centers drive optical transport market to $19B by 2029
After a challenging 2024 that saw the optical transport market contract by 9%, the telecommunications infrastructure sector is poised for a significant rebound. The catalyst driving this recovery isn't traditional telecom growth, but rather the explosive expansion of artificial intelligence computing infrastructure. The optical transport market—which encompasses the fiber-optic cables, switches, and networking equipment that carry data across long distances—is projected to grow at a steady 5% annually through 2029, reaching $19 billion by the end of the forecast period. This turnaround represents a dramatic shift from the broader telecom industry slowdown that characterized much of 2024. AI data centers...
read Jul 17, 2025UK launches Isambard-AI, Britain’s most powerful supercomputer
The UK's most powerful supercomputer, Isambard-AI, has been made fully operational in Bristol, with Technology Secretary Peter Kyle officially launching the machine as part of the government's expanded AI strategy. The supercomputer will join Cambridge's Dawn machine to form the UK's "AI Research Resource," designed to tackle public challenges like reducing NHS waiting lists and developing climate change solutions while positioning Britain as an "AI maker rather than an AI taker." What you should know: Isambard-AI represents a significant leap in the UK's public AI computing capacity, built with cutting-edge hardware and substantial government investment. The supercomputer uses more than...
read Jul 17, 2025NSF invests $20M in AI supercomputer despite 57% budget cut
The US National Science Foundation is financing a $20-million supercomputer at Georgia Institute of Technology despite facing a proposed 57% budget cut under the Trump administration. The supercomputer, called Nexus, will use AI to advance scientific research across multiple disciplines and represents one of the NSF's most significant funding commitments since the administration proposed slashing its budget from $6.8 billion to $3.9 billion. What you should know: Nexus will be one of the most powerful supercomputers dedicated to AI-driven scientific research, capable of calculating more than 400 quadrillion operations per second. The supercomputer is designed to find new cures for...
read Jul 16, 2025MIT study reveals 3 key barriers blocking AI from real software engineering
MIT researchers have mapped the key challenges preventing AI from achieving autonomous software engineering, arguing that current systems excel at basic code generation but struggle with the complex, large-scale tasks that define real-world software development. The comprehensive study, published by MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL), outlines a research agenda to move beyond today's "autocomplete sidekick" capabilities toward genuine engineering partnership. The big picture: While AI coding tools have made impressive strides, they remain fundamentally limited by narrow benchmarks, poor human-machine communication, and inability to handle enterprise-scale codebases. Current evaluation metrics like SWE-Bench focus on small, self-contained problems...
read Jul 16, 2025China takes the plunge with $223M wind-powered underwater data center off Shanghai
China has launched construction of a wind-powered underwater data center six miles off the coast of Shanghai, marking a significant expansion of its pioneering approach to ocean-based computing infrastructure. The project represents China's bid to address the massive water consumption challenges facing AI and cloud computing while positioning itself as a global leader in sustainable digital infrastructure. The big picture: China is pulling ahead of other nations in deploying underwater data centers as an alternative cooling solution for AI computing, moving from pilot projects to commercial-scale deployments in less than 30 months. Why this matters: Traditional data centers consume hundreds...
read Jul 15, 2025Broadcom’s Tomahawk Ultra chip connects 4x more AI processors than Nvidia
Broadcom has launched the Tomahawk Ultra networking chip, designed to accelerate AI data processing by connecting hundreds of chips within data centers. The processor directly challenges Nvidia's dominance in AI infrastructure, offering four times the connectivity capacity of Nvidia's competing NVLink Switch chip while using enhanced ethernet protocols instead of proprietary systems. Why this matters: The chip represents Broadcom's strategic push to compete with Nvidia in the lucrative AI hardware market, particularly as companies like Google seek alternatives to Nvidia's graphics processors for AI workloads. Key technical advantages: The Tomahawk Ultra serves as a traffic controller for data moving between...
read Jul 15, 2025CoreWeave invests $6B in Pennsylvania AI data center, strategically serving east coast markets
CoreWeave, an artificial intelligence cloud computing firm, has committed up to $6 billion to build a new data center in Lancaster, Pennsylvania. This massive investment represents one of the largest AI infrastructure projects announced to date, highlighting the growing demand for specialized computing resources to support AI workloads. Why this matters: The substantial investment underscores the explosive growth in AI computing demand and the critical need for specialized infrastructure to support increasingly complex AI models and applications. Key details: CoreWeave's announcement signals a major expansion of AI-focused data center capacity in the United States. The Lancaster facility will specifically serve...
read Jul 14, 2025AWS launches Kiro to transform chaotic AI coding into structured workflow
Amazon Web Services has launched Kiro, a new AI coding tool designed to formalize "vibe coding"—the informal process of generating custom code through AI chatbot interactions. The tool aims to address the unstructured nature of current AI coding practices, which a recent study found actually increased task completion time for experienced software engineers by 19%. What you should know: Kiro transforms the chaotic process of AI-assisted coding into a structured workflow with built-in project planning and quality controls. Developers start by entering specifications for each project component, then use AI to generate code that meets those requirements. The tool creates...
read Jul 11, 2025$1B Solo.io’s Kagent Studio brings AI agents to Kubernetes workflows
Solo.io has won the "Most Likely to Succeed" award at VB Transform 2025's Innovation Showcase, where the billion-dollar cloud-native application networking company announced Kagent Studio, a framework for building and managing AI agents in Kubernetes environments. The recognition validates enterprise interest in Solo.io's platform engineering solutions, particularly as companies increasingly adopt AI-driven automation within their cloud infrastructure. What you should know: Kagent Studio is the first cloud-native framework specifically designed for DevOps and platform engineers to build, secure, run and manage AI agents in Kubernetes. The framework integrates natively with VSCode (a popular code editor) and provides real-time incident response...
read Jul 11, 2025Swiss universities to release 70B parameter open-source LLM in 2025
ETH Zurich and EPFL will release a fully open-source large language model in late summer 2025, trained on the "Alps" supercomputer at the Swiss National Supercomputing Centre. The model represents a significant milestone in open AI development, offering multilingual fluency in over 1,000 languages and positioning European institutions as credible alternatives to closed commercial systems from the US and China. What you should know: The LLM will be completely transparent, with source code, weights, and training data publicly available under the Apache 2.0 License. Unlike commercial models developed behind closed doors, this approach enables high-trust applications and supports regulatory compliance...
read