News/Computing

Jan 16, 2025

Transforming into AI factories the rise of decision-making engines in modern data centers

Artificial Intelligence data centers are rapidly transforming from traditional computing facilities into AI factories, with significant implications for infrastructure, energy consumption, and economic output. The transformation underway: Google's conversion of the Widows Creek Fossil Plant into a data center powered by renewable energy exemplifies the broader shift from industrial to digital infrastructure. These AI factories function as decision-making engines that require massive computing, networking, and storage resources The facilities face many of the same challenges as traditional industrial factories, including power management, scalability, and reliability concerns Data centers are being constructed at unprecedented rates to meet the growing demand for...

read
Jan 16, 2025

How edge computing + infrastructure arbitrage are redefining AI by optimizing global compute workloads

AI edge computing is driving increased cloud usage rather than replacing it, as revealed by new research from Hong Kong University of Science and Technology and Microsoft Research Asia showing the intricate dependencies between cloud and edge infrastructure. Key research findings: The study utilized a three-layer architecture consisting of Azure cloud servers, GeForce RTX 4090 edge servers, and Jetson Nano client devices to analyze the relationship between edge and cloud computing. Testing revealed that edge-only inference struggled with low bandwidth, while client-only processing couldn't handle complex tasks A hybrid approach combining edge and cloud resources proved most effective, maintaining performance...

read
Jan 16, 2025

Insatiable demand for AI services is fueling a new industry of custom chips

Custom silicon chips and artificial intelligence are driving a fundamental shift in data center computing, with custom semiconductor solutions emerging as a critical strategy for managing escalating power consumption and performance demands. Current state and challenges: The rapid growth of AI services is creating unprecedented demands on data center infrastructure, with AI power consumption expected to increase by 44.7% annually through 2028. Data center power consumption is projected to reach 857 terawatt hours by 2028, equivalent to the electricity usage of a nation ranking just behind Japan Traditional approaches to chip design and power management are struggling to keep pace...

read
Jan 14, 2025

Building Colossus: Inside the AI supercomputer Supermicro is building for xAI

A cutting-edge AI supercomputer, claimed to be the world's largest liquid-cooled GPU cluster, is being built through a partnership between xAI, Supermicro, and NVIDIA in Memphis, Tennessee. Project scale and specifications: The supercomputer installation represents an unprecedented deployment of computing power in terms of both size and speed. The system incorporates over 100,000 NVIDIA HGX H100 GPUs across four compute halls, with each hall housing approximately 25,000 GPUs The facility includes exabytes of storage capacity (an exabyte is equivalent to one billion gigabytes) The entire system was transformed from an empty building to a functioning AI supercomputer in just 122...

read
Jan 10, 2025

DIMON AI outperforms supercomputers in solving complex equations

A new AI framework called Diffeomorphic Mapping Operator Learning (DIMON) can solve complex partial differential equations faster on a personal computer than traditional methods using supercomputers. Key innovation: DIMON represents a significant advancement in computational methods by efficiently solving partial differential equations (mathematical formulas that model how forces, fluids, or other factors interact with different materials and shapes) across multiple geometries. The framework can handle diverse engineering challenges, from predicting air movement around airplane wings to analyzing building stress and car crash deformations Traditional methods require substantial computing power and time to process these complex calculations DIMON achieves superior results...

read
Jan 8, 2025

AI and crypto together will mean a huge electric bill for the internet

Key developments: The integration of AI into everyday internet functions and the continuous operation of cryptocurrency mining are creating unprecedented demands on the power grid. A single ChatGPT prompt consumes approximately 10 times more electricity than a traditional Google search Generating one AI image requires energy equivalent to a full smartphone charge Major tech companies like Google and Amazon are increasingly incorporating AI assistants into their services through products like Gemini and Rufus Technical context: AI systems' energy consumption stems from their complex computational requirements and the physical infrastructure needed to support them. AI tools rely on large language models...

read
Jan 4, 2025

Google and Synaptics partner to boost Edge AI in IoT devices

Synaptics and Google have entered a pivotal collaboration to advance edge AI computing for Internet of Things (IoT) devices, combining their respective strengths in hardware and machine learning capabilities. The core partnership: Synaptics and Google announced a strategic collaboration aimed at optimizing multimodal processing for context-aware computing in IoT devices. The initiative combines Google's MLIR-compliant machine learning core with Synaptics' Astra AI-Native compute platform This integration leverages open-source software and tools to accelerate AI device development The partnership focuses on supporting multiple input types including vision, image, voice, and sound processing Technical implementation: The collaboration centers on Synaptics' Astra platform,...

read
Jan 3, 2025

The biggest breakthroughs in small language models in 2024

The emergence of smaller, more efficient language models marks a significant shift in AI development, with major tech companies now focusing on creating compact versions of their flagship AI systems. The evolution of AI models: The AI industry's initial focus on larger models, sparked by OpenAI's GPT-3 in 2020, is giving way to a more nuanced approach that prioritizes efficiency and specialized performance. OpenAI research scientist Noam Brown emphasized at TEDAI San Francisco that scale has been the primary driver of AI progress over the past five years Researchers have discovered that smaller, more focused models can match or exceed...

read
Dec 31, 2024

OpenAI’s new AI model charges up to $1,000 per query

OpenAI's latest AI model o3 demonstrates significant performance improvements but comes with unprecedented computational costs, reaching over $1,000 per query for complex tasks. Key developments: OpenAI's new o3 model employs a "test-time compute" technique that allows it to spend more time processing and exploring multiple possibilities before providing answers. The model achieved an impressive 87.5 percent score on the ARC-AGI benchmark, nearly triple the performance of its predecessor o1's 32 percent O3 utilizes a unique "high-compute mode" that enables more thorough analysis of complex problems A "mini" version of o3 is scheduled for release in January Cost implications: The computational...

read
Dec 30, 2024

AI data centers are growing to staggering sizes

AI data centers are rapidly expanding to unprecedented physical sizes, with some facilities now stretching over 1.24 miles in length, according to Ciena CEO Gary Smith. Scale of expansion: The growing demand for GPU-powered AI computing is pushing data centers to expand both horizontally and vertically, creating massive multi-story facilities that challenge traditional networking approaches. Cloud providers are building data centers that span more than two kilometers in length These facilities are increasingly utilizing multi-story designs, adding vertical scale to horizontal sprawl Current campuses are blurring traditional boundaries between wide-area networks and data center infrastructure Power requirements: The energy demands...

read
Dec 29, 2024

The case against LLMs in software development

Software industry veteran offers a critical analysis of Large Language Models and the degradation of software quality over time. The core argument: The rise of Large Language Models (LLMs) represents a concerning shift in computing, where corporations prioritize profit over software quality and user experience. Historical context: Earlier software development emphasized different priorities and characteristics compared to today's landscape: Programs were faster and more efficient despite limited hardware capabilities Quality control was paramount due to the difficulty of distributing patches Software was typically standalone, purchasable, and didn't require internet connectivity Applications were simpler, focused on specific use cases, and supported...

read
Dec 29, 2024

NeoClouds and the growing trend of renting GPUs

Investors have poured $20 billion into companies specializing in GPU rental services for AI computing, marking a significant shift in cloud infrastructure investments. Market Evolution: The cloud computing landscape has transformed dramatically since 2014, with Graphics Processing Units (GPUs) evolving from gaming hardware to essential components for artificial intelligence development. Vultr, a Florida-based startup, recently secured $333 million in funding from AMD and LuminArx Capital, achieving a $3.5 billion valuation The company has established over 30 data centers globally, competing with major cloud providers like AWS, Google Cloud, and Microsoft Azure GPU rentals have become Vultr's primary growth driver, reflecting...

read
Dec 25, 2024

Microsoft to expand its global infrastructure to meet growing AI demand

Microsoft's cloud computing chief, Scott Guthrie, announces major expansion and restructuring of the company's artificial intelligence infrastructure to meet growing demand for AI services. Key context: Microsoft is significantly expanding its AI computing capacity while simultaneously restructuring its cloud operations to better support advanced AI workloads. The company plans to double its AI-specific computing infrastructure by the end of 2024 This expansion includes new specialized hardware installations across multiple global regions The restructuring will integrate AI capabilities more deeply into existing cloud services Investment details: The expansion represents one of Microsoft's largest infrastructure investments to date, though specific financial figures...

read
Dec 25, 2024

Regulatory uncertainty and AI made 2024 a bad year for sustainability

Key challenges facing sustainability: 2024 witnessed unprecedented setbacks in corporate environmental and social governance initiatives, particularly in the United States. Political turbulence and contentious elections created uncertainty around environmental regulations and climate action policies Many companies either withdrew from or remained silent about their diversity, equity, and inclusion (DEI) and environmental, social, and governance (ESG) commitments The implementation of new sustainability reporting requirements temporarily diverted corporate resources from actual sustainability initiatives Mixed signals in clean energy transition: Despite overall challenges, the clean energy sector showed signs of reaching critical mass while facing new obstacles. The clean economy achieved several important...

read
Dec 18, 2024

How edge AI and 5G will power a new generation of Industry 4.0 apps

Rapid adoption of edge computing and private wireless networks is driving industrial digital transformation, with Nokia executives highlighting how these technologies are enabling more sophisticated Industry 4.0 applications. Key market trends; According to Nokia's Industrial Digitalization report, 40% of organizations deploying private wireless networks have also implemented on-premises edge computing capabilities to support their digital transformation initiatives. Industrial enterprises are increasingly leveraging AI beyond basic analytics to create business value across multiple use cases The combination of edge computing and mission-critical connectivity is emerging as a foundational platform for industrial digitalization Companies are prioritizing operational technology factors including reliability, accuracy,...

read
Dec 18, 2024

MIT engineers are building 3D ‘high-rise’ AI chips for faster computing

MIT engineers are unveiling a groundbreaking method for creating multilayered 3D computer chips that could transform computing capabilities, particularly for artificial intelligence applications. Technical breakthrough; MIT researchers have developed a novel approach to stack multiple layers of semiconducting materials directly on top of each other, creating what they call "high-rise" chips. The process operates at relatively low temperatures (below 400°C), ensuring the preservation of underlying circuitry while enabling vertical chip construction Using innovative "seed pockets" at mask opening edges, the team successfully grew single-crystalline 2D materials called transition-metal dichalcogenides (TMDs) The researchers demonstrated the technique by creating alternating layers of...

read
Dec 17, 2024

Verizon and NVIDIA partner on 5G, mobile edge computing initiative

Verizon is expanding its enterprise offerings by combining private 5G networks, edge computing, and NVIDIA's AI capabilities to deliver advanced artificial intelligence services for businesses. The core innovation: Verizon has partnered with NVIDIA to develop an integrated solution that merges private 5G networks, Mobile Edge Computing (MEC), and NVIDIA AI Enterprise software platform. The new platform is designed to support demanding computing needs, including generative AI models, computer vision, and augmented reality applications The infrastructure is built to be plug-and-play and can support multiple users or use cases simultaneously Services can be deployed either on-premise or through portable private network...

read
Dec 16, 2024

How Arm boosts AI performance with its new chip designs

The advancement of AI and machine learning technologies has positioned Arm as a key enabler of AI acceleration across diverse computing platforms, leveraging its decade-long expertise in processor architecture and optimization. Core technology capabilities: Arm's processor architecture incorporates multiple technologies that enable efficient AI workload processing without dedicated neural processing units (NPUs). Matrix multiplication, the fundamental operation in AI processing, is accelerated through built-in hardware features in both Arm v8 and v9 architectures Technologies like Neon, Scalable Vector Extensions (SVE), and Scalable Matrix Extensions (SME) enable CPUs to perform accelerated matrix operations independently These capabilities span across Arm's processor families,...

read
Dec 15, 2024

STMicro unveils edge AI microcontroller for smart devices

The rapid advancement of artificial intelligence is driving innovation in microcontroller technology, with major semiconductor manufacturers developing specialized hardware for running AI applications directly on devices rather than in the cloud. Major industry development: STMicroelectronics, a leading European semiconductor company, has introduced its STM32N6 series of microcontrollers specifically designed for edge AI and machine learning applications. The new microcontroller series represents STMicroelectronics' first entry into the edge AI computing market These chips are intended for both consumer and industrial electronics applications, particularly focusing on image and audio processing tasks The microcontrollers are designed to perform computations that traditionally required more...

read
Dec 15, 2024

Quantum computers may produce results humans can never truly verify

The emergence of quantum computers capable of solving previously insurmountable computational problems marks a significant shift in human-machine relationships, challenging traditional notions of knowledge verification and understanding. The quantum leap forward: Google's quantum computer, Willow, can reportedly solve problems in minutes that would take conventional supercomputers billions of years to process. The system tackles calculations that would require approximately 10 septillion years for traditional supercomputers to complete This breakthrough demonstrates computational capabilities that exceed the age of the universe itself The achievement, while remarkable, creates a paradox where the results cannot be verified through conventional means Technical foundations: Quantum computing...

read
Dec 14, 2024

How photonic and quantum breakthroughs are ushering in new computing era

Computing hardware is reaching new frontiers with breakthrough developments in photonic and quantum technologies that could fundamentally change how computers process information. Key innovations in photonic computing: MIT researchers have developed a groundbreaking photonic chip that processes both matrix multiplication and non-linear operations on a single platform, marking a significant advance in optical computing technology. The chip uses light instead of traditional electrical circuits to process data, resulting in lower energy consumption and improved efficiency By creating specialized non-linear optical function units (NOFUs), researchers enabled data to remain in the optical domain throughout processing The system demonstrates low latency, reduced...

read
Dec 11, 2024

Michigan state board approves $100M supercomputing lab for U-M

The state of Michigan is making significant investments in advanced technology and manufacturing, with major funding approvals for a cutting-edge supercomputing facility and electric vehicle production expansion. Major investment details: The Michigan Strategic Fund board has approved approximately $100 million for a University of Michigan (U-M) supercomputing lab and nearly $28 million for Detroit Diesel Corporation's electric vehicle components expansion. The U-M supercomputing project, partnering with Los Alamos National Laboratory, represents a total investment of $785 million Detroit Diesel's funding will support a $285 million expansion of electric component manufacturing for medium and heavy-duty electric trucks Both projects require final...

read
Dec 11, 2024

EU invests €1.5B to create 7 new AI factories across Europe

The European Union is dramatically expanding its artificial intelligence capabilities with a EUR 1.5 billion investment in seven AI Factories across the continent, marking a significant step in its strategy to compete globally in AI development and applications. Strategic Investment Overview; The European High Performance Computing Joint Undertaking (EuroHPC) has selected seven locations across Europe to establish AI Factories, with funding split between EU programs and national sources. The initiative aims to double Europe's AI computing capacity through a combination of new facilities and upgraded existing infrastructure Five locations will receive entirely new AI-optimized supercomputers, while two sites will upgrade...

read
Dec 11, 2024

Cloud trends 2024: Serverless, sovereign and AI-enabled

Public cloud platforms are experiencing significant transformation through artificial intelligence integration, serverless architecture adoption, and growing emphasis on digital sovereignty across different global markets. Major market shifts: The public cloud platform landscape is being reshaped by three primary forces: artificial intelligence adoption, serverless-first approaches, and regional sovereignty requirements. Hyperscalers are adapting their core infrastructure to support generative AI capabilities while expanding their service offerings beyond traditional enterprise IT boundaries Chinese cloud providers are driving innovation in AI services and foundation model support across multiple domains European cloud providers are capitalizing on sovereignty and sustainability requirements to compete effectively across the...

read
Load More