News/Open-source
SambaNova Announces Challenger to OpenAI’s o1 Model with New Demo
SambaNova challenges OpenAI with high-speed AI demo: SambaNova Systems has unveiled a new demo on Hugging Face, showcasing a fast, open-source alternative to OpenAI's o1 model using Meta's Llama 3.1 Instruct model. The demo, powered by SambaNova's SN40L chips, allows developers to interact with the 405B parameter Llama 3.1 model, achieving speeds of 405 tokens per second. This release represents a significant step in SambaNova's efforts to compete in the enterprise AI infrastructure market, challenging both OpenAI and hardware providers like Nvidia. The demo emphasizes speed and efficiency, which are crucial for practical business applications of AI technology. Open-source vs....
read Sep 11, 2024Mistral AI Unveils Pixtral 12B, Its First Multimodal AI Model
Mistral AI expands into multimodal AI: Mistral AI, the French AI startup, has released Pixtral 12B, its first multimodal AI model combining language and vision processing capabilities. The model is not yet available on the public web, but its source code can be downloaded from Hugging Face or GitHub for testing on individual instances. Mistral initially released the model through a torrent link, continuing its unconventional approach to AI model releases. Sophia Yang, head of developer relations at Mistral, announced that the model will soon be available through the company's web chatbot and La Platforme API. Key features of Pixtral...
read Sep 11, 2024The Power of Open-Source in AI Implementation
Artificial intelligence (AI) adoption is rapidly increasing in organizations, with generative AI usage nearly doubling in less than a year according to a recent McKinsey survey. This surge in AI implementation comes amid shifting enterprise priorities and an increasingly complex technology landscape. The evolving tech landscape: Businesses are grappling with cloud migration, enhanced security needs, legacy IT modernization, and workflow digitization, all of which contribute to a more intricate technological environment. Complexity in technology can hinder decision-making, increase risks, and negatively impact customer loyalty and employee engagement. There is a growing need for organizations to navigate this complexity confidently to...
read Sep 10, 2024Arcee AI Launches SuperNova, A 70B AI Model for Enterprise Data Control
SuperNova, a new 70 billion parameter language model designed for enterprise deployment, has been unveiled by Arcee AI. This model aims to provide a customizable, instruction-adherent alternative to cloud-based AI services, addressing key enterprise concerns such as data privacy, model stability, and customization. Technical innovations and development process: SuperNova is built on Meta's Llama-3.1-70B-Instruct architecture and employs a novel post-training process to enhance its capabilities. The development involved training three models simultaneously, including one distilled from Llama 405B and another trained with Arcee's EvolKit-generated dataset. A proprietary merging technique combines the strengths of these models, resulting in advanced instruction-following capabilities....
read Sep 10, 2024Global AI Robotics Advancements To Showcase at ROSCon Events
Global robotics community gathers for ROSCon events: The Robot Operating System (ROS) ecosystem is set to take center stage at three upcoming ROSCon conferences, highlighting the latest advancements in AI robotics development. Event details and locations: ROS developers and enthusiasts will have multiple opportunities to engage with the community and explore cutting-edge technologies. ROSCon events are scheduled in Tokyo (September 25), Odense, Denmark (October 21-23), and China (November 2-3, location to be confirmed) These conferences aim to bring together robotics developers of all skill levels for learning, networking, and collaboration NVIDIA's role in advancing ROS ecosystem: The tech giant is...
read Sep 10, 2024DeepSeek-V2.5 Advances Open-Source AI With Powerful Language Model
Breakthrough in open-source AI: DeepSeek, a Chinese AI company, has launched DeepSeek-V2.5, a powerful new open-source language model that combines general language processing and advanced coding capabilities. DeepSeek-V2.5 was released on September 6, 2024, and is available on Hugging Face with both web and API access. The model is optimized for writing, instruction-following, and coding tasks, introducing function calling capabilities for external tool interaction. It outperforms its predecessors in several benchmarks, including AlpacaEval 2.0 (50.5 accuracy), ArenaHard (76.2 accuracy), and HumanEval Python (89 score). In internal Chinese evaluations, DeepSeek-V2.5 surpassed GPT-4o mini and ChatGPT-4o-latest. Expert recognition and praise: The new...
read Sep 7, 2024Capx AI Launches 8B-Parameter Multimodal Vision Model
Groundbreaking multimodal AI model unveiled: Capx AI has released Llama-3.1-vision, an 8 billion parameter Vision model that combines Meta AI's Llama 3.1 8B language model with the SigLIP vision encoder. The model, released under the Apache 2.0 License, is designed to excel in instruction-following tasks and create rich visual representations. Built upon BAAI's Bunny repository, the architecture consists of a vision encoder, a connector module, and a language model. The model leverages Low-Rank Adaptation (LoRA) for efficient training on limited computational resources. Innovative two-stage training approach: The development process involved a pretraining stage to align visual and text embeddings, followed...
read Sep 6, 2024Yi-Coder Has Launched and Wants to be Your New Coding Buddy
Yi-Coder's launch by 01.AI marks a significant shift in AI-assisted coding, challenging the notion that bigger models are always better and potentially democratizing access to advanced AI tools for developers worldwide. A new paradigm in AI coding assistance: Yi-Coder, released by 01.AI, offers state-of-the-art coding performance with a surprisingly compact model of fewer than 10 billion parameters. The AI coding assistant is available in 9 billion and 1.5 billion parameter versions, capable of handling tasks across 52 programming languages. Yi-Coder's 128,000 token context length allows it to process massive code snippets, potentially revolutionizing work on complex, large-scale projects. The tool...
read Sep 5, 2024Hugging Face Partners with Truffle Security to Protect Code Repositories
Hugging Face bolsters security with TruffleHog integration: Hugging Face has partnered with Truffle Security to incorporate TruffleHog's secret scanning capabilities into its platform, enhancing security measures for users and developers. Key partnership details: The collaboration between Hugging Face and Truffle Security aims to prevent accidental leaks of sensitive information in code repositories. TruffleHog is an open-source tool that detects and verifies secret leaks in code, scanning for credentials, tokens, and encryption keys. The partnership focuses on two main initiatives: enhancing Hugging Face's automated scanning pipeline and creating a native Hugging Face scanner in TruffleHog. Automated scanning pipeline improvements: Hugging Face...
read Sep 4, 2024Meta’s Llama Models Have Been Downloaded Nearly 350 Million Times This Year
Open-source AI surge led by Meta: Meta's Llama family of AI models has experienced explosive growth, with downloads approaching 350 million on Hugging Face, representing a tenfold increase from the previous year. Meta's mid-year update reveals significant adoption of its open-source AI models, particularly among large enterprises such as Zoom, Spotify, Infosys, AT&T, and Goldman Sachs. The performance of open-source AI is now matching that of closed-source alternatives, gaining substantial traction at the enterprise level. This trend is reshaping the AI landscape, putting pressure on closed-model companies to innovate and reduce costs. Meta's strategic open approach: The company's decision to...
read Sep 4, 2024Yi-Coder is a Small But Mighty Open-Source LLM for Coding
Yi-Coder, a new series of open-source code language models, has emerged as a powerful tool for developers, offering state-of-the-art coding performance with fewer than 10 billion parameters. Model overview and key features: Yi-Coder is available in two sizes—1.5B and 9B parameters—with both base and chat versions designed for efficient inference and flexible training. The models are built upon a foundation of 2.4 trillion high-quality tokens sourced from GitHub repositories and filtered code-related data from CommonCrawl. Yi-Coder supports a maximum context window of 128K tokens, enabling project-level code comprehension and generation. The 9B parameter version outperforms similar-sized models and even rivals...
read Aug 27, 2024China-Based CogVideoX Launches Open-Source AI Video Model
Groundbreaking open-source AI for video generation: CogVideoX, a new text-to-video model developed by researchers from Tsinghua University and Zhipu AI, has the potential to revolutionize video creation and democratize advanced AI capabilities. The open-source model generates high-quality, coherent videos up to six seconds long from text prompts, rivaling proprietary systems from companies like Runway, Luma AI, and Pika Labs. CogVideoX-5B, the most advanced version, boasts 5 billion parameters and produces 720x480 resolution videos at 8 frames per second. The model outperforms competitors like VideoCrafter-2.0 and OpenSora across multiple metrics, according to the researchers' benchmarks. Technical innovations driving performance: The researchers...
read Aug 27, 2024Hugging Face Releases New Tutorials to Build AI-Powered Robots
Hugging Face democratizes robotics development: The AI community platform has released comprehensive tutorials aimed at empowering developers to build and train AI-powered robots, marking a significant step towards making robotics more accessible. The tutorials cover a wide range of topics, from sourcing components to deploying AI models, enabling users to train neural networks for controlling low-cost robotic hardware. Developers can learn to program neural networks via laptop to control robots, teaching them tasks such as object detection, grasping, and manipulation. The guides are designed to be user-friendly, catering to individuals with varying levels of robotics experience, including information on 3D...
read Aug 27, 2024AI Startup Aleph Alpha Launches Open-Source Language Models
Aleph Alpha's release of open-source AI models signals a shift towards transparent and EU-compliant machine learning, potentially reshaping the landscape of AI development. Breakthrough in open-source AI: German startup Aleph Alpha has unveiled two new large language models (LLMs) under an open license, challenging the closed-source approach of many tech giants. The models, Pharia-1-LLM-7B-control and Pharia-1-LLM-7B-control-aligned, each have 7 billion parameters and are designed to deliver concise, length-controlled responses in multiple European languages. Aleph Alpha claims their performance matches leading open-source models in the 7-8 billion parameter range. The company has also open-sourced its training codebase, called "Scaling," allowing researchers...
read Aug 26, 2024This New AI-Powered Platform Simplifies Health Insurance Appeals
Tech worker's personal mission sparks AI-powered insurance appeal platform: Holden Karau, a San Francisco-based tech professional, has launched Fight Health Insurance, an open-source platform that uses AI to help patients generate health insurance appeals. The genesis of a novel solution: Karau's experience with numerous insurance claim denials for her own healthcare needs led her to develop an innovative approach to a common problem. As a trans woman and motorcycle enthusiast, Karau faced multiple insurance denials for gender-affirming care and accident-related treatments. She successfully appealed over 90% of roughly 40 denials, demonstrating the effectiveness of challenging insurance decisions. This personal experience...
read Aug 25, 2024Europe Risks Falling Behind in AI Race, Tech CEOs Warn
Open-source AI's potential and Europe's regulatory challenge: Mark Zuckerberg of Meta and Daniel Ek of Spotify have jointly advocated for Europe to embrace open-source AI technology while warning about the region's complex regulatory environment. The tech leaders argue that Europe's fragmented regulations could cause the continent to fall behind in the global AI innovation race. They emphasize AI's transformative potential in boosting productivity, accelerating scientific progress, and contributing to the global economy. However, they note that AI benefits are not being equally distributed, with disparities emerging between those with and without access to cutting-edge AI technology. The case for open-source...
read Aug 25, 2024Tech Giants Push for Open-Source AI to Fuel Innovation
AI industry leaders advocate for open-source models: Mark Zuckerberg and Daniel Ek make a compelling case for open-sourcing AI software, particularly in Europe, to prevent power concentration and foster innovation. Zuckerberg and Ek argue that open-sourcing AI models creates a level playing field and ensures power isn't concentrated among a few large players. The approach aligns with Meta's recent shift in priorities, focusing more on AI investments rather than the "metaverse." This stance marks a notable change in perception for Zuckerberg, who has faced criticism for past decisions but is now gaining support for his AI-focused strategy. The future of...
read Aug 23, 2024The Open Source Initiative Creates New Definition for Open-Source
Defining open-source AI: The Open Source Initiative (OSI) has unveiled a new definition for open-source AI systems, aiming to provide clarity in a field where the concept was previously ambiguous. The definition outlines key criteria for AI systems to be considered open-source, including unrestricted use, inspectability, modifiability, and shareability. Transparency requirements extend to training data, source code, and model weights, ensuring a comprehensive understanding of the AI system's components. The definition stipulates that sufficient information must be provided to allow a skilled person to recreate a substantially equivalent system using the same or similar data. Collaborative effort and development process:...
read Aug 21, 2024Microsoft Unveils 3 Powerful New Open-Source AI Models
Breakthrough in AI development: Microsoft has unveiled three new powerful AI models in its Phi series, marking a significant advancement in the field of artificial intelligence and machine learning. The new models, Phi-3.5-mini-instruct, Phi-3.5-MoE-instruct, and Phi-3.5-vision-instruct, are designed for various tasks ranging from basic reasoning to complex vision-related problems. These models are now available on Hugging Face under an MIT License, allowing for both commercial use and modification, which could potentially accelerate AI innovation across various sectors. In benchmark tests, the Phi-3.5 models have demonstrated impressive performance, surpassing some models from tech giants like Google, Meta, and OpenAI in certain...
read Aug 21, 2024OpenAI Announces GPT-4o Fine-Tuning for Developers
OpenAI introduces fine-tuning for GPT-4o: OpenAI has announced the ability for third-party developers to fine-tune custom versions of its latest large multimodal model, GPT-4o, enhancing its applicability for specific applications or organizational needs. Key features and benefits: Fine-tuning allows developers to adjust the model's tone, follow specific instructions, and improve accuracy in technical tasks, even with small datasets. Developers can access this feature through OpenAI's fine-tuning dashboard by selecting the gpt-4o-2024-08-06 base model. The company claims strong results can be achieved with as few as a dozen examples in the training data. Promotional offer and pricing: OpenAI is offering up...
read Aug 18, 2024Inside Be.Ta Labs’ 100% Solar-Powered AI Operation
Pioneering sustainable AI: Be.Ta Labs harnesses solar power for eco-friendly AI operations. Be.Ta Labs has made a significant breakthrough in the field of artificial intelligence by powering their entire AI and large language model (LLM) infrastructure exclusively with sustainable solar energy, setting a new standard for environmental responsibility in the tech industry. Environmental impact of AI development: The carbon footprint of artificial intelligence has become a growing concern in recent years, with significant implications for climate change and sustainability efforts. Training a single AI model can potentially emit as much CO2 as five cars over their entire lifetimes, highlighting the...
read Aug 16, 2024An Inside Look at Google’s Gemma Open-Source AI Models
The Gemma model family represents a significant advancement in open-source AI, offering lightweight yet powerful alternatives to larger language models. Introducing Gemma: Gemma is a family of open-source AI models derived from the same research and technology as Google's Gemini models, designed to be lightweight and state-of-the-art for various applications. Gemma models are built to cater to different use cases and modalities, offering flexibility for developers and researchers. The family includes variations like Gemma 1, CodeGemma, Gemma 2, RecurrentGemma, and PaliGemma, each optimized for specific tasks. All Gemma models utilize a decoder-only Transformer architecture, building on proven techniques in natural...
read Aug 16, 2024AI Model Hermes 3 Shows Advanced Skills and Unexpected Behavior
Hermes 3, a powerful new open-source AI model developed by Lambda and Nous Research, demonstrates advanced capabilities while exhibiting unusual existential crises when given blank prompts. Model overview and development: Hermes 3 is a fine-tuned version of Meta's open-source Llama 3 large language model, created through a collaboration between AI infrastructure company Lambda and Nous Research. The model was developed across three parameter sizes: 8 billion, 70 billion, and 405 billion. Hermes 3 is based on Meta's Llama 3.1-405 billion parameter model, representing a significant advancement in open-source AI technology. Impressive capabilities: Hermes 3 showcases a range of powerful text-based...
read Aug 15, 2024Chinese Researchers Create AI Model That Generates 10,000-Word Texts
Breakthrough in AI-generated content: Researchers at Tsinghua University in Beijing have developed an AI system capable of producing coherent texts exceeding 10,000 words, challenging the boundaries of machine-generated writing. The system, named "LongWriter," is detailed in a paper titled "LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs." This development addresses the longstanding challenge of generating extensive, high-quality written content using artificial intelligence. The research team discovered a correlation between an AI model's output length and the length of texts it encounters during its training phase. Technical innovations: The LongWriter system incorporates novel approaches to enhance AI's capacity for long-form...
read