Google’s Gemini AI platform is receiving significant updates coinciding with Samsung’s S25 launch, introducing action chaining capabilities and enhanced multimodal features.
Key Updates: Gemini’s latest improvements focus on interconnected actions and expanded device compatibility, particularly for Samsung’s newest phones and Google Pixel devices.
- Action chaining now enables users to perform sequential tasks across different apps, such as finding restaurants in Google Maps and drafting invitation texts in Messages
- The feature depends on app-specific extensions, with Google and Samsung apps being among the first to support this functionality
- Implementation requires developer-written extensions to connect individual apps with Gemini
Multimodal Enhancements: Gemini Live is expanding its conversational capabilities to include multimedia interactions on select devices.
- Users can now upload images, files, and YouTube videos directly into Gemini conversations
- The system can analyze visual content and provide feedback or suggestions
- These features are exclusively available on Galaxy S24, S25, and Pixel 9 devices
Project Astra Integration: Google’s prototype AI assistant is set to debut in the coming months, bringing advanced environmental interaction capabilities.
- The system allows users to interact with their surroundings through their phone’s camera
- Users can point their devices at objects or locations to receive relevant information
- Project Astra will initially launch on Galaxy S25 and Pixel phones
- The technology is designed to work with Google’s upcoming AI glasses, enabling hands-free interactions
Market Context: The development signals Google’s strategic positioning in the evolving AI wearables market.
- Google is preparing to compete with Meta’s Ray-Ban smart glasses
- The release date for Google’s AI glasses remains unannounced
- These developments represent a significant step toward more intuitive AI interactions in daily life
Looking Forward: While these updates mark substantial progress in AI assistance capabilities, the success of features like action chaining will largely depend on developer adoption and the creation of compatible extensions across popular apps. The integration with future wearable technology could particularly impact how users interact with AI in their daily lives.
Google has just announced the ability to chain actions in Gemini and it could change the way we use AI for good