Google Gemini is a next-generation AI model developed by Google's AI research labs DeepMind and Google Research1. It is designed to be multimodal, meaning it can generalize and seamlessly understand, operate across, and combine different types of information, including text, code, audio, image, and video. Gemini is available in three different sizes: Gemini Ultra, Gemini Pro, and Gemini Nano, each optimized for different tasks and devices.
Some primary capabilities of Google Gemini mentioned in the news content include:
Sophisticated multimodal reasoning: Gemini can handle complex reasoning tasks, such as understanding and interpreting information from various sources, including text, images, and audio3.
Advanced coding capabilities: Gemini can understand, explain, and generate high-quality code in popular programming languages like Python, Java, C++, and Go.
Flexibility to run on different devices: Gemini's three size variants (Ultra, Pro, and Nano) allow it to efficiently run on everything from data centers to mobile devices3.
Personal health advice: A version of Gemini, called Personal Health Large Language Model (PH-LLM), has been fine-tuned to understand and reason on time-series personal health data from wearables such as smartwatches and heart rate monitors. This model can provide sleep and fitness advice more accurately than human experts.
Integration with Google products and services: Gemini is integrated into various Google products, including Bard, Pixel 8 Pro, Search, Ads, Chrome, and Duet AI, to enhance their capabilities and provide a more seamless user experience.
Customizability: Gemini Advanced subscribers can create customized versions of Gemini called "Gems" to cater to specific needs and preferences, such as a gym buddy, sous chef, coding partner, or creative writing guide.
Speech technology: Gemini Live, a new mobile conversational experience, uses state-of-the-art speech technology to make speaking with Gemini more intuitive and natural, allowing users to talk to Gemini and choose from a variety of natural-sounding voices for responses.
Wearable devices collect various types of data for health monitoring that PH-LLM analyzes. Some specific types of data collected include:
These data are collected passively and continuously, providing a rich source of information for personal health monitoring. PH-LLM utilizes this data to generate insights, identify potential causes for observed behaviors, and provide recommendations to improve sleep hygiene and fitness outcomes4.
The Personal Health Large Language Model (PH-LLM) utilizes data from wearables to provide health advice by analyzing and reasoning over numerical time-series personal health data. This data is collected from wearable devices such as smartwatches and heart rate monitors, which provide a rich and continuous source of information for personal health monitoring. The model is trained to make recommendations, answer professional examination questions, and predict self-reported sleep disruption and results of sleep impairment. By processing and interpreting the data from wearables, PH-LLM can offer personalized insights and recommendations for improving sleep hygiene and fitness outcomes.