Google DeepMind unveiled Gemini Robotics 1.0, an end-to-end multimodal policy model that directly maps vision + language + proprioception inputs to continuous robot actions, demonstrating unprecedented dexterity in household tasks such as folding clothes and opening drawers without task-specific fine-tuning.