Back to List
TechnologyAIInnovationUser Experience

Google Research Unveils Generative UI: AI Now Creates Interactive Interfaces from Simple Prompts, Transforming User Experience in Gemini and Search

Google Research has introduced Generative UI, a groundbreaking interactive technology that enables AI models to generate complete, visual, and interactive user interfaces, including web pages, tools, games, and applications, from natural language prompts. This innovation expands AI's capability beyond mere content generation to full interactive experience creation. Integrated into Gemini App's 'Dynamic View' and Google Search's AI Mode, Generative UI addresses the limitations of traditional AI's linear text output, which struggles with complex knowledge and interactive tasks. The system allows AI to instantly design and implement functional interfaces, such as animated DNA explanations or social media galleries, rather than just providing textual descriptions. This feature is currently experimental in Gemini and available to Google AI Pro and Ultra users in the US for Search's AI Mode, leveraging tool access, system-level instructions, and post-processing for robust and safe interface generation.

Xiaohu.AI 日报

Google Research has announced the launch of Generative UI, a new interactive technology paradigm. This breakthrough allows AI models to not only 'generate content' but also automatically create complete visual and interactive user interfaces, such as web pages, tools, games, and applications. The core concept is that this functionality is now integrated into the Gemini App (Gemini Dynamic View) and Google Search's AI Mode, signifying Google's expansion of generative AI from 'language output' to 'interactive experience generation'.

The traditional AI interaction method has a fundamental limitation: regardless of how powerful a language model is, users ultimately see linear text output. This is highly disadvantageous for presenting complex knowledge, spatial relationships, or interactive tasks. Google's research team posed a question, leading to the concept of Generative UI: an AI system capable of instantly designing and implementing interactive interfaces based on user prompts.

Generative UI is defined as a new AI capability that enables models to generate complete, interactive, visual, and task-oriented user experiences based on natural language prompts. The generated results can include functional web pages, operable tools, visual dashboards, interactive simulation scenarios, or educational/experimental environments. Unlike previous approaches, these interfaces are not pre-designed templates but are instantly generated by AI at the moment a question is posed.

For example, if a user inputs a query about DNA transcription, a traditional AI might only output a descriptive text. With Generative UI, the AI would generate a dynamic page displaying animations of DNA strands and RNA polymerase, color-coding transcription steps, allowing clicks to view cellular differences, and even enabling users to drag sliders to 'observe' the entire process. This illustrates the fundamental difference between 'generating content' and 'generating experiences'.

Generative UI is being experimentally implemented in two core Google platforms:

1. **Gemini App: 'Dynamic View' & Visual Layout Mode**
Gemini 3, through its Agentic Coding capabilities, can generate independent interface logic for each prompt. Features include automatic design of interactive experiences, content level adjustment based on target users (children, adults, professionals), and the ability to generate educational tools, learning games, business presentations, or plan management interfaces. For instance, inputting: 'Help me explain the microbiome so a five-year-old can understand' would lead the AI to generate an educational page with illustrations, animations, and voiceovers. Inputting: 'Create a social media gallery for a brand' would result in a display interface with images and layout. This functionality is currently an experimental feature (Dynamic View and Visual Layout modes).

2. **Google Search: Dynamic Search Experience in AI Mode**
Generative UI is also integrated into Google Search's AI Mode. The search engine no longer just provides text summaries or web links but directly constructs interactive explanatory environments. For example, if a user searches for 'How does photosynthesis work?', the AI will instantly generate dynamic step-by-step animations, molecular structure diagrams, operable molecular models, and mixed text-and-image explanation areas. This feature is currently available to Google AI Pro and Ultra users in the United States and requires activation in the 'Thinking' option within Search's AI Mode.

Google's paper, 'Generative UI: LLMs are Effective UI Generators,' outlines its implementation architecture, which consists of three core components:

1. **Tool Access:** The AI can access a suite of external tools, such as image generation (e.g., Imagen system), search engine results, code execution or data retrieval modules, and graphical drawing simulation environments. The results from these tools can be used by the model to generate higher-quality content or be directly passed to the user's browser to reduce latency, enabling the AI to generate truly functional interfaces, not just text.

2. **System-Level Instructions:** The AI receives a strict set of background system instructions, including the type of interface to generate, code format, design style, and error avoidance specifications. These instructions act as a design blueprint, ensuring that the AI's generated results are functional, clearly structured, and stylistically consistent.

3. **Post-Processing:** After AI output, multiple layers of algorithms perform corrections and safety checks, including verifying code functionality, correcting common errors, maintaining a consistent visual style (e.g., 'Wizard Green' style), and ensuring safe and uncluttered output. Ultimately, the AI outputs a directly usable interface, not just pure text.

This research represents a paradigm shift in Human-Computer Interaction (HCI). Generative UI propels HCI into an 'AI-generated interactive interface' stage. Future interfaces will no longer be pre-designed by human designers but will be instantly generated by AI based on context, entirely centered on user needs. This is not merely a new form of UI but also the starting point for AI moving towards 'environment-generating intelligence,' where AI doesn't just answer questions but constructs an environment for users to understand and explore.

Related News

Technology

Google Unveils Antigravity: A New AI-Powered Autonomous Platform for End-to-End Software Development, Integrating with Gemini 3 for Agentic Coding

Google has launched Antigravity, a novel platform designed for "AI agent-led development," moving beyond traditional IDEs. This autonomous agent collaboration system enables AI to independently plan, execute, and verify complete software development tasks. Deeply integrated with the Gemini 3 model, Antigravity represents Google's key product in "Agentic Coding." It addresses limitations of previous AI tools, which were primarily assistive and required manual operation and step-by-step human prompts. Antigravity allows AI to work across editors, terminals, and browsers, plan complex multi-step tasks, automatically execute actions via tool calls, and self-check results. It shifts the development paradigm from human-operated tools to AI-operated tools with human supervision and collaboration. The platform's core philosophy revolves around Trust, Autonomy, Feedback, and Self-Improvement, providing transparency into AI's decision-making, enabling autonomous cross-environment operations, facilitating real-time human feedback, and allowing AI to learn from past experiences.

Technology

Google Vids Unlocks Advanced AI Features for All Gmail Users: Free Access to AI Voiceovers, Redundancy Removal, and Image Editing

Google has made several advanced AI features in its Vids video editing platform available to all users with a Gmail account, previously exclusive to paid subscribers. These newly accessible tools include AI voiceovers, automatic removal of redundant speech, and AI image editing. The transcription trimming feature automatically eliminates filler words like "um" and "ah," along with long pauses, significantly enhancing video quality. Users can also generate professional-grade voiceovers from text scripts, choosing from seven different voice options, many of which sound natural. Additionally, the AI image editing tool allows for easy modifications such as background removal, descriptive editing, and transforming static photos into dynamic videos. Google aims to empower both beginners and experienced creators to produce high-quality video content, anticipating significant growth in the video editing market despite Vids being in its early stages.

Technology

Quora's Poe AI Platform Launches Group Chat Feature Supporting Up to 200 Users for Enhanced Collaborative AI Interactions

Quora has introduced a new group chat feature for its AI platform, Poe, allowing up to 200 users to collaborate with various AI models and bots in a single conversation. This innovation supports multi-modal interactions including text, image, video, and audio generation. The launch coincides with OpenAI's ChatGPT piloting similar group chat functionalities in select markets, signaling a shift in AI interaction methods. Quora highlights that this feature will offer new interactive experiences for AI users, such as family trip planning using Gemini 2.5 and o3Deep Research, or team brainstorming with image models to create mood boards. Users can also engage in intellectual games with Q&A bots. Group chats can be created from Poe's homepage, with real-time synchronization across devices, ensuring seamless transitions between desktop and mobile. Quora developed this feature over six months and plans to optimize it based on user feedback, emphasizing the unexplored potential for group interaction and collaboration in AI mediums. Poe also enables users to create and share custom bots.