Loading...
「ツール」は右上に移動しました。
利用したサーバー: wtserver1
2いいね 188 views回再生

Testing Gemini 2.0's Multimodal Live API with Video in Google AI Studio

2:05 - TLDR
Gemini processes text, audio, and video inputs, delivering real-time text and audio outputs

Quick Start to test it:⭐
Navigate to Google AI Studio.
Obtain your API Key from the dashboard.
Click on Stream Realtime.
Select the Video Chat option to initiate a live multimodal interaction with Gemini.
-----------------------------------------------------------------------------------------------------------------------------------------
More detailed instructions on how to test it and use it in projects:

https://ai.google.dev/gemini-api/docs...

Getting Started with the Multimodal Live API using Gen AI SDK:

https://github.com/GoogleCloudPlatfor...
----------------------------------------------------------------------------------------------------------------------------------------
The Multimodal Live API enables developers to implement low-latency, bidirectional voice and video interactions with Gemini, providing natural, human-like conversations with voice-interrupt capabilities.
---------------------------------------------------------------------------------------------------------------------------------------
🐶🖥️
This video demonstrates Gemini’s advanced multimodal capabilities by correctly identifying a broken external monitor through its vision and voice-driven interaction features. Gemini visually recognizes the device, reads small identifying text, and guides the user through replacement instructions—highlighting practical applications for developers.

Explore the Multimodal Live API today in Google AI Studio.

コメント