-
Notifications
You must be signed in to change notification settings - Fork 2.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: Add intro to multimodal live api notebook #1528
base: main
Are you sure you want to change the base?
Conversation
Check out this pull request on See visual diffs & provide feedback on Jupyter Notebooks. Powered by ReviewNB |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @gericdong and reviewers,
I'm currently reviewing this pull request and will post my detailed review in a few minutes. In the meantime, here's a quick summary to help everyone get up to speed:
This PR adds a new introductory Jupyter notebook (gemini/multimodal-live-api/intro_multimodal_live_api.ipynb
) to the generative-ai
repository. The notebook demonstrates basic usage of the Multimodal Live API, showcasing text-to-text generation, text-to-audio generation, and text-to-audio conversation. The notebook includes setup instructions, authentication (for Colab users), and clear examples for each functionality. It also provides links to open the notebook in various environments (Colab, Colab Enterprise, Vertex AI Workbench, and GitHub).
The changes are comprehensive, adding a complete, well-structured notebook with code examples and explanations. The notebook is well-documented and easy to follow, making it a valuable resource for users wanting to learn about the Multimodal Live API.
Happy reviewing!
Code whispers secrets,
Lines of logic entwine,
Meaning starts to bloom.
Gemini encountered an error creating the review. You can try again by commenting |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Minor suggestions
"source": [ | ||
"## Use the Multimodal Live API\n", | ||
"\n", | ||
"Multimodal Live API is a stateful API that uses WebSockets. This section shows some basic examples of how to use Multimodal Live API for text-to-text and text-to-audio generation." |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"Multimodal Live API is a stateful API that uses WebSockets. This section shows some basic examples of how to use Multimodal Live API for text-to-text and text-to-audio generation." | |
"Multimodal Live API is a stateful API that uses [WebSockets](https://en.wikipedia.org/wiki/WebSocket). This section shows some basic examples of how to use Multimodal Live API for text-to-text and text-to-audio generation." |
gemini/multimodal-live-api/intro_multimodal_live_api_genai_sdk.ipynb
Outdated
Show resolved
Hide resolved
gemini/multimodal-live-api/intro_multimodal_live_api_genai_sdk.ipynb
Outdated
Show resolved
Hide resolved
Co-authored-by: Holt Skinner <[email protected]>
….ipynb Co-authored-by: Holt Skinner <[email protected]>
….ipynb Co-authored-by: Holt Skinner <[email protected]>
Add intro to multimodal live api notebook to demonstrate the some basic examples such as