A comprehensive Rust client library for Google's Gemini 2.5 API.
- π Complete Gemini 2.5 API Implementation - Full support for all Gemini API endpoints
- π οΈ Function Calling & Tools - Custom functions and Google Search integration with OpenAPI schema support
- π¦ Batch Processing - Efficient batch content generation and embedding
- πΎ Content Caching - Cache system instructions and conversation history for cost optimization
- π Streaming Responses - Real-time streaming of generated content
- π§ Thinking Mode - Support for Gemini 2.5 thinking capabilities
- π¨ Image Generation - Text-to-image generation and image editing capabilities
- π€ Speech Generation - Text-to-speech with single and multi-speaker support
- πΌοΈ Multimodal Support - Images and binary data processing
- π Text Embeddings - Advanced embedding generation with multiple task types
- βοΈ Highly Configurable - Custom models, endpoints, and generation parameters with HTTP client builder
- π Type Safe - Comprehensive type definitions with full
serde
support - β‘ Async/Await - Built on
tokio
for high-performance async operations - π Comprehensive Tracing - Built-in structured logging and telemetry with
tracing
for observability
Add this to your Cargo.toml
:
[dependencies]
gemini-rust = "1.5.1"
Get started with simple text generation, system prompts, and conversations. See basic_generation.rs
for complete examples including simple messages, system prompts, and multi-turn conversations.
Enable real-time content streaming for interactive applications. See basic_streaming.rs
for examples of processing content as it's generated with immediate display.
The library provides comprehensive access to all Gemini 2.5 capabilities through an intuitive Rust API:
Advanced reasoning capabilities with thought process visibility and custom thinking budgets. See thinking_basic.rs
and thinking_advanced.rs
.
- Custom function declarations with OpenAPI schema support (using
schemars
) - Google Search integration for real-time information
- Type-safe function definitions with automatic schema generation
- See
tools.rs
andcomplex_function.rs
- Image Generation: Text-to-image with detailed prompts and editing capabilities
- Speech Generation: Text-to-speech with single and multi-speaker support
- Image Processing: Analyze images, videos, and binary data
- See
image_generation.rs
andmulti_speaker_tts.rs
Efficient processing of multiple requests with automatic file handling for large jobs. See batch_generate.rs
.
Cache system instructions and conversation history to reduce costs and improve performance. See cache_basic.rs
.
Advanced embedding generation with multiple task types for document retrieval and semantic search. See embedding.rs
.
Real-time streaming of generated content for interactive applications. See streaming.rs
.
- Custom models and endpoints
- Detailed generation parameters (temperature, tokens, etc.)
- HTTP client customization with timeouts and proxies
- See
generation_config.rs
andcustom_base_url.rs
Built-in structured logging and telemetry with tracing
for comprehensive monitoring and debugging.
Configure different Gemini models including Flash, Pro, Lite, and custom models. See custom_models.rs
for examples of all model configuration options including convenience methods, enum variants, and custom model strings.
Use custom API endpoints and configurations. See custom_base_url.rs
for examples of configuring custom endpoints with different models.
For advanced HTTP configuration (timeouts, proxies, custom headers), use the builder pattern. See http_client_builder.rs
for a complete example with custom timeouts, user agents, connection pooling, and proxy configuration.
The library is instrumented with the tracing
crate to provide detailed telemetry data for monitoring and debugging. This allows you to gain deep insights into the library's performance and behavior.
Key tracing features include:
- HTTP Request Tracing: Captures detailed information about every API call, including HTTP method, URL, and response status, to help diagnose network-related issues
- Token Usage Monitoring: Records the number of prompt, candidate, and total tokens for each generation request, enabling cost analysis and optimization
- Structured Logging: Emits traces as structured events, compatible with modern log aggregation platforms like Elasticsearch, Datadog, and Honeycomb, allowing for powerful querying and visualization
- Performance Metrics: Provides timing information for each API request, allowing you to identify and address performance bottlenecks
To use these features, you will need to integrate a tracing
subscriber into your application. See tracing_telemetry.rs
for comprehensive examples including basic console logging, structured logging for production, and environment-based log level filtering.
The repository includes 30+ comprehensive examples demonstrating all features. See examples/README.md
for detailed information.
basic_generation.rs
- Simple content generation for beginnersbasic_streaming.rs
- Real-time streaming responsessimple.rs
- Comprehensive example with function callingthinking_basic.rs
- Gemini 2.5 thinking modebatch_generate.rs
- Batch content generationimage_generation.rs
- Text-to-image generationgoogle_search.rs
- Google Search integrationurl_context.rs
- URL Context tool for web content analysis
Run any example:
GEMINI_API_KEY="your-api-key" cargo run --example basic_generation
Get your API key from Google AI Studio and set it as an environment variable:
export GEMINI_API_KEY="your-api-key-here"
- Gemini 2.5 Flash - Fast, efficient model (default) -
Model::Gemini25Flash
- Gemini 2.5 Flash Lite - Lightweight model -
Model::Gemini25FlashLite
- Gemini 2.5 Pro - Advanced model with thinking capabilities -
Model::Gemini25Pro
- Text Embedding 004 - Latest embedding model -
Model::TextEmbedding004
- Custom models - Use
Model::Custom(String)
or string literals for other models
Contributions are welcome! Please feel free to submit a Pull Request. For major changes, please open an issue first to discuss what you would like to change.
For guidelines on developing agents and applications, see the Agent Development Guide.
This project is licensed under the MIT License - see the LICENSE file for details.
- Google for providing the Gemini API
- The Rust community for excellent async and HTTP libraries
- Special thanks to @npatsakula for major contributions that made this project more complete
- All contributors who have helped improve this library