Gemini Integration
How Simili Bot uses Google Gemini for AI analysis.Services
Simili Bot uses Gemini for:- Text Embeddings - Convert text to vectors for semantic search
- LLM Analysis - AI reasoning, duplicate detection, routing, and triage
Embeddings
gemini-embedding-001
Default embedding model in v0.2.0:- Convert issue text to vector for similarity search
- Generate embeddings for all issues during bulk indexing
- Embed PR content (title + body + changed files) for PR duplicate detection
v0.1.0 used
gemini-embedding-001 (768 dimensions). v0.2.0 uses gemini-embedding-001 (3072 dimensions). If you’re migrating from v0.1.0, you must re-index your collection after updating.LLM analysis
Default LLM model:gemini-2.5-flash
1. Duplicate detection
2. Quality assessment
3. Issue routing
4. Auto triage
Models
| Model | Type | Default |
|---|---|---|
gemini-embedding-001 | Embeddings | Yes |
gemini-2.5-flash | LLM | Yes |
gemini-2.0-flash-lite | LLM | No (previous default) |
Configuration
API quotas
Free Tier:- Embeddings: 50 requests/minute
- LLM Calls: 15 requests/minute
- Generous monthly limits
- Pay-as-you-go
- Higher rate limits available
Error handling
Graceful degradation if Gemini is unavailable:Performance
Typical latencies:- Embedding request: 200-500ms
- LLM analysis: 2-5 seconds
- Batch embedding: 500ms-2s
Migration from v0.1.0
If upgrading from v0.1.0, update yoursimili.yaml:
Next steps
Gemini configuration
Setup Gemini for Simili Bot
OpenAI integration
Use OpenAI as an alternative

