V2
This commit is contained in:
@@ -1,84 +1,66 @@
|
||||
# AI Providers Guide
|
||||
|
||||
## Supported AI Providers
|
||||
## Scope
|
||||
|
||||
Your AI Interview Assistant now supports multiple AI providers! Here's how to set up and use each one:
|
||||
This guide covers **chat/response providers** used by the extension after transcription.
|
||||
|
||||
## 🤖 **OpenAI (GPT)**
|
||||
- **Models Available**: GPT-4o, GPT-4o-mini, GPT-4-turbo, GPT-3.5-turbo
|
||||
- **API Key**: Get from [OpenAI Platform](https://platform.openai.com/account/api-keys)
|
||||
- **Recommended Model**: GPT-4o-mini (good balance of speed and quality)
|
||||
- **Cost**: Pay per token usage
|
||||
Note: Speech-to-text is configured separately in Assistant Setup (`STT Provider`, `STT Model`, language/task/VAD/beam settings).
|
||||
|
||||
## 🧠 **Anthropic (Claude)**
|
||||
- **Models Available**: Claude-3.5-Sonnet, Claude-3.5-Haiku, Claude-3-Opus
|
||||
- **API Key**: Get from [Anthropic Console](https://console.anthropic.com/)
|
||||
- **Recommended Model**: Claude-3.5-Sonnet (excellent reasoning)
|
||||
- **Cost**: Pay per token usage
|
||||
## Supported Chat Providers
|
||||
|
||||
## 🔍 **Google (Gemini)**
|
||||
- **Models Available**: Gemini-1.5-Pro, Gemini-1.5-Flash, Gemini-Pro
|
||||
- **API Key**: Get from [Google AI Studio](https://aistudio.google.com/app/apikey)
|
||||
- **Recommended Model**: Gemini-1.5-Flash (fast and efficient)
|
||||
- **Cost**: Free tier available, then pay per token
|
||||
### OpenAI
|
||||
- Default models in UI: `gpt-4o`, `gpt-4o-mini`, `gpt-4-turbo`, `gpt-3.5-turbo`
|
||||
- API key: https://platform.openai.com/account/api-keys
|
||||
- Good default: `gpt-4o-mini` (speed/cost balance)
|
||||
|
||||
## 🌊 **DeepSeek**
|
||||
- **Models Available**: DeepSeek-Chat, DeepSeek-Reasoner
|
||||
- **API Key**: Get from [DeepSeek Platform](https://platform.deepseek.com/)
|
||||
- **Recommended Model**: DeepSeek-Chat (general use)
|
||||
- **Cost**: Pay per token usage
|
||||
### Anthropic
|
||||
- Default models in UI: `claude-3-5-sonnet-20241022`, `claude-3-5-haiku-20241022`, `claude-3-opus-20240229`
|
||||
- API key: https://console.anthropic.com/
|
||||
- Good default: `claude-3-5-sonnet-20241022`
|
||||
|
||||
## 🏠 **Ollama (Local)**
|
||||
- **Models Available**: Llama3.2, Llama3.1, Mistral, CodeLlama, Phi3
|
||||
- **Setup**: Install [Ollama](https://ollama.ai/) locally
|
||||
- **No API Key Required**: Runs completely on your machine
|
||||
- **Cost**: Free (uses your computer's resources)
|
||||
### Google Gemini
|
||||
- Default models in UI: `gemini-1.5-pro`, `gemini-1.5-flash`, `gemini-pro`
|
||||
- API key: https://aistudio.google.com/app/apikey
|
||||
- Good default: `gemini-1.5-flash`
|
||||
|
||||
## 🚀 **How to Setup**
|
||||
### DeepSeek
|
||||
- Default models in UI: `deepseek-chat`, `deepseek-reasoner`
|
||||
- API key: https://platform.deepseek.com/
|
||||
- Good default: `deepseek-chat`
|
||||
|
||||
### 1. **Choose Your Provider**
|
||||
- Open the extension side panel
|
||||
- Select your preferred AI provider from the dropdown
|
||||
### Ollama (local)
|
||||
- Default models in UI: `llama3.2`, `llama3.1`, `mistral`, `codellama`, `phi3`
|
||||
- API key: not required
|
||||
- Endpoint used by extension: `http://localhost:11434`
|
||||
|
||||
### 2. **Select Model**
|
||||
- Choose the specific model you want to use
|
||||
- Different models have different capabilities and speeds
|
||||
## Model List Behavior
|
||||
|
||||
### 3. **Add API Key** (if required)
|
||||
- Enter your API key for the selected provider
|
||||
- Ollama doesn't require an API key
|
||||
- Keys are stored securely in Chrome's storage
|
||||
- For cloud providers, if an API key is saved, the extension attempts to fetch live model lists.
|
||||
- If model fetch fails, the extension falls back to the built-in default model list above.
|
||||
- For Ollama, the extension reads models from `/api/tags`.
|
||||
|
||||
### 4. **Start Using**
|
||||
- Click "Start Listening" to begin audio capture
|
||||
- The extension will use your selected AI provider for responses
|
||||
## Setup Steps
|
||||
|
||||
## 💡 **Tips**
|
||||
1. Open side panel -> `Assistant Setup`.
|
||||
2. Choose `AI Provider`.
|
||||
3. Save provider API key (not needed for Ollama).
|
||||
4. Select model.
|
||||
5. Start listening.
|
||||
|
||||
- **For Speed**: Use GPT-4o-mini, Gemini-1.5-Flash, or Claude-3.5-Haiku
|
||||
- **For Quality**: Use GPT-4o, Claude-3.5-Sonnet, or Gemini-1.5-Pro
|
||||
- **For Privacy**: Use Ollama (runs locally, no data sent to servers)
|
||||
- **For Free Usage**: Try Google Gemini's free tier or set up Ollama
|
||||
## Recommended Defaults
|
||||
|
||||
## 🔧 **Ollama Setup**
|
||||
- Fastest general: `gpt-4o-mini` / `gemini-1.5-flash` / `claude-3-5-haiku-20241022`
|
||||
- Highest quality: `gpt-4o` / `claude-3-5-sonnet-20241022` / `gemini-1.5-pro`
|
||||
- Local-only privacy: `ollama` + local STT
|
||||
|
||||
If you want to use Ollama (local AI):
|
||||
## Troubleshooting
|
||||
|
||||
1. Install Ollama from [ollama.ai](https://ollama.ai/)
|
||||
2. Run: `ollama pull llama3.2` (or your preferred model)
|
||||
3. Make sure Ollama is running: `ollama serve`
|
||||
4. Select "Ollama (Local)" in the extension
|
||||
- `API key not set`: save provider key in Assistant Setup.
|
||||
- `Failed to fetch models`: key may be invalid, provider API unavailable, or network blocked. Default model list is used as fallback.
|
||||
- `Ollama connection failed`: ensure `ollama serve` is running and model is pulled.
|
||||
- Slow or expensive responses: switch to smaller/faster model and enable Speed mode.
|
||||
|
||||
## 🆘 **Troubleshooting**
|
||||
## Storage Note
|
||||
|
||||
- **"API key not set"**: Make sure you've entered a valid API key
|
||||
- **"Failed to connect"**: Check your internet connection (or Ollama service for local)
|
||||
- **"Invalid API key"**: Verify your API key is correct and has sufficient credits
|
||||
- **Slow responses**: Try switching to a faster model like GPT-4o-mini or Gemini-1.5-Flash
|
||||
|
||||
## 🔒 **Privacy & Security**
|
||||
|
||||
- API keys are stored locally in Chrome's secure storage
|
||||
- Only the selected provider receives your audio transcriptions
|
||||
- Ollama option keeps everything completely local
|
||||
- No audio data is stored permanently
|
||||
- Provider API keys are stored in extension sync storage (`chrome.storage.sync`).
|
||||
- Keep least-privilege keys where possible and rotate keys regularly.
|
||||
|
||||
Reference in New Issue
Block a user