AI Providers
MemryLab supports 9 LLM providers including 8 with free tiers. You can switch providers at any time without losing data. LLM and embedding providers can be configured independently — for example, use Gemini for analysis and Ollama for private local embeddings.
Recommended Local Models (Ollama)
For local inference, pick models based on your GPU VRAM. Embedding model is always nomic-embed-text.
| VRAM | LLM Model | Speed | Install |
|---|---|---|---|
| 4 GB | llama3.2:3b | ~40 tok/s | ollama pull llama3.2:3b |
| 8 GB | llama3.1:8b | ~35 tok/s | ollama pull llama3.1:8b |
| 12 GB ★ | qwen2.5:14b-instruct-q5_K_M | ~25 tok/s | ollama pull qwen2.5:14b-instruct-q5_K_M |
| 16 GB+ | qwen2.5:32b-instruct-q4_K_M | ~15 tok/s | ollama pull qwen2.5:32b-instruct-q4_K_M |
Larger models produce better belief extraction and contradiction detection. For no-GPU setups, use Gemini Flash (free) or Groq (free) as cloud providers.
💡 Pro Tip: Mix Providers
In Settings → Embedding Provider, you can use a different provider for embeddings vs LLM. The best privacy setup: use Gemini Flash (free, fast) for analysis + Ollama (local) for embeddings. This way your search index stays fully private while analysis uses a powerful cloud model.
Quick Comparison
| Provider | Free Tier | Details |
|---|---|---|
| Google Gemini | Yes | 1500 req/day on Gemini 1.5 Flash |
| Groq | Yes | Free tier with rate limits |
| OpenRouter | Yes | Free models available (Llama, Mistral, etc.) |
| Cerebras | Yes | Free tier, fastest inference |
| Mistral | Yes | Free tier for small models |
| SambaNova | Yes | Free tier with generous limits |
| Cohere | Yes | Free trial API key |
| OpenAI | Paid | Pay-per-use, starts at $0.002/1K tokens |
| Custom (OpenAI-compatible) | Paid | Any OpenAI-compatible endpoint (LM Studio, Ollama, etc.) |
Google Gemini
1. Go to ai.google.dev and sign in
2. Create an API key
3. In MemryLab: Settings → LLM → Select "Gemini"
4. Paste your API key and test connectionGroq
1. Go to console.groq.com and create an account
2. Generate an API key from the dashboard
3. In MemryLab: Settings → LLM → Select "Groq"
4. Paste your API key and test connectionOpenRouter
1. Go to openrouter.ai and create an account
2. Generate an API key
3. In MemryLab: Settings → LLM → Select "OpenRouter"
4. Paste your API key — free models are auto-selectedCerebras
1. Go to cloud.cerebras.ai and sign up
2. Create an API key from the dashboard
3. In MemryLab: Settings → LLM → Select "Cerebras"
4. Paste your API key and test connectionMistral
1. Go to console.mistral.ai and create an account
2. Generate an API key
3. In MemryLab: Settings → LLM → Select "Mistral"
4. Paste your API key and test connectionSambaNova
1. Go to cloud.sambanova.ai and sign up
2. Create an API key
3. In MemryLab: Settings → LLM → Select "SambaNova"
4. Paste your API key and test connectionCohere
1. Go to dashboard.cohere.com and sign up
2. Copy your trial API key
3. In MemryLab: Settings → LLM → Select "Cohere"
4. Paste your API key and test connectionOpenAI
1. Go to platform.openai.com and sign in
2. Create an API key under API Keys
3. In MemryLab: Settings → LLM → Select "OpenAI"
4. Paste your API key and test connectionCustom (OpenAI-compatible)
1. In MemryLab: Settings → LLM → Select "Custom"
2. Enter the base URL (e.g., http://localhost:1234/v1)
3. Enter API key if required (or leave blank for local)
4. Select a model name and test connectionAPI Key Security
MemryLab stores API keys in your operating system's secure keychain — never in plain text files, environment variables, or the SQLite database.
- Windows: Windows Credential Manager
- macOS: Keychain Access
- Linux: libsecret (GNOME Keyring / KWallet)