🔐 Mem0 is now SOC 2 and HIPAA compliant! We're committed to the highest standards of data security and privacy, enabling secure memory for enterprises, healthcare, and beyond. Learn more
To use DeepSeek LLM models, you have to set the DEEPSEEK_API_KEY environment variable. You can also optionally set DEEPSEEK_API_BASE if you need to use a different API endpoint (defaults to “https://api.deepseek.com”).
import osfrom mem0 import Memoryos.environ["DEEPSEEK_API_KEY"] = "your-api-key"os.environ["OPENAI_API_KEY"] = "your-api-key" # for embedder modelconfig = { "llm": { "provider": "deepseek", "config": { "model": "deepseek-chat", # default model "temperature": 0.2, "max_tokens": 2000, "top_p": 1.0 } }}m = Memory.from_config(config)messages = [ {"role": "user", "content": "I'm planning to watch a movie tonight. Any recommendations?"}, {"role": "assistant", "content": "How about a thriller movies? They can be quite engaging."}, {"role": "user", "content": "I’m not a big fan of thriller movies but I love sci-fi movies."}, {"role": "assistant", "content": "Got it! I'll avoid thriller recommendations and suggest sci-fi movies in the future."}]m.add(messages, user_id="alice", metadata={"category": "movies"})
You can also configure the API base URL in the config: