🔐 Mem0 is now SOC 2 and HIPAA compliant! We're committed to the highest standards of data security and privacy, enabling secure memory for enterprises, healthcare, and beyond. Learn more
To use mistral’s models, please obtain the Mistral AI api key from their console. Set the MISTRAL_API_KEY environment variable to use the model as given below in the example.
import osfrom mem0 import Memoryos.environ["OPENAI_API_KEY"] = "your-api-key" # used for embedding modelos.environ["MISTRAL_API_KEY"] = "your-api-key"config = { "llm": { "provider": "litellm", "config": { "model": "open-mixtral-8x7b", "temperature": 0.1, "max_tokens": 2000, } }}m = Memory.from_config(config)messages = [ {"role": "user", "content": "I'm planning to watch a movie tonight. Any recommendations?"}, {"role": "assistant", "content": "How about a thriller movies? They can be quite engaging."}, {"role": "user", "content": "I’m not a big fan of thriller movies but I love sci-fi movies."}, {"role": "assistant", "content": "Got it! I'll avoid thriller recommendations and suggest sci-fi movies in the future."}]m.add(messages, user_id="alice", metadata={"category": "movies"})