Groq is the creator of the world’s first Language Processing Unit (LPU), providing exceptional speed performance for AI workloads running on their LPU Inference Engine.
In order to use LLMs from Groq, go to their platform and get the API key. Set the API key as GROQ_API_KEY environment variable to use the model as given below in the example.
import osfrom mem0 import Memoryos.environ["OPENAI_API_KEY"]="your-api-key"# used for embedding modelos.environ["GROQ_API_KEY"]="your-api-key"config ={"llm":{"provider":"groq","config":{"model":"mixtral-8x7b-32768","temperature":0.1,"max_tokens":2000,}}}m = Memory.from_config(config)messages =[{"role":"user","content":"I'm planning to watch a movie tonight. Any recommendations?"},{"role":"assistant","content":"How about a thriller movies? They can be quite engaging."},{"role":"user","content":"I’m not a big fan of thriller movies but I love sci-fi movies."},{"role":"assistant","content":"Got it! I'll avoid thriller recommendations and suggest sci-fi movies in the future."}]m.add(messages, user_id="alice", metadata={"category":"movies"})