docs/components/llms/models/langchain.mdx
Mem0 supports LangChain as a provider to access a wide range of LLM models. LangChain is a framework for developing applications powered by language models, making it easy to integrate various LLM providers through a consistent interface.
For a complete list of available chat models supported by LangChain, refer to the LangChain Chat Models documentation.
os.environ["OPENAI_API_KEY"] = "your-api-key"
openai_model = ChatOpenAI( model="gpt-5-mini", temperature=0.2, max_tokens=2000 )
config = { "llm": { "provider": "langchain", "config": { "model": openai_model } } }
m = Memory.from_config(config) messages = [ {"role": "user", "content": "I'm planning to watch a movie tonight. Any recommendations?"}, {"role": "assistant", "content": "How about thriller movies? They can be quite engaging."}, {"role": "user", "content": "I'm not a big fan of thriller movies but I love sci-fi movies."}, {"role": "assistant", "content": "Got it! I'll avoid thriller recommendations and suggest sci-fi movies in the future."} ] m.add(messages, user_id="alice", metadata={"category": "movies"})
```typescript TypeScript
import { Memory } from 'mem0ai/oss';
import { ChatOpenAI } from "@langchain/openai";
// Initialize a LangChain model directly
const openaiModel = new ChatOpenAI({
modelName: "gpt-4",
temperature: 0.2,
maxTokens: 2000,
apiKey: process.env.OPENAI_API_KEY,
});
const config = {
llm: {
provider: 'langchain',
config: {
model: openaiModel,
},
},
};
const memory = new Memory(config);
const messages = [
{"role": "user", "content": "I'm planning to watch a movie tonight. Any recommendations?"},
{"role": "assistant", "content": "How about thriller movies? They can be quite engaging."},
{"role": "user", "content": "I'm not a big fan of thriller movies but I love sci-fi movies."},
{"role": "assistant", "content": "Got it! I'll avoid thriller recommendations and suggest sci-fi movies in the future."}
]
await memory.add(messages, { userId: "alice", metadata: { category: "movies" } });
LangChain supports a wide range of LLM providers, including:
ChatOpenAI)ChatAnthropic)ChatGoogleGenerativeAI, ChatGooglePalm)ChatMistralAI)ChatOllama)AzureChatOpenAI)HuggingFaceChatEndpoint)You can use any of these model instances directly in your configuration. For a complete and up-to-date list of available providers, refer to the LangChain Chat Models documentation.
When using LangChain as a provider, you'll need to:
All available parameters for the langchain config are present in Master List of All Params in Config.