Skip to main content


Here's an example of calling a HugggingFaceInference model as an LLM:

npm install @huggingface/inference@2

We're unifying model params across all packages. We now suggest using model instead of modelName, and apiKey for API keys.

import { HuggingFaceInference } from "langchain/llms/hf";

const model = new HuggingFaceInference({
model: "gpt2",
apiKey: "YOUR-API-KEY", // In Node.js defaults to process.env.HUGGINGFACEHUB_API_KEY
const res = await model.invoke("1 + 1 =");
console.log({ res });

Was this page helpful?

You can also leave detailed feedback on GitHub.