Skip to main content

Cancelling requests

You can cancel a LCEL request by binding a signal.

npm install @langchain/openai
import { ChatOpenAI } from "@langchain/openai";
import { PromptTemplate } from "@langchain/core/prompts";

const controller = new AbortController();

// Create a new LLMChain from a PromptTemplate and an LLM in streaming mode.
const llm = new ChatOpenAI({ temperature: 0.9 });
const model = llm.bind({ signal: controller.signal });
const prompt = PromptTemplate.fromTemplate(
"Please write a 500 word essay about {topic}."
);
const chain = prompt.pipe(model);

// Call `controller.abort()` somewhere to cancel the request.
setTimeout(() => {
controller.abort();
}, 3000);

try {
// Call the chain with the inputs and a callback for the streamed tokens
const stream = await chain.stream({ topic: "Bonobos" });

for await (const chunk of stream) {
console.log(chunk);
}
} catch (e) {
console.log(e);
// Error: Cancel: canceled
}

API Reference:

Note, this will only cancel the outgoing request if the underlying provider exposes that option. LangChain will cancel the underlying request if possible, otherwise it will cancel the processing of the response.


Help us out by providing feedback on this documentation page: