How to do extraction without using function calling
This guide assumes familiarity with the following:
LLMs that are able to follow prompt instructions well can be tasked with outputting information in a given format without using function calling.
This approach relies on designing good prompts and then parsing the output of the LLMs to make them extract information well, though it lacks some of the guarantees provided by function calling or JSON mode.
Here, we’ll use Claude which is great at following instructions! See here for more about Anthropic models.
First, we’ll install the integration package:
- npm
- yarn
- pnpm
npm i @langchain/anthropic @langchain/core zod zod-to-json-schema
yarn add @langchain/anthropic @langchain/core zod zod-to-json-schema
pnpm add @langchain/anthropic @langchain/core zod zod-to-json-schema
import { ChatAnthropic } from "@langchain/anthropic";
const model = new ChatAnthropic({
model: "claude-3-sonnet-20240229",
temperature: 0,
});
All the same considerations for extraction quality apply for parsing approach.
This tutorial is meant to be simple, but generally should really include reference examples to squeeze out performance!
Using StructuredOutputParser
The following example uses the built-in
StructuredOutputParser
to
parse the output of a chat model. We use the built-in prompt formatting
instructions contained in the parser.
import { z } from "zod";
import { StructuredOutputParser } from "langchain/output_parsers";
import { ChatPromptTemplate } from "@langchain/core/prompts";
let personSchema = z
.object({
name: z.optional(z.string()).describe("The name of the person"),
hair_color: z
.optional(z.string())
.describe("The color of the person's hair, if known"),
height_in_meters: z
.optional(z.string())
.describe("Height measured in meters"),
})
.describe("Information about a person.");
const parser = StructuredOutputParser.fromZodSchema(personSchema);
const prompt = ChatPromptTemplate.fromMessages([
[
"system",
"Answer the user query. Wrap the output in `json` tags\n{format_instructions}",
],
["human", "{query}"],
]);
const partialedPrompt = await prompt.partial({
format_instructions: parser.getFormatInstructions(),
});
Let’s take a look at what information is sent to the model
const query = "Anna is 23 years old and she is 6 feet tall";
const promptValue = await partialedPrompt.invoke({ query });
console.log(promptValue.toChatMessages());
[
SystemMessage {
lc_serializable: true,
lc_kwargs: {
content: "Answer the user query. Wrap the output in `json` tags\n" +
"You must format your output as a JSON value th"... 1444 more characters,
additional_kwargs: {}
},
lc_namespace: [ "langchain_core", "messages" ],
content: "Answer the user query. Wrap the output in `json` tags\n" +
"You must format your output as a JSON value th"... 1444 more characters,
name: undefined,
additional_kwargs: {}
},
HumanMessage {
lc_serializable: true,
lc_kwargs: {
content: "Anna is 23 years old and she is 6 feet tall",
additional_kwargs: {}
},
lc_namespace: [ "langchain_core", "messages" ],
content: "Anna is 23 years old and she is 6 feet tall",
name: undefined,
additional_kwargs: {}
}
]
const chain = partialedPrompt.pipe(model).pipe(parser);
await chain.invoke({ query });
{ name: "Anna", hair_color: "", height_in_meters: "1.83" }
Custom Parsing
You can also create a custom prompt and parser with LangChain
and
LCEL
.
You can use a raw function to parse the output from the model.
In the below example, we’ll pass the schema into the prompt as JSON
schema. For convenience, we’ll declare our schema with Zod, then use the
zod-to-json-schema
utility to convert it to JSON schema.
import { z } from "zod";
import { zodToJsonSchema } from "zod-to-json-schema";
personSchema = z
.object({
name: z.optional(z.string()).describe("The name of the person"),
hair_color: z
.optional(z.string())
.describe("The color of the person's hair, if known"),
height_in_meters: z
.optional(z.string())
.describe("Height measured in meters"),
})
.describe("Information about a person.");
const peopleSchema = z.object({
people: z.array(personSchema),
});
const SYSTEM_PROMPT_TEMPLATE = [
"Answer the user's query. You must return your answer as JSON that matches the given schema:",
"```json\n{schema}\n```.",
"Make sure to wrap the answer in ```json and ``` tags. Conform to the given schema exactly.",
].join("\n");
const customParsingPrompt = ChatPromptTemplate.fromMessages([
["system", SYSTEM_PROMPT_TEMPLATE],
["human", "{query}"],
]);
const extractJsonFromOutput = (message) => {
const text = message.content;
// Define the regular expression pattern to match JSON blocks
const pattern = /```json\s*((.|\n)*?)\s*```/gs;
// Find all non-overlapping matches of the pattern in the string
const matches = pattern.exec(text);
if (matches && matches[1]) {
try {
return JSON.parse(matches[1].trim());
} catch (error) {
throw new Error(`Failed to parse: ${matches[1]}`);
}
} else {
throw new Error(`No JSON found in: ${message}`);
}
};
const customParsingQuery = "Anna is 23 years old and she is 6 feet tall";
const customParsingPromptValue = await customParsingPrompt.invoke({
schema: zodToJsonSchema(peopleSchema),
customParsingQuery,
});
customParsingPromptValue.toString();
"System: Answer the user's query. You must return your answer as JSON that matches the given schema:\n"... 170 more characters
const customParsingChain = prompt.pipe(model).pipe(extractJsonFromOutput);
await customParsingChain.invoke({
schema: zodToJsonSchema(peopleSchema),
customParsingQuery,
});
{ name: "Anna", age: 23, height: { feet: 6, inches: 0 } }
Next steps
You’ve now learned how to perform extraction without using tool calling.
Next, check out some of the other guides in this section, such as some tips on how to improve extraction quality with examples.