Class to parse the output of an LLM call into a dictionary.

Hierarchy

Constructors

  • Parameters

    • fields: RegexParserFields

    Returns RegexParser

  • Parameters

    • regex: string | RegExp
    • outputKeys: string[]
    • Optional defaultOutputKey: string

    Returns RegexParser

Properties

outputKeys: string[]
regex: string | RegExp
defaultOutputKey?: string

Methods

  • Default implementation of batch, which calls invoke N times. Subclasses should override this method if they can batch more efficiently.

    Parameters

    Returns Promise<Record<string, string>[]>

    An array of RunOutputs, or mixed RunOutputs and errors if batchOptions.returnExceptions is set

  • Parameters

    Returns Promise<(Error | Record<string, string>)[]>

  • Parameters

    Returns Promise<(Error | Record<string, string>)[]>

  • Returns a string with instructions on how the LLM output should be formatted to match the regex pattern.

    Returns string

    A string with formatting instructions.

  • Calls the parser with a given input and optional configuration options. If the input is a string, it creates a generation with the input as text and calls parseResult. If the input is a BaseMessage, it creates a generation with the input as a message and the content of the input as text, and then calls parseResult.

    Parameters

    • input: string | BaseMessage

      The input to the parser, which can be a string or a BaseMessage.

    • Optional options: BaseCallbackConfig

      Optional configuration options.

    Returns Promise<Record<string, string>>

    A promise of the parsed output.

  • Parses the given text using the regex pattern and returns a dictionary with the parsed output. If the regex pattern does not match the text and no defaultOutputKey is provided, throws an OutputParserException.

    Parameters

    • text: string

      The text to be parsed.

    Returns Promise<Record<string, string>>

    A dictionary with the parsed output.

  • Parses the result of an LLM call. This method is meant to be implemented by subclasses to define how the output from the LLM should be parsed.

    Parameters

    Returns Promise<Record<string, string>>

    A promise of the parsed output.

  • Parses the result of an LLM call with a given prompt. By default, it simply calls parseResult.

    Parameters

    Returns Promise<Record<string, string>>

    A promise of the parsed output.

  • Parameters

    Returns Promise<Record<string, string>>

  • Create a new runnable sequence that runs each individual runnable in series, piping the output of one runnable into another runnable or runnable-like.

    Type Parameters

    • NewRunOutput

    Parameters

    • coerceable: RunnableLike<Record<string, string>, NewRunOutput>

      A runnable, function, or object whose values are functions or runnables.

    Returns RunnableSequence<string | BaseMessage, Exclude<NewRunOutput, Error>>

    A new runnable sequence.

  • Stream output in chunks.

    Parameters

    Returns Promise<IterableReadableStream<Record<string, string>>>

    A readable stream that is also an iterable.

  • Stream all output from a runnable, as reported to the callback system. This includes all inner runs of LLMs, Retrievers, Tools, etc. Output is streamed as Log objects, which include a list of jsonpatch ops that describe how the state of the run has changed in each step, and the final state of the run. The jsonpatch ops can be applied in order to construct state.

    Parameters

    • input: string | BaseMessage
    • Optional options: Partial<BaseCallbackConfig>
    • Optional streamOptions: Omit<LogStreamCallbackHandlerInput, "autoClose">

    Returns AsyncGenerator<RunLogPatch, any, unknown>

  • Default implementation of transform, which buffers input and then calls stream. Subclasses should override this method if they can start producing output while input is still being generated.

    Parameters

    Returns AsyncGenerator<Record<string, string>, any, unknown>

Generated using TypeDoc