Class BaseTransformOutputParser<T>Abstract

Class to parse the output of an LLM call that also allows streaming inputs.

Type Parameters

  • T = unknown

Hierarchy

Constructors

Properties

name?: string

Methods

  • Assigns new fields to the dict output of this runnable. Returns a new runnable.

    Parameters

    • mapping: RunnableMapLike<Record<string, unknown>, Record<string, unknown>>

    Returns RunnableSequence<any, any>

  • Default implementation of batch, which calls invoke N times. Subclasses should override this method if they can batch more efficiently.

    Parameters

    • inputs: (string | BaseMessage)[]

      Array of inputs to each batch call.

    • Optional options: Partial<RunnableConfig> | Partial<RunnableConfig>[]

      Either a single call options object to apply to each batch call or an array for each call.

    • Optional batchOptions: RunnableBatchOptions & {
          returnExceptions?: false;
      }

    Returns Promise<T[]>

    An array of RunOutputs, or mixed RunOutputs and errors if batchOptions.returnExceptions is set

  • Parameters

    Returns Promise<(Error | T)[]>

  • Parameters

    Returns Promise<(Error | T)[]>

  • Return a string describing the format of the output.

    Parameters

    Returns string

    Format instructions.

    Example

    {
    "foo": "bar"
    }
  • Parameters

    • Optional suffix: string

    Returns string

  • Calls the parser with a given input and optional configuration options. If the input is a string, it creates a generation with the input as text and calls parseResult. If the input is a BaseMessage, it creates a generation with the input as a message and the content of the input as text, and then calls parseResult.

    Parameters

    • input: string | BaseMessage

      The input to the parser, which can be a string or a BaseMessage.

    • Optional options: RunnableConfig

      Optional configuration options.

    Returns Promise<T>

    A promise of the parsed output.

  • Parse the output of an LLM call.

    Parameters

    • text: string

      LLM output to parse.

    • Optional callbacks: Callbacks

    Returns Promise<T>

    Parsed output.

  • Parses the result of an LLM call. This method is meant to be implemented by subclasses to define how the output from the LLM should be parsed.

    Parameters

    Returns Promise<T>

    A promise of the parsed output.

  • Parses the result of an LLM call with a given prompt. By default, it simply calls parseResult.

    Parameters

    • generations: Generation[] | ChatGeneration[]

      The generations from an LLM call.

    • _prompt: BasePromptValueInterface

      The prompt used in the LLM call.

    • Optional callbacks: Callbacks

      Optional callbacks.

    Returns Promise<T>

    A promise of the parsed output.

  • Parameters

    • text: string
    • _prompt: BasePromptValueInterface
    • Optional callbacks: Callbacks

    Returns Promise<T>

  • Pick keys from the dict output of this runnable. Returns a new runnable.

    Parameters

    • keys: string | string[]

    Returns RunnableSequence<any, any>

  • Create a new runnable sequence that runs each individual runnable in series, piping the output of one runnable into another runnable or runnable-like.

    Type Parameters

    • NewRunOutput

    Parameters

    • coerceable: RunnableLike<T, NewRunOutput>

      A runnable, function, or object whose values are functions or runnables.

    Returns RunnableSequence<string | BaseMessage, Exclude<NewRunOutput, Error>>

    A new runnable sequence.

  • Stream output in chunks.

    Parameters

    Returns Promise<IterableReadableStream<T>>

    A readable stream that is also an iterable.

  • Stream all output from a runnable, as reported to the callback system. This includes all inner runs of LLMs, Retrievers, Tools, etc. Output is streamed as Log objects, which include a list of jsonpatch ops that describe how the state of the run has changed in each step, and the final state of the run. The jsonpatch ops can be applied in order to construct state.

    Parameters

    • input: string | BaseMessage
    • Optional options: Partial<RunnableConfig>
    • Optional streamOptions: Omit<LogStreamCallbackHandlerInput, "autoClose">

    Returns AsyncGenerator<RunLogPatch, any, unknown>

  • Transforms an asynchronous generator of input into an asynchronous generator of parsed output.

    Parameters

    • inputGenerator: AsyncGenerator<string | BaseMessage, any, unknown>

      An asynchronous generator of input.

    • options: BaseCallbackConfig

      A configuration object.

    Returns AsyncGenerator<T, any, unknown>

    An asynchronous generator of parsed output.

  • Bind lifecycle listeners to a Runnable, returning a new Runnable. The Run object contains information about the run, including its id, type, input, output, error, startTime, endTime, and any tags or metadata added to the run.

    Parameters

    • params: {
          onEnd?: ((run, config?) => void | Promise<void>);
          onError?: ((run, config?) => void | Promise<void>);
          onStart?: ((run, config?) => void | Promise<void>);
      }

      The object containing the callback functions.

      • Optional onEnd?: ((run, config?) => void | Promise<void>)
          • (run, config?): void | Promise<void>
          • Called after the runnable finishes running, with the Run object.

            Parameters

            Returns void | Promise<void>

      • Optional onError?: ((run, config?) => void | Promise<void>)
          • (run, config?): void | Promise<void>
          • Called if the runnable throws an error, with the Run object.

            Parameters

            Returns void | Promise<void>

      • Optional onStart?: ((run, config?) => void | Promise<void>)
          • (run, config?): void | Promise<void>
          • Called before the runnable starts running, with the Run object.

            Parameters

            Returns void | Promise<void>

    Returns Runnable<string | BaseMessage, T, RunnableConfig>

Generated using TypeDoc