Interface LlamaCppCallOptions

Hierarchy

Properties

callbacks?: Callbacks

Callbacks for this call and any sub-calls (eg. a Chain calling an LLM). Tags are passed to all callbacks, metadata is passed to handle*Start callbacks.

maxTokens?: number

The maximum number of tokens the response should contain.

metadata?: Record<string, unknown>

Metadata for this call and any sub-calls (eg. a Chain calling an LLM). Keys should be strings, values should be JSON-serializable.

runName?: string

Name for the tracer run for this call. Defaults to the name of the class.

signal?: AbortSignal

Abort signal for this call. If provided, the call will be aborted when the signal is aborted.

stop?: string[]

Stop tokens to use for this call. If not provided, the default stop tokens for the model will be used.

streamingModel?: string

This lets the streaming method know which model we are using, valid options are: 'llama2' - A Llama2 model, this is the default. 'chatML' - A ChatML model 'falcon' - A Falcon model 'genral' - Any other model, uses "### Human\n", "### Assistant\n" format

tags?: string[]

Tags for this call and any sub-calls (eg. a Chain calling an LLM). You can use these to filter calls.

timeout?: number

Timeout for this call in milliseconds.

Generated using TypeDoc