Optional
callbacksOptional
maxThe maximum number of tokens the response should contain.
Optional
metadataMetadata for this call and any sub-calls (eg. a Chain calling an LLM). Keys should be strings, values should be JSON-serializable.
Optional
runName for the tracer run for this call. Defaults to the name of the class.
Optional
signalAbort signal for this call. If provided, the call will be aborted when the signal is aborted.
Optional
stopStop tokens to use for this call. If not provided, the default stop tokens for the model will be used.
Optional
streamingThis lets the streaming method know which model we are using, valid options are: 'llama2' - A Llama2 model, this is the default. 'chatML' - A ChatML model 'falcon' - A Falcon model 'genral' - Any other model, uses "### Human\n", "### Assistant\n" format
Optional
tagsTags for this call and any sub-calls (eg. a Chain calling an LLM). You can use these to filter calls.
Optional
timeoutTimeout for this call in milliseconds.
Generated using TypeDoc
Callbacks for this call and any sub-calls (eg. a Chain calling an LLM). Tags are passed to all callbacks, metadata is passed to handle*Start callbacks.