Optional callbacksOptional maxThe maximum number of tokens the response should contain.
Optional metadataMetadata for this call and any sub-calls (eg. a Chain calling an LLM). Keys should be strings, values should be JSON-serializable.
Optional runName for the tracer run for this call. Defaults to the name of the class.
Optional signalAbort signal for this call. If provided, the call will be aborted when the signal is aborted.
Optional stopStop tokens to use for this call. If not provided, the default stop tokens for the model will be used.
Optional streamingThis lets the streaming method know which model we are using, valid options are: 'llama2' - A Llama2 model, this is the default. 'chatML' - A ChatML model 'falcon' - A Falcon model 'genral' - Any other model, uses "### Human\n", "### Assistant\n" format
Optional tagsTags for this call and any sub-calls (eg. a Chain calling an LLM). You can use these to filter calls.
Optional timeoutTimeout for this call in milliseconds.
Generated using TypeDoc
Callbacks for this call and any sub-calls (eg. a Chain calling an LLM). Tags are passed to all callbacks, metadata is passed to handle*Start callbacks.