Class: abstract
ToolCallLLM<AdditionalChatOptions>
Unified language model interface
Extends
BaseLLM
<AdditionalChatOptions
,ToolCallLLMMessageOptions
>
Extended by
Type Parameters
• AdditionalChatOptions extends object
= object
Constructors
new ToolCallLLM()
new ToolCallLLM<
AdditionalChatOptions
>():ToolCallLLM
<AdditionalChatOptions
>
Returns
ToolCallLLM
<AdditionalChatOptions
>
Inherited from
Properties
metadata
abstract
metadata:LLMMetadata
Inherited from
Defined in
packages/core/dist/llms/index.d.ts:167
supportToolCall
abstract
supportToolCall:boolean
Defined in
packages/core/dist/llms/index.d.ts:174
Methods
chat()
chat(params)
abstract
chat(params
):Promise
<AsyncIterable
<ChatResponseChunk
<object
>>>
Get a chat response from the LLM
Parameters
• params: LLMChatParamsStreaming
<AdditionalChatOptions
, ToolCallLLMMessageOptions
>
Returns
Promise
<AsyncIterable
<ChatResponseChunk
<object
>>>
Inherited from
Defined in
packages/core/dist/llms/index.d.ts:170
chat(params)
abstract
chat(params
):Promise
<ChatResponse
<ToolCallLLMMessageOptions
>>
Get a chat response from the LLM
Parameters
• params: LLMChatParamsNonStreaming
<AdditionalChatOptions
, ToolCallLLMMessageOptions
>
Returns
Promise
<ChatResponse
<ToolCallLLMMessageOptions
>>
Inherited from
Defined in
packages/core/dist/llms/index.d.ts:171
complete()
complete(params)
complete(
params
):Promise
<AsyncIterable
<CompletionResponse
>>
Get a prompt completion from the LLM
Parameters
• params: LLMCompletionParamsStreaming
Returns
Promise
<AsyncIterable
<CompletionResponse
>>
Inherited from
Defined in
packages/core/dist/llms/index.d.ts:168
complete(params)
complete(
params
):Promise
<CompletionResponse
>
Get a prompt completion from the LLM
Parameters
• params: LLMCompletionParamsNonStreaming
Returns
Promise
<CompletionResponse
>
Inherited from
Defined in
packages/core/dist/llms/index.d.ts:169