Class: OllamaEmbedding
OllamaEmbedding is an alias for Ollama that implements the BaseEmbedding interface.
Extends
Implements
Constructors
new OllamaEmbedding()
new OllamaEmbedding(
params
):OllamaEmbedding
Parameters
• params: OllamaParams
Returns
Inherited from
Source
packages/core/src/llm/ollama.ts:75
Properties
embedBatchSize
embedBatchSize:
number
=DEFAULT_EMBED_BATCH_SIZE
Implementation of
BaseEmbedding
. embedBatchSize
Inherited from
Source
packages/core/src/embeddings/types.ts:13
hasStreaming
readonly
hasStreaming:true
=true
Inherited from
Source
packages/core/src/llm/ollama.ts:61
model
model:
string
Inherited from
Source
packages/core/src/llm/ollama.ts:66
ollama
ollama:
Ollama
Inherited from
Source
packages/core/src/llm/ollama.ts:63
options
options:
Partial
<Omit
<Options
,"temperature"
|"top_p"
|"num_ctx"
>> &Pick
<Options
,"temperature"
|"top_p"
|"num_ctx"
>
Inherited from
Source
packages/core/src/llm/ollama.ts:68
Accessors
metadata
get
metadata():LLMMetadata
Returns
Source
packages/core/src/llm/ollama.ts:87
Methods
abort()
abort():
void
Returns
void
Inherited from
Source
packages/core/src/llm/ollama.ts:205
chat()
chat(params)
chat(
params
):Promise
<AsyncIterable
<ChatResponseChunk
>>
Get a chat response from the LLM
Parameters
• params: LLMChatParamsStreaming
<object
, object
>
Returns
Promise
<AsyncIterable
<ChatResponseChunk
>>
Inherited from
Source
packages/core/src/llm/ollama.ts:99
chat(params)
chat(
params
):Promise
<ChatResponse
<object
>>
Parameters
• params: LLMChatParamsNonStreaming
<object
, object
>
Returns
Promise
<ChatResponse
<object
>>
Inherited from
Source
packages/core/src/llm/ollama.ts:102
complete()
complete(params)
complete(
params
):Promise
<AsyncIterable
<CompletionResponse
>>
Get a prompt completion from the LLM
Parameters
• params: LLMCompletionParamsStreaming
Returns
Promise
<AsyncIterable
<CompletionResponse
>>