mirror of
https://github.com/zadam/trilium.git
synced 2025-11-17 02:30:42 +01:00
yes, this finally does set streaming to true
This commit is contained in:
@@ -20,6 +20,18 @@ export interface StreamChunk {
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Options for chat completion requests
|
||||
*
|
||||
* Key properties:
|
||||
* - stream: If true, the response will be streamed
|
||||
* - model: Model name to use
|
||||
* - provider: Provider to use (openai, anthropic, ollama, etc.)
|
||||
* - enableTools: If true, enables tool support
|
||||
*
|
||||
* The stream option is particularly important and should be consistently handled
|
||||
* throughout the pipeline. It should be explicitly set to true or false.
|
||||
*/
|
||||
export interface ChatCompletionOptions {
|
||||
model?: string;
|
||||
temperature?: number;
|
||||
|
||||
Reference in New Issue
Block a user