File tree 2 files changed +16
-8
lines changed
2 files changed +16
-8
lines changed Original file line number Diff line number Diff line change @@ -8,13 +8,17 @@ import { IAgentRuntime, ModelProvider } from "./types.ts";
8
8
*/
9
9
export async function embed ( runtime : IAgentRuntime , input : string ) {
10
10
// get the charcter, and handle by model type
11
- const model = models [ runtime . character . settings . model ] ;
12
-
13
- if ( model !== ModelProvider . OPENAI ) {
11
+ const provider = runtime . modelProvider ;
12
+
13
+ // Use local llama for non-OpenAI/OpenRouter providers
14
+ if ( provider !== ModelProvider . OPENAI && provider !== ModelProvider . OPENROUTER ) {
14
15
return await runtime . llamaService . getEmbeddingResponse ( input ) ;
15
16
}
16
17
17
- const embeddingModel = models [ runtime . modelProvider ] . model . embedding ;
18
+ const embeddingModel = getModel ( provider , ModelClass . EMBEDDING , runtime ) ;
19
+ const endpoint = provider === ModelProvider . OPENROUTER ?
20
+ "https://openrouter.ai/api/v1" :
21
+ models [ provider ] . endpoint ;
18
22
19
23
// Check if we already have the embedding in the lore
20
24
const cachedEmbedding = await retrieveCachedEmbedding ( runtime , input ) ;
Original file line number Diff line number Diff line change @@ -70,9 +70,13 @@ export async function generateText({
70
70
71
71
switch ( provider ) {
72
72
case ModelProvider . OPENAI :
73
- case ModelProvider . LLAMACLOUD : {
74
- prettyConsole . log ( "Initializing OpenAI model." ) ;
75
- const openai = createOpenAI ( { apiKey } ) ;
73
+ case ModelProvider . LLAMACLOUD :
74
+ case ModelProvider . OPENROUTER : {
75
+ prettyConsole . log ( `Initializing ${ provider } model.` ) ;
76
+ const openai = createOpenAI ( {
77
+ apiKey,
78
+ baseURL : provider === ModelProvider . OPENROUTER ? "https://openrouter.ai/api/v1" : undefined
79
+ } ) ;
76
80
77
81
console . log ( '****** CONTEXT\n' , context )
78
82
@@ -88,7 +92,7 @@ export async function generateText({
88
92
console . log ( "****** RESPONSE\n" , openaiResponse ) ;
89
93
90
94
response = openaiResponse ;
91
- prettyConsole . log ( " Received response from OpenAI model." ) ;
95
+ prettyConsole . log ( ` Received response from ${ provider } model.` ) ;
92
96
break ;
93
97
}
94
98
You can’t perform that action at this time.
0 commit comments