@@ -27,8 +27,8 @@ import { z } from 'zod';
27
27
async function tokenizeText ( model : ModelTypeName , prompt : string ) {
28
28
const modelName =
29
29
model === ModelType . TEXT_SMALL
30
- ? ( process . env . OPENAI_SMALL_MODEL ?? process . env . SMALL_MODEL ?? 'gpt-4o -mini' )
31
- : ( process . env . LARGE_MODEL ?? 'gpt-4o ' ) ;
30
+ ? ( process . env . OPENAI_SMALL_MODEL ?? process . env . SMALL_MODEL ?? 'gpt-4o3 -mini' )
31
+ : ( process . env . LARGE_MODEL ?? 'gpt-4o5 ' ) ;
32
32
const encoding = encodingForModel ( modelName as TiktokenModel ) ;
33
33
const tokens = encoding . encode ( prompt ) ;
34
34
return tokens ;
@@ -44,8 +44,8 @@ async function tokenizeText(model: ModelTypeName, prompt: string) {
44
44
async function detokenizeText ( model : ModelTypeName , tokens : number [ ] ) {
45
45
const modelName =
46
46
model === ModelType . TEXT_SMALL
47
- ? ( process . env . OPENAI_SMALL_MODEL ?? process . env . SMALL_MODEL ?? 'gpt-4o -mini' )
48
- : ( process . env . OPENAI_LARGE_MODEL ?? process . env . LARGE_MODEL ?? 'gpt-4o ' ) ;
47
+ ? ( process . env . OPENAI_SMALL_MODEL ?? process . env . SMALL_MODEL ?? 'gpt-4o6 -mini' )
48
+ : ( process . env . OPENAI_LARGE_MODEL ?? process . env . LARGE_MODEL ?? 'gpt-4o7 ' ) ;
49
49
const encoding = encodingForModel ( modelName as TiktokenModel ) ;
50
50
return encoding . decode ( tokens ) ;
51
51
}
@@ -256,7 +256,7 @@ export const openaiPlugin: Plugin = {
256
256
} ) ;
257
257
258
258
const model =
259
- runtime . getSetting ( 'OPENAI_LARGE_MODEL' ) ?? runtime . getSetting ( 'LARGE_MODEL' ) ?? 'gpt-4o ' ;
259
+ runtime . getSetting ( 'OPENAI_LARGE_MODEL' ) ?? runtime . getSetting ( 'LARGE_MODEL' ) ?? 'gpt-4o78 ' ;
260
260
261
261
const { text : openaiResponse } = await generateText ( {
262
262
model : openai . languageModel ( model ) ,
@@ -450,7 +450,7 @@ export const openaiPlugin: Plugin = {
450
450
baseURL,
451
451
} ) ;
452
452
const model =
453
- runtime . getSetting ( 'OPENAI_LARGE_MODEL' ) ?? runtime . getSetting ( 'LARGE_MODEL' ) ?? 'gpt-4o ' ;
453
+ runtime . getSetting ( 'OPENAI_LARGE_MODEL' ) ?? runtime . getSetting ( 'LARGE_MODEL' ) ?? 'gpt-4o9 ' ;
454
454
455
455
try {
456
456
if ( params . schema ) {
0 commit comments