@@ -68,64 +68,61 @@ export async function generateText({
68
68
69
69
switch ( provider ) {
70
70
case ModelProvider . OPENAI :
71
- case ModelProvider . LLAMACLOUD :
72
- {
73
- console . log ( "Initializing OpenAI model." ) ;
74
- const openai = createOpenAI ( { apiKey } ) ;
75
-
76
- const { text : openaiResponse } = await aiGenerateText ( {
77
- model : openai . languageModel ( model ) ,
78
- prompt : context ,
79
- temperature : temperature ,
80
- maxTokens : max_response_length ,
81
- frequencyPenalty : frequency_penalty ,
82
- presencePenalty : presence_penalty ,
83
- } ) ;
84
-
85
- response = openaiResponse ;
86
- console . log ( "Received response from OpenAI model." ) ;
87
- break ;
88
- }
89
-
90
- case ModelProvider . ANTHROPIC :
91
- {
92
- console . log ( "Initializing Anthropic model." ) ;
93
- const anthropicVertex = createAnthropicVertex ( ) ;
94
-
95
- const { text : anthropicResponse } = await aiGenerateText ( {
96
- model : anthropicVertex ( model ) ,
97
- prompt : context ,
98
- temperature : temperature ,
99
- maxTokens : max_response_length ,
100
- frequencyPenalty : frequency_penalty ,
101
- presencePenalty : presence_penalty ,
102
- } ) ;
103
-
104
- response = anthropicResponse ;
105
- console . log ( "Received response from Anthropic model." ) ;
106
- break ;
107
- }
108
-
109
- case ModelProvider . GROK :
110
- {
111
- console . log ( "Initializing Grok model." ) ;
112
- const grok = createGroq ( { apiKey } ) ;
113
-
114
- const { text : grokResponse } = await aiGenerateText ( {
115
- model : grok . languageModel ( model , {
116
- parallelToolCalls : false ,
117
- } ) ,
118
- prompt : context ,
119
- temperature : temperature ,
120
- maxTokens : max_response_length ,
121
- frequencyPenalty : frequency_penalty ,
122
- presencePenalty : presence_penalty ,
123
- } ) ;
124
-
125
- response = grokResponse ;
126
- console . log ( "Received response from Grok model." ) ;
127
- break ;
128
- }
71
+ case ModelProvider . LLAMACLOUD : {
72
+ console . log ( "Initializing OpenAI model." ) ;
73
+ const openai = createOpenAI ( { apiKey } ) ;
74
+
75
+ const { text : openaiResponse } = await aiGenerateText ( {
76
+ model : openai . languageModel ( model ) ,
77
+ prompt : context ,
78
+ temperature : temperature ,
79
+ maxTokens : max_response_length ,
80
+ frequencyPenalty : frequency_penalty ,
81
+ presencePenalty : presence_penalty ,
82
+ } ) ;
83
+
84
+ response = openaiResponse ;
85
+ console . log ( "Received response from OpenAI model." ) ;
86
+ break ;
87
+ }
88
+
89
+ case ModelProvider . ANTHROPIC : {
90
+ console . log ( "Initializing Anthropic model." ) ;
91
+ const anthropicVertex = createAnthropicVertex ( ) ;
92
+
93
+ const { text : anthropicResponse } = await aiGenerateText ( {
94
+ model : anthropicVertex ( model ) ,
95
+ prompt : context ,
96
+ temperature : temperature ,
97
+ maxTokens : max_response_length ,
98
+ frequencyPenalty : frequency_penalty ,
99
+ presencePenalty : presence_penalty ,
100
+ } ) ;
101
+
102
+ response = anthropicResponse ;
103
+ console . log ( "Received response from Anthropic model." ) ;
104
+ break ;
105
+ }
106
+
107
+ case ModelProvider . GROK : {
108
+ console . log ( "Initializing Grok model." ) ;
109
+ const grok = createGroq ( { apiKey } ) ;
110
+
111
+ const { text : grokResponse } = await aiGenerateText ( {
112
+ model : grok . languageModel ( model , {
113
+ parallelToolCalls : false ,
114
+ } ) ,
115
+ prompt : context ,
116
+ temperature : temperature ,
117
+ maxTokens : max_response_length ,
118
+ frequencyPenalty : frequency_penalty ,
119
+ presencePenalty : presence_penalty ,
120
+ } ) ;
121
+
122
+ response = grokResponse ;
123
+ console . log ( "Received response from Grok model." ) ;
124
+ break ;
125
+ }
129
126
130
127
case ModelProvider . LLAMALOCAL :
131
128
console . log ( "Using local Llama model for text completion." ) ;
@@ -140,12 +137,11 @@ export async function generateText({
140
137
console . log ( "Received response from local Llama model." ) ;
141
138
break ;
142
139
143
- default :
144
- {
145
- const errorMessage = `Unsupported provider: ${ provider } ` ;
146
- console . error ( errorMessage ) ;
147
- throw new Error ( errorMessage ) ;
148
- }
140
+ default : {
141
+ const errorMessage = `Unsupported provider: ${ provider } ` ;
142
+ console . error ( errorMessage ) ;
143
+ throw new Error ( errorMessage ) ;
144
+ }
149
145
}
150
146
151
147
return response ;
0 commit comments