Skip to content

Commit b39d2aa

Browse files
committed
newer tests
1 parent 71cadbf commit b39d2aa

10 files changed

+407
-549
lines changed

.env.example

+48-62
Original file line numberDiff line numberDiff line change
@@ -117,77 +117,63 @@ EXPRESS_MAX_PAYLOAD= # Default: 100kb
117117
#######################################
118118

119119
### Provider Settings ###
120-
PROVIDER_ENDPOINT=https://api.openai.com/v1/chat/completions
120+
PROVIDER_ENDPOINT=https://api.openai.com/v1
121121
PROVIDER_NAME=openai
122122
PROVIDER_API_KEY=
123123

124-
### Model Names ###
125-
DEFAULT_MODEL=gpt-3.5-turbo
126-
SMALL_MODEL=gpt-3.5-turbo-instruct
124+
# Model Settings - Default
125+
DEFAULT_MODEL=gpt-4
126+
DEFAULT_MODEL_MAX_INPUT_TOKENS=4096
127+
DEFAULT_MODEL_MAX_OUTPUT_TOKENS=1024
128+
DEFAULT_MODEL_TEMPERATURE=0.7
129+
DEFAULT_MODEL_FREQUENCY_PENALTY=0
130+
DEFAULT_MODEL_PRESENCE_PENALTY=0
131+
DEFAULT_MODEL_REPETITION_PENALTY=1.0
132+
DEFAULT_MODEL_STOP=
133+
134+
# Model Settings - Small
135+
SMALL_MODEL=gpt-3.5-turbo
136+
SMALL_MODEL_MAX_INPUT_TOKENS=2048
137+
SMALL_MODEL_MAX_OUTPUT_TOKENS=1024
138+
SMALL_MODEL_TEMPERATURE=0.7
139+
SMALL_MODEL_FREQUENCY_PENALTY=0
140+
SMALL_MODEL_PRESENCE_PENALTY=0
141+
SMALL_MODEL_REPETITION_PENALTY=1.0
142+
SMALL_MODEL_STOP=
143+
144+
# Model Settings - Medium
127145
MEDIUM_MODEL=gpt-4
146+
MEDIUM_MODEL_MAX_INPUT_TOKENS=2048
147+
MEDIUM_MODEL_MAX_OUTPUT_TOKENS=1024
148+
MEDIUM_MODEL_TEMPERATURE=0.7
149+
MEDIUM_MODEL_FREQUENCY_PENALTY=0
150+
MEDIUM_MODEL_PRESENCE_PENALTY=0
151+
MEDIUM_MODEL_REPETITION_PENALTY=1.0
152+
MEDIUM_MODEL_STOP=
153+
154+
# Model Settings - Large
128155
LARGE_MODEL=gpt-4-turbo
129-
EMBEDDING_MODEL=text-embedding-ada-002
130-
USE_OPENAI_EMBEDDING=true
131-
IMAGE_MODEL=dall-e-3
132-
133-
### Default Model Settings ###
134-
DEFAULT_MAX_INPUT_TOKENS=4096
135-
DEFAULT_MAX_OUTPUT_TOKENS=1024
136-
DEFAULT_TEMPERATURE=0.7
137-
DEFAULT_STOP_SEQUENCES=
138-
DEFAULT_FREQUENCY_PENALTY=0
139-
DEFAULT_PRESENCE_PENALTY=0
140-
DEFAULT_REPETITION_PENALTY=1.0
141-
142-
### Model-Specific Token Limits ###
143-
# Small Model
144-
SMALL_INPUT_TOKENS=2048
145-
SMALL_OUTPUT_TOKENS=512
146-
147-
## Small Model Settings ###
148-
SMALL_MAX_INPUT_TOKENS=4096
149-
SMALL_MAX_OUTPUT_TOKENS=1024
150-
SMALL_TEMPERATURE=0.7
151-
SMALL_STOP_SEQUENCES=###,END
152-
SMALL_FREQUENCY_PENALTY=0
153-
SMALL_PRESENCE_PENALTY=0
154-
SMALL_REPETITION_PENALTY=1.0
155-
156-
# Medium Model
157-
MEDIUM_INPUT_TOKENS=8192
158-
MEDIUM_OUTPUT_TOKENS=2048
159-
160-
## Medium Model Settings ###
161-
MEDIUM_MAX_INPUT_TOKENS=4096
162-
MEDIUM_MAX_OUTPUT_TOKENS=1024
163-
MEDIUM_TEMPERATURE=0.7
164-
MEDIUM_STOP_SEQUENCES=###,END
165-
MEDIUM_FREQUENCY_PENALTY=0
166-
MEDIUM_PRESENCE_PENALTY=0
167-
MEDIUM_REPETITION_PENALTY=1.0
168-
169-
# Large Model
170-
LARGE_INPUT_TOKENS=128000
171-
LARGE_OUTPUT_TOKENS=4096
172-
173-
## Large Model Settings ###
174-
LARGE_MAX_INPUT_TOKENS=4096
175-
LARGE_MAX_OUTPUT_TOKENS=1024
176-
LARGE_TEMPERATURE=0.7
177-
LARGE_STOP_SEQUENCES=###,END
178-
LARGE_FREQUENCY_PENALTY=0
179-
LARGE_PRESENCE_PENALTY=0
180-
LARGE_REPETITION_PENALTY=1.0
181-
182-
183-
184-
### Specialized Model Settings ###
185-
# Embedding Model
156+
LARGE_MODEL_MAX_INPUT_TOKENS=4096
157+
LARGE_MODEL_MAX_OUTPUT_TOKENS=1024
158+
LARGE_MODEL_TEMPERATURE=0.7
159+
LARGE_MODEL_FREQUENCY_PENALTY=0
160+
LARGE_MODEL_PRESENCE_PENALTY=0
161+
LARGE_MODEL_REPETITION_PENALTY=1.0
162+
LARGE_MODEL_STOP=
163+
164+
# Model Settings - Embedding
165+
EMBEDDING_MODEL=text-embedding-3-small
186166
EMBEDDING_DIMENSIONS=1536
187167

188-
# Image Generation
168+
# Model Settings - Image
169+
IMAGE_MODEL=dall-e-3
189170
IMAGE_STEPS=50
190171

172+
# Model Settings - Vision
173+
IMAGE_VISION_MODEL=gpt-4-vision-preview
174+
175+
176+
191177

192178
# Community Plugin for OpenAI Configuration
193179
ENABLE_OPEN_AI_COMMUNITY_PLUGIN=false

.gitignore

+4-1
Original file line numberDiff line numberDiff line change
@@ -91,4 +91,7 @@ lit-config.json
9191

9292
# Configuration to exclude the extra and local_docs directories
9393
extra
94-
**/dist/**
94+
**/dist/**
95+
96+
**/local_cache/**
97+
**/local_cache

packages/client-direct/src/index.ts

+1-1
Original file line numberDiff line numberDiff line change
@@ -302,7 +302,7 @@ export class DirectClient {
302302

303303
// save response to memory
304304
const responseMessage: Memory = {
305-
id: stringToUuid(messageId + "-" + runtime.agentId),
305+
id: stringToUuid(`${messageId}-${runtime.agentId}`),
306306
...userMessage,
307307
userId: runtime.agentId,
308308
content: response,

0 commit comments

Comments
 (0)