feat:max_token配置
This commit is contained in:
@@ -48,6 +48,8 @@ def create_vana():
|
||||
"api_key": config('CHAT_MODEL_API_KEY', default=''),
|
||||
"api_base": config('CHAT_MODEL_BASE_URL', default=''),
|
||||
"model": config('CHAT_MODEL_NAME', default=''),
|
||||
'temperature':config('CHAT_MODEL_TEMPERATURE', default=0.7, cast=float),
|
||||
'max_tokens':config('CHAT_MODEL_MAX_TOKEN', default=20000),
|
||||
},
|
||||
)
|
||||
|
||||
|
||||
@@ -26,7 +26,7 @@ class OpenAICompatibleLLM(VannaBase):
|
||||
def __init__(self, client=None, config_file=None):
|
||||
VannaBase.__init__(self, config=config_file)
|
||||
# default parameters - can be overrided using config
|
||||
self.temperature = 0.5
|
||||
self.temperature = 0.6
|
||||
self.max_tokens = 5000
|
||||
|
||||
if "temperature" in config_file:
|
||||
@@ -175,6 +175,7 @@ class OpenAICompatibleLLM(VannaBase):
|
||||
print(
|
||||
f"Using model {self.config['model']} for {num_tokens} tokens (approx)"
|
||||
)
|
||||
print(self.config)
|
||||
response = self.client.chat.completions.create(
|
||||
model=self.config["model"],
|
||||
messages=prompt,
|
||||
@@ -197,7 +198,6 @@ class OpenAICompatibleLLM(VannaBase):
|
||||
stop=None,
|
||||
temperature=self.temperature,
|
||||
)
|
||||
|
||||
for choice in response.choices:
|
||||
if "text" in choice:
|
||||
return choice.text
|
||||
|
||||
Reference in New Issue
Block a user