Add messages regarding the use of token during query (#147)
* Add messages regarding the use of token during query * fix: apply ruff format
This commit is contained in:
@@ -834,6 +834,11 @@ class OpenAIChat(LLMInterface):
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
response = self.client.chat.completions.create(**params)
|
response = self.client.chat.completions.create(**params)
|
||||||
|
print(
|
||||||
|
f"Total tokens = {response.usage.total_tokens}, prompt tokens = {response.usage.prompt_tokens}, completion tokens = {response.usage.completion_tokens}"
|
||||||
|
)
|
||||||
|
if response.choices[0].finish_reason == "length":
|
||||||
|
print("The query is exceeding the maximum allowed number of tokens")
|
||||||
return response.choices[0].message.content.strip()
|
return response.choices[0].message.content.strip()
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Error communicating with OpenAI: {e}")
|
logger.error(f"Error communicating with OpenAI: {e}")
|
||||||
|
|||||||
Reference in New Issue
Block a user