Refactor OpenAI response handling to use updated response types
This commit is contained in:
parent
f0f4e3c9b7
commit
7a6b4d8fce
1 changed files with 4 additions and 4 deletions
8
misc.py
8
misc.py
|
|
@ -15,7 +15,7 @@ if TYPE_CHECKING:
|
||||||
from discord.guild import GuildChannel
|
from discord.guild import GuildChannel
|
||||||
from discord.interactions import InteractionChannel
|
from discord.interactions import InteractionChannel
|
||||||
from openai import OpenAI
|
from openai import OpenAI
|
||||||
from openai.types.chat.chat_completion import ChatCompletion
|
from openai.types.responses import Response
|
||||||
|
|
||||||
|
|
||||||
logger: logging.Logger = logging.getLogger(__name__)
|
logger: logging.Logger = logging.getLogger(__name__)
|
||||||
|
|
@ -194,11 +194,11 @@ def chat( # noqa: PLR0913, PLR0917
|
||||||
"User message ends here.\n"
|
"User message ends here.\n"
|
||||||
)
|
)
|
||||||
|
|
||||||
completion: ChatCompletion = openai_client.chat.completions.create(
|
resp: Response = openai_client.responses.create(
|
||||||
model="gpt-5-chat-latest",
|
model="gpt-5-chat-latest",
|
||||||
messages=[{"role": "system", "content": prompt}],
|
input=[{"role": "user", "content": prompt}],
|
||||||
)
|
)
|
||||||
response: str | None = completion.choices[0].message.content
|
response: str | None = resp.output_text
|
||||||
logger.info("AI response: %s", response)
|
logger.info("AI response: %s", response)
|
||||||
|
|
||||||
return response
|
return response
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue