Open bachittle opened 1 month ago
+1
+1
I want to add support for the GPT-4o-realtime model, which relies on WebSocket technology. This necessitates the use of a WebSocket library, but introducing such a dependency conflicts with the zero-dependency philosophy of the existing library, as I previously discussed with @sashabaranov.
As a result, I've decided to create a new library dedicated exclusively to GPT-4o-realtime. This library will serve as a complement to go-openai, focusing solely on supporting GPT-4o-realtime functionality.
The new library is called go-openai-realtime. Feel free to check it out!
@WqyJh, thank you for your effort on this! I think websockets are a fair case to introduce a dependency to this library, and I would love to merge your changes if you'll decide to contribute 🙌🏻
They just released support for audio input and output in the chat completions endpoint, using the gpt-4o-audio-preview model. This could be supported first in the meantime: https://platform.openai.com/docs/guides/audio/quickstart
They just released support for audio input and output in the chat completions endpoint, using the gpt-4o-audio-preview model. This could be supported first in the meantime: https://platform.openai.com/docs/guides/audio/quickstart
I just added support for gpt-4o-audio-preview. See https://github.com/sashabaranov/go-openai/pull/895
@WqyJh, thank you for your effort on this! I think websockets are a fair case to introduce a dependency to this library, and I would love to merge your changes if you'll decide to contribute 🙌🏻
I'd like to contribute to the project. Since it contains a lot of code and examples, it will take some time to complete. Mixing all the code together would create a mess, so I suggest organizing all the real-time code into a folder named realtime.
https://platform.openai.com/docs/guides/realtime
it uses websockets to get and set audio files from gpt-4o-realtime.