Description
Your current environment
Hi everyone,
I'm trying to run the code in examples/online_serving/openai_transcription_client.py
I'm using vllm 0.8.4
🐛 Describe the bug
INFO 04-28 17:06:48 [init.py:239] Automatically detected platform cuda.
transcription result: The first words I spoke in the original phonograph, a little piece of practical poetry. Mary had a little lamb, its streets were quite as slow, and everywhere that Mary went the lamb was sure to go.
transcription result: Traceback (most recent call last):
File "/home/aptikal/abdalfar/InterviewSim/vllm_project/openai_transcription_client.py", line 66, in
asyncio.run(stream_openai_response())
File "/usr/lib/python3.11/asyncio/runners.py", line 190, in run
return runner.run(main)
^^^^^^^^^^^^^^^^
File "/usr/lib/python3.11/asyncio/runners.py", line 118, in run
return self._loop.run_until_complete(task)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/lib/python3.11/asyncio/base_events.py", line 653, in run_until_complete
return future.result()
^^^^^^^^^^^^^^^
File "/home/aptikal/abdalfar/InterviewSim/vllm_project/openai_transcription_client.py", line 60, in stream_openai_response
content = chunk['choices'][0].get('delta',
~~~~~^^^^^^^^^^^
KeyError: 'choices'
Any idea how to solve this ?
Thank you
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.