You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Confirm this is an issue with the Python library and not an underlying OpenAI API
This is an issue with the Python library
Describe the bug
For a resume-writing program with multiple levels of async calls, launching relatively small scale async processing causes the API to fail catastrophically.
Attempted the OpenAIClient and httpx.AsyncClient solutions which were suggested here and elsewhere: #769
When called synchronously, code processes 50 resumes sequentially with no problem, and perhaps 3 or 4 'Timeout' failures in aggregate that are successfully completed using exponential backoff. The average completion time for each document is 50 seconds with a std of perhaps 10 seconds.
When the same 50 documents are run simultaneously using asyncio:
await asyncio.gather(*tasks)
Several hundred - several thousand timeout errors occur in aggregate, and most of the time, the processes will fail catastrophically as None is returned by the OpenAI api, which then fails cascadingly throughout the system.
Average completion time rises to 240 seconds with an std of perhaps 30 seconds.
I've confirmed that unique clients are created for each document:
OpenAIClient object at 0x7f9a57762fb0
OpenAIClient object at 0x7f9a5764f430
OpenAIClient object at 0x7f9a57249870
...
Running with a clean new environment updated today:
python==3.10.13
openai==1.12.0
httpx==0.27.0
#769 seems to indicate that the problem was resolved in open 1.3.8, but we can't fix.
To Reproduce
Initiate 50 top-level tasks, each of which fires of approx 100 tasks, each of which may fire 0-5 additional tasks, and may reiterate
Create an AsyncOpenAI Client for each of the 50 toplevel tasks
Observe that OpenAI repeatedly returns thousands of timeout errors
Spawning 5000-25000 requests concurrently would likely hit many of the rate-limiting caps like requests/sec and tokens/sec. Are you ramping up requests or just starting thousands all at once?
Good call, and that’s worth looking in to. In PROD we are fewer than 1/second. Peak testing has been 16,000/ hour, so we haven’t come close to our rate limit of 10,000 RPM (see screenshot)
You can use with_raw_response then extract the remaining tokens and requests for your quota. There's some more details in the docs
response=awaitopenai_client.embeddings.with_raw_response.create(
model="ada-text-002-etc",
input="Ground control to Major Tom",
)
# Get rate limit informationprint(response.headers.get("x-ratelimit-remaining-requests")
print(response.headers.get("x-ratelimit-remaining-tokens")
embeddings=response.parse() # get back the Embeddings object
If you're trying to simulate a production environment at load, I'd recommend ramping up requests or using something like locust. That's what we're using to load test OpenAI endpoints and models.
Confirm this is an issue with the Python library and not an underlying OpenAI API
Describe the bug
For a resume-writing program with multiple levels of async calls, launching relatively small scale async processing causes the API to fail catastrophically.
Attempted the OpenAIClient and httpx.AsyncClient solutions which were suggested here and elsewhere:
#769
When called synchronously, code processes 50 resumes sequentially with no problem, and perhaps 3 or 4 'Timeout' failures in aggregate that are successfully completed using exponential backoff. The average completion time for each document is 50 seconds with a std of perhaps 10 seconds.
When the same 50 documents are run simultaneously using asyncio:
await asyncio.gather(*tasks)
Several hundred - several thousand timeout errors occur in aggregate, and most of the time, the processes will fail catastrophically as None is returned by the OpenAI api, which then fails cascadingly throughout the system.
Average completion time rises to 240 seconds with an std of perhaps 30 seconds.
I've confirmed that unique clients are created for each document:
OpenAIClient object at 0x7f9a57762fb0
OpenAIClient object at 0x7f9a5764f430
OpenAIClient object at 0x7f9a57249870
...
Running with a clean new environment updated today:
python==3.10.13
openai==1.12.0
httpx==0.27.0
#769 seems to indicate that the problem was resolved in open 1.3.8, but we can't fix.
To Reproduce
Code snippets
OS
Amazon Linux
Python version
3.10.13
Library version
openai 1.12.0
The text was updated successfully, but these errors were encountered: