concurrent model execution #346
Unanswered
leeseungchae
asked this question in
Q&A
Replies: 2 comments 2 replies
-
|
Hi, Can you post a code snippet demonstrating your usage and the performance issue? |
Beta Was this translation helpful? Give feedback.
2 replies
-
|
Same problem. I think we need continuous batching. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment

Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I'm Steve Lee, an AI developer.
The performance improvement of Faster-Whisper is quite surprising.
However, while it's very fast when running a single script, the speed becomes significantly slower
when running multiple requests at the same time. Is there any solution for this concurrency issue?
Additionally, despite adjusting the num_workers and threads,
there's no performance improvement. What could be the reason for this?
Beta Was this translation helpful? Give feedback.
All reactions