You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello
I would like to understand where comes an extra latency when executing a model inference:
if I run for (N times) model.infer(); sleep(1s); repeat ...
the very first model.infer() (within a loop) has always a penalty (instead of 60ms, it goes up to 110ms). Without the sleep(), no penalty (the first of the first loop always has, but then the repeated loops dont show this latency).
is this expected? how to mitigate ? (without doing busy dummy infer)
is that related to how tbb worker threads are rescheduled/initialized ? something else?
thanks!
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello
I would like to understand where comes an extra latency when executing a model inference:
if I run for (N times) model.infer(); sleep(1s); repeat ...
the very first model.infer() (within a loop) has always a penalty (instead of 60ms, it goes up to 110ms). Without the sleep(), no penalty (the first of the first loop always has, but then the repeated loops dont show this latency).
is this expected? how to mitigate ? (without doing busy dummy infer)
is that related to how tbb worker threads are rescheduled/initialized ? something else?
thanks!
Beta Was this translation helpful? Give feedback.
All reactions