You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I am trying to batch the input to the pytorch model, but the batched input works for when the model is in cpu. But when the model is in gpu, the inference only works for the first time. After that it shows the below error occurs
Hi I am having the same issue. The reason for batched inference on GPU is for large batches in very low-latency streaming scenario. The cpu version is relatively slow when batch size exceeds a couple tens of items. Are there any plans to make this model usable on GPU?
I ran a sanity check using a 4-second clip at 16 KHz. I split the array into chunks of 512. Unfortunately, I get very different probabilities in batched mode vs. frame-by-frame inference.
The frame-by-frame probabilities seem correct. When there is speech, the pytorch model outputs a value close to 1. The same frame will output a much lower value, e.g. 0.15, if it is part of a batch.
馃悰 Bug
Hi,
I am trying to batch the input to the pytorch model, but the batched input works for when the model is in cpu. But when the model is in gpu, the inference only works for the first time. After that it shows the below error occurs
code
If I run the above code twice, then this error occurs.
The text was updated successfully, but these errors were encountered: