You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this demo, the chunk_size is defined as chunk_size = lookahead_size + ENCODER_STEP_LENGTH, but the default setting results in chunks that are too small for my use case. I’d like to process the audio input in approximately 1-second increments.
However, simply changing ENCODER_STEP_LENGTH to 1000 didn’t produce the expected results. Could anyone provide guidance on how to properly adjust the chunk_size or any other relevant parameters to achieve this? Any detailed explanation or example would be greatly appreciated!
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Hi, I’m using the Online_ASR_Microphone_Demo_Cache_Aware_Streaming.ipynb notebook and referring to the following code snippet:
In this demo, the chunk_size is defined as
chunk_size = lookahead_size + ENCODER_STEP_LENGTH
, but the default setting results in chunks that are too small for my use case. I’d like to process the audio input in approximately 1-second increments.However, simply changing ENCODER_STEP_LENGTH to 1000 didn’t produce the expected results. Could anyone provide guidance on how to properly adjust the chunk_size or any other relevant parameters to achieve this? Any detailed explanation or example would be greatly appreciated!
Beta Was this translation helpful? Give feedback.
All reactions