Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add duration of audio and VAD removed duration to BatchedInferencePipeline #1186

Open
wants to merge 12 commits into
base: master
Choose a base branch
from

Conversation

greenw0lf
Copy link

With the non-batched version of the WhisperModel, you would get logging output like:

Processing audio with duration 01:33:59.990
VAD filter removed 06:55.648 of audio

Whereas, when calling the BatchedInferencePipeline's transcribe() method, that is no longer the case.

This PR tries to bring that logging back as I believe it does not add extra overhead to the logging and it is quite useful for developers who wish to know how much audio gets processed in the end.

If this is not the case and it is an issue related to my usage of the model, I apologize in advance!

@@ -114,6 +114,7 @@ def __init__(
self,
model,
):
self.logger = get_logger()
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

self.model already has a logger, so I'd rather we use the same logger instead of having duplicate loggers in both the pipeline and the model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants