You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to alignment my audio recording files with corresponding transcripts. There are a lot of pauses and silence in my audios. I want multi-level alignment (mainly word-level and segment/paragraph-level) as well as alignment for the pauses and silence. It is important for me to know how long or how short the inter-segment pauses are. However when I use the command below, it detects no pauses/silence for between segments. While if I use is_text_type=plain instead of mplain, I receive the alignment for those inter-segment pauses (as well as the segments).
I want to alignment my audio recording files with corresponding transcripts. There are a lot of pauses and silence in my audios. I want multi-level alignment (mainly word-level and segment/paragraph-level) as well as alignment for the pauses and silence. It is important for me to know how long or how short the inter-segment pauses are. However when I use the command below, it detects no pauses/silence for between segments. While if I use is_text_type=plain instead of mplain, I receive the alignment for those inter-segment pauses (as well as the segments).
I want to alignment my audio recording files with corresponding transcripts. There are a lot of pauses and silence in my audios. I want multi-level alignment (mainly word-level and segment/paragraph-level) as well as alignment for the pauses and silence. It is important for me to know how long or how short the inter-segment pauses are. However when I use the command below, it detects no pauses/silence for between segments. While if I use
is_text_type=plain
instead ofmplain
, I receive the alignment for those inter-segment pauses (as well as the segments).python -m aeneas.tools.execute_task sample_audio.mp3 sample_audio_transcript.txt "task_language=eng|os_task_file_format=json|is_text_type=mplain|task_adjust_boundary_nonspeech_min=0.0100|task_adjust_boundary_nonspeech_string=(sil)|task_adjust_boundary_algorithm=auto" sample_audio_output.multilevel.json
Why?
The text was updated successfully, but these errors were encountered: