Skip to content

Instantly share code, notes, and snippets.

Show Gist options
  • Select an option

  • Save NickyDark1/d44537bdedd727f415b8d69b0ca2e107 to your computer and use it in GitHub Desktop.

Select an option

Save NickyDark1/d44537bdedd727f415b8d69b0ca2e107 to your computer and use it in GitHub Desktop.

Revisions

  1. @Vaibhavs10 Vaibhavs10 created this gist Jan 24, 2023.
    26 changes: 26 additions & 0 deletions live_whisper_inference.py
    Original file line number Diff line number Diff line change
    @@ -0,0 +1,26 @@
    #pip install git+https://github.com/huggingface/transformers.git

    import datetime
    import sys
    from transformers import pipeline
    from transformers.pipelines.audio_utils import ffmpeg_microphone_live

    pipe = pipeline("automatic-speech-recognition", model="openai/whisper-base", device=0)
    sampling_rate = pipe.feature_extractor.sampling_rate


    start = datetime.datetime.now()

    chunk_length_s = 5
    stream_chunk_s = 0.1
    mic = ffmpeg_microphone_live(
    sampling_rate=sampling_rate,
    chunk_length_s=chunk_length_s,
    stream_chunk_s=stream_chunk_s,
    )
    print("Start talking...")
    for item in pipe(mic):
    sys.stdout.write("\033[K")
    print(item["text"], end="\r")
    if not item["partial"][0]:
    print("")