Skip to content

Instantly share code, notes, and snippets.

@jamespantalones
Created July 2, 2020 01:07
Show Gist options
  • Save jamespantalones/75f3a8ac438c6ff03f9e4cfe70c2cb51 to your computer and use it in GitHub Desktop.
Save jamespantalones/75f3a8ac438c6ff03f9e4cfe70c2cb51 to your computer and use it in GitHub Desktop.

Revisions

  1. jamespantalones created this gist Jul 2, 2020.
    152 changes: 152 additions & 0 deletions index.html
    Original file line number Diff line number Diff line change
    @@ -0,0 +1,152 @@
    <!--https://stackoverflow.com/questions/51687308/how-to-use-web-audio-api-to-get-raw-pcm-audio/54213814#54213814-->

    <html><head><meta http-equiv="Content-Type" content="text/html; charset=ISO-8859-1">
    <title>capture microphone then show time & frequency domain output</title>

    <script type="text/javascript">

    var webaudio_tooling_obj = function () {

    var audioContext = new AudioContext();

    console.log("audio is starting up ...");

    var BUFF_SIZE_RENDERER = 16384;
    var SIZE_SHOW = 3; // number of array elements to show in console output

    var audioInput = null,
    microphone_stream = null,
    gain_node = null,
    script_processor_node = null,
    script_processor_analysis_node = null,
    analyser_node = null;

    if (!navigator.getUserMedia)
    navigator.getUserMedia = navigator.getUserMedia || navigator.webkitGetUserMedia ||
    navigator.mozGetUserMedia || navigator.msGetUserMedia;

    if (navigator.getUserMedia){

    navigator.getUserMedia({audio:true},
    function(stream) {
    start_microphone(stream);
    },
    function(e) {
    alert('Error capturing audio.');
    }
    );

    } else { alert('getUserMedia not supported in this browser.'); }

    // ---

    function show_some_data(given_typed_array, num_row_to_display, label) {

    var size_buffer = given_typed_array.length;
    var index = 0;

    console.log("__________ " + label);

    if (label === "time") {

    for (; index < num_row_to_display && index < size_buffer; index += 1) {

    var curr_value_time = (given_typed_array[index] / 128) - 1.0;

    console.log(curr_value_time);
    }

    } else if (label === "frequency") {

    for (; index < num_row_to_display && index < size_buffer; index += 1) {

    console.log(given_typed_array[index]);
    }

    } else {

    throw new Error("ERROR - must pass time or frequency");
    }
    }

    function process_microphone_buffer(event) {

    var i, N, inp, microphone_output_buffer;

    // not needed for basic feature set
    // microphone_output_buffer = event.inputBuffer.getChannelData(0); // just mono - 1 channel for now
    }

    function start_microphone(stream){

    gain_node = audioContext.createGain();
    gain_node.connect( audioContext.destination );

    microphone_stream = audioContext.createMediaStreamSource(stream);
    microphone_stream.connect(gain_node);

    script_processor_node = audioContext.createScriptProcessor(BUFF_SIZE_RENDERER, 1, 1);
    script_processor_node.onaudioprocess = process_microphone_buffer;

    microphone_stream.connect(script_processor_node);

    // --- enable volume control for output speakers

    document.getElementById('volume').addEventListener('change', function() {

    var curr_volume = this.value;
    gain_node.gain.value = curr_volume;

    console.log("curr_volume ", curr_volume);
    });

    // --- setup FFT

    script_processor_analysis_node = audioContext.createScriptProcessor(2048, 1, 1);
    script_processor_analysis_node.connect(gain_node);

    analyser_node = audioContext.createAnalyser();
    analyser_node.smoothingTimeConstant = 0;
    analyser_node.fftSize = 2048;

    microphone_stream.connect(analyser_node);

    analyser_node.connect(script_processor_analysis_node);

    var buffer_length = analyser_node.frequencyBinCount;

    var array_freq_domain = new Uint8Array(buffer_length);
    var array_time_domain = new Uint8Array(buffer_length);

    console.log("buffer_length " + buffer_length);

    script_processor_analysis_node.onaudioprocess = function() {

    // get the average for the first channel
    analyser_node.getByteFrequencyData(array_freq_domain);
    analyser_node.getByteTimeDomainData(array_time_domain);

    // draw the spectrogram
    if (microphone_stream.playbackState == microphone_stream.PLAYING_STATE) {

    show_some_data(array_freq_domain, SIZE_SHOW, "frequency");
    show_some_data(array_time_domain, SIZE_SHOW, "time"); // store this to record to aggregate buffer/file
    }
    };
    }

    }(); // webaudio_tooling_obj = function()

    </script>

    </head>
    <body>

    <p>Volume</p>
    <input id="volume" type="range" min="0" max="1" step="0.1" value="0.0"/>

    <p> </p>
    <button onclick="webaudio_tooling_obj()">start audio</button>

    </body>
    </html>