Source: core/speech-input.js

/** 
 * Experimental speech input, uses annyang library.
 * Seems to work well on chrome and edge on PC.
 * Kind of works on android chrome, with limitations, but WebRTC disables it for good, thus, disabled if MediaStreams are enabled.  
 * */
export class SpeechInput {
  static instances = [];
  static enabled = true;
  static active = false;
  static android = (typeof navigator !== 'undefined' && navigator.userAgent.toLowerCase().indexOf('android') > -1);
  static touchListener = null;
  static mediaStreams = null;
  constructor() {
    this.commands = {};
    this.noMatch = null;
    this.onlyLetters = false;
    this.lowercase = true;
    this.removePeriod = true;
    this.spoke = false;
    this.constructor.instances.push(this);
    // this should go to static block but then jsdoc fails:
    if ( ! this.constructor.touchListener ) {
      this.constructor.touchListener = (e) => {
        this.continue();
      }
      document.addEventListener('touchstart', this.constructor.touchListener);
    }
  }
  continue() {
    if ( this.constructor.active && this.constructor.android && !this.constructor.mediaStreams ) {
      //console.log("Android speech recognition (re) starting");
      if ( this.constructor.mediaStreams ) {
        this.constructor.mediaStreams.publishAudio(false);
      }
      annyang.start({autoRestart:false, continuous:true});
    }
  }
  addCommand(command, callback, text) {
    if ( text ) {
      text = " "+text;
    } else {
      text = "";
    }
    let callbacks = [ callback ];
    if ( this.commands[command] ) {
      // adding another callback to existing command
      callbacks.push(callback);
      console.log("Callback added to "+command, callbacks);
    }
    this.commands[command+text] = (text) => this.callback(command, text, callbacks);
    // microsoft apparently attempts to add punctuation
    this.commands[command+'.'+text] = (text) => this.callback(command, text, callbacks);
    this.commands[command+','+text] = (text) => this.callback(command, text, callbacks);
  }
  
  callback(command, text, callbacks) {
    //console.log("Executing "+text, callback);
    if ( text ) {
      if (this.lowercase) {
        text = text.toLowerCase();
      }
      if (this.onlyLetters) {
        text = text.replace(/[^a-zA-Z ]/g, "");
      }
      if (this.removePeriod && text.endsWith(".")) {
        text = text.substring(0,text.length-1);
      }
    }
    this.spoke=true;
    //console.log("Spoke:"+ command+" "+text);
    callbacks.forEach(callback=>callback(text));
  }
  callNoMatch(phrases) {
    this.spoke=true;
    //console.log("Spoke:"+ phrases);
    if ( this.noMatch ) {
      this.noMatch(phrases);
    }
  }
  addNoMatch(callback) {
    this.noMatch = callback;
  }
  endCallback() {
    //console.log("Speech recognition ended, spoke: "+this.spoke+" active:"+this.constructor.active);
    if ( this.spoke ) {
      this.spoke = false;
      this.continue();
    } else {
      // silence/stop
      //console.log("Speech recognition ended in silence");
      if ( this.constructor.mediaStreams ) {
        this.constructor.mediaStreams.publishAudio(true);
      }
    }
  }
  
  static available() {
    return typeof(annyang) != 'undefined' && annyang;
  }
  
  static isEnabled() {
    return SpeechInput.enabled && SpeechInput.available();
  }
  
  start() {
    if ( SpeechInput.enabled && SpeechInput.available() ) {
      let index = this.constructor.instances.indexOf(this);
      if ( index < 0 ) {
        // this instance might have been disposed, kept elsewhere, and restarted
        this.constructor.instances.push(this);
      }
      // Add our commands to annyang
      if ( this.commands ) {
        annyang.addCommands(this.commands);
        //console.log(this.commands);
      }
      if ( this.noMatch ) {
        annyang.addCallback('resultNoMatch', (phrases)=>this.callNoMatch(phrases));
      }
      if ( this.constructor.android && ! this.end ) {
        this.end = () => this.endCallback();
        annyang.addCallback('end', this.end );
      }
      // Start listening. You can call this here, or attach this call to an event, button, etc.
      if ( this.constructor.android ) {
        //console.log("Speech recognition will start on touch, to prevent annoying beeping on android");
      } else {
        annyang.start();
        //console.log("Speech recognition started: "+annyang.isListening(), this.commands);
      }
      this.constructor.active = true;
    } else {
      console.log("Speech recognition unavailable");
    }
  }
  stop() {
    if ( annyang ) {
      //console.log("speech recognition stopped");
      annyang.abort();
      this.constructor.active = false;
    }
  }
  dispose() {
    let index = this.constructor.instances.indexOf(this);
    if ( index >= 0 ) {
      // index could be -1 if dispose is called more than once
      this.constructor.instances.splice(index,1);
      if ( this.constructor.instances.length > 0 ) {
        this.constructor.instances[this.constructor.instances.length -1].start();
      }
    }
    if (typeof annyang !=='undefined' && annyang) {
      this.stop();
      if ( this.commands ) {
        // annyang expects array of phrases as argument
        annyang.removeCommands(Object.keys(this.commands));
        //console.log(' disabled commands:', Object.keys(this.commands));
      }
      if ( this.noMatch ) {
        annyang.removeCallback('resultNoMatch', this.noMatch);
      }
      if ( this.end ) {
        annyang.removeCallback('end', this.end);
        delete this.end;
      }
    }
  }
}