<?xml version="1.0" encoding="UTF-8" standalone="yes" ?>
<!DOCTYPE bugzilla SYSTEM "https://www.w3.org/Bugs/Public/page.cgi?id=bugzilla.dtd">

<bugzilla version="5.0.4"
          urlbase="https://www.w3.org/Bugs/Public/"
          
          maintainer="sysbot+bugzilla@w3.org"
>

    <bug>
          <bug_id>26336</bug_id>
          
          <creation_ts>2014-07-15 02:12:00 +0000</creation_ts>
          <short_desc>Support speech recognition on specific media stream</short_desc>
          <delta_ts>2018-08-06 10:32:18 +0000</delta_ts>
          <reporter_accessible>1</reporter_accessible>
          <cclist_accessible>1</cclist_accessible>
          <classification_id>1</classification_id>
          <classification>Unclassified</classification>
          <product>Speech API</product>
          <component>Speech API</component>
          <version>unspecified</version>
          <rep_platform>PC</rep_platform>
          <op_sys>All</op_sys>
          <bug_status>RESOLVED</bug_status>
          <resolution>WONTFIX</resolution>
          
          
          <bug_file_loc></bug_file_loc>
          <status_whiteboard></status_whiteboard>
          <keywords></keywords>
          <priority>P2</priority>
          <bug_severity>normal</bug_severity>
          <target_milestone>---</target_milestone>
          
          
          <everconfirmed>1</everconfirmed>
          <reporter name="Shih-Chiang Chien">schien</reporter>
          <assigned_to name="Glen Shires">gshires</assigned_to>
          <cc>philip</cc>
    
    <cc>xians</cc>
          
          

      

      

      

          <comment_sort_order>oldest_to_newest</comment_sort_order>  
          <long_desc isprivate="0" >
    <commentid>108932</commentid>
    <comment_count>0</comment_count>
    <who name="Shih-Chiang Chien">schien</who>
    <bug_when>2014-07-15 02:12:00 +0000</bug_when>
    <thetext>Current speech recognition API cannot specify media stream, we could introduce an optional parameter in SpeechRecognition.start() to enable following use cases:
  1. multiple microphone selection via gUM media constraints
  2. remote audio stream (WebRTC)
  3. stream from audio file</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>112564</commentid>
    <comment_count>1</comment_count>
    <who name="xians">xians</who>
    <bug_when>2014-10-02 10:08:32 +0000</bug_when>
    <thetext>We are working on hooking up gUM audio track with WebSpeech in Chrome, but we are not going to support 2, 3 due to concerns of server abuse. That says, we only allow hooking up audio track from microphone to WebSpeech, a track uses non-microphone source (like a file or remote audio track) will be thrown a exception when connecting to WebSpeech.

The new API allows WebSpeech to benefit from gUM technologies, ex. AEC. This will substantially improve the recognition performance during a conference call.</thetext>
  </long_desc><long_desc isprivate="0" >
    <commentid>129312</commentid>
    <comment_count>2</comment_count>
    <who name="Philip Jägenstedt">philip</who>
    <bug_when>2018-08-06 10:32:18 +0000</bug_when>
    <thetext>This work was started in Chrome in https://crbug.com/408940 but wasn&apos;t finished and then removed again. To revisit this issue, I suggest filing an issue on https://github.com/w3c/speech-api and getting implementers talking to each other.</thetext>
  </long_desc>
      
      

    </bug>

</bugzilla>