16:31:50 RRSAgent has joined #htmlspeech 16:31:50 logging to http://www.w3.org/2010/11/11-htmlspeech-irc 16:36:36 zakim, nick burn is DanBurnett 16:36:36 sorry, burn, I do not see a party named 'DanBurnett' 16:40:11 rrsagent, bye 16:41:28 rrsagent, draft minutes 16:41:28 I have made the request to generate http://www.w3.org/2010/11/11-htmlspeech-minutes.html burn 16:41:31 rrsagent, bye 16:57:03 marc has joined #htmlspeech 16:57:08 INC_(HTMLSPEECH)12:00PM has now started 16:57:15 +[Microsoft] 16:57:46 zakim, code? 16:57:46 the conference code is 48657 (tel:+1.617.761.6200 tel:+33.4.26.46.79.03 tel:+44.203.318.0479), burn 16:58:32 ddahl has joined #htmlspeech 16:58:32 +??P2 16:58:37 +??P7 16:59:14 +Debbie_Dahl 16:59:29 trackbot start telcon 16:59:54 zakim, mute P7 16:59:54 sorry, burn, I do not know which phone connection belongs to P7 17:00:02 mbodell has joined #htmlspeech 17:00:04 trackbot, start telcon 17:00:04 zakim, mute ??P7 17:00:04 ??P7 should now be muted 17:00:06 +??P14 17:00:07 RRSAgent, make logs public 17:00:09 Zakim, this will be 17:00:09 I don't understand 'this will be', trackbot 17:00:10 Meeting: HTML Speech Incubator Group Teleconference 17:00:10 Date: 11 November 2010 17:00:11 yes 17:00:16 zakim, unmute ??P7 17:00:16 ??P7 should no longer be muted 17:00:22 Zakim, ??P14 is Olli_Pettay 17:00:22 +Olli_Pettay; got it 17:00:32 Zakim, smaug is Olli_Pettay 17:00:32 sorry, smaug_, I do not recognize a party named 'smaug' 17:00:35 + +39.335.766.aaaa 17:00:38 Zakim, smaug_ is Olli_Pettay 17:00:38 sorry, smaug_, I do not recognize a party named 'smaug_' 17:00:39 zakim, ??P7 is Marc_Schroeder 17:00:40 +Marc_Schroeder; got it 17:00:43 chair: Dan_Burnett 17:00:55 zakim, ??P2 is Dan_Burnett 17:00:55 +Dan_Burnett; got it 17:00:56 +Michael_Bodell 17:01:10 zakim, nick burn is Dan_Burnett 17:01:10 ok, burn, I now associate you with Dan_Burnett 17:01:18 ah 17:01:29 Zakim, nick smaug_ is Olli_Pettay 17:01:29 ok, smaug_, I now associate you with Olli_Pettay 17:01:40 zakim, nick marc is Marc_Schroeder 17:01:40 ok, marc, I now associate you with Marc_Schroeder 17:01:45 zakim, who's on the phone? 17:01:45 On the phone I see [Microsoft], Dan_Burnett, Marc_Schroeder, Debbie_Dahl, Olli_Pettay, +39.335.766.aaaa, Michael_Bodell 17:01:51 zakim, nick ddahl is Debbie_Dahl 17:01:51 ok, ddahl, I now associate you with Debbie_Dahl 17:01:55 zakim, aaaa is Paolo_Baggia 17:01:55 +Paolo_Baggia; got it 17:02:00 +Milan_Young 17:02:05 + +1.760.705.aabb 17:02:35 zakim, aabb is Bjorn_Bringert 17:02:37 +Bjorn_Bringert; got it 17:03:06 Milan has joined #htmlspeech 17:03:16 zakim, who's on the phone? 17:03:16 On the phone I see [Microsoft], Dan_Burnett, Marc_Schroeder, Debbie_Dahl, Olli_Pettay, Paolo_Baggia, Michael_Bodell, Milan_Young, Bjorn_Bringert 17:03:26 rrsagent, start logging 17:03:26 I'm logging. I don't understand 'start logging', ddahl. Try /msg RRSAgent help 17:04:01 + +1.425.391.aacc 17:04:38 zakim, [Microsoft] is temporarily Robert_Brown 17:04:39 +Robert_Brown; got it 17:04:59 zakim, aacc is Dan_Druta 17:04:59 +Dan_Druta; got it 17:05:11 zakim, who's here? 17:05:20 On the phone I see Robert_Brown, Dan_Burnett, Marc_Schroeder, Debbie_Dahl, Olli_Pettay, Paolo_Baggia, Michael_Bodell, Milan_Young, Bjorn_Bringert, Dan_Druta 17:05:23 On IRC I see Milan, mbodell, ddahl, marc, RRSAgent, Zakim, burn, smaug_, trackbot 17:05:47 zakim, nick Milan is Milan_Young 17:05:47 ok, burn, I now associate Milan with Milan_Young 17:05:59 zakim, nick ddahl is Debbie_Dahl 17:05:59 ok, burn, I now associate ddahl with Debbie_Dahl 17:06:13 zakim, nick mbodell is Michael_Bodell 17:06:13 ok, burn, I now associate mbodell with Michael_Bodell 17:06:56 topic: minute takers and expectations 17:07:30 scribe: ddahl 17:08:03 dan: important to start and end on time 17:08:49 ... in the future we will track when people take minutes, also will prefer to to select people who join late and haven't taken minutes 17:09:16 bringert has joined #htmlspeech 17:09:20 ...will start asking newer people in the coming weeks 17:09:52 ...suggestion that we check on f2f minutes and requirements draft 17:10:25 ...send minor corrections to minutes by email, or bring up major corrections now. 17:10:55 ...new requirements draft sent out based on f2f discussion by Michael Bodell 17:11:48 http://lists.w3.org/Archives/Public/public-xg-htmlspeech/2010Nov/0102.html 17:11:49 michael: started new section and struck out requirements that have been deleted. the new draft just covers f2f 17:12:08 dan: concerns about requirements? 17:12:12 agenda is at http://lists.w3.org/Archives/Public/public-xg-htmlspeech/2010Nov/0104.html 17:12:33 zakim, who is noisy? 17:12:44 burn, listening for 10 seconds I heard sound from the following: Marc_Schroeder (85%), Paolo_Baggia (14%) 17:13:02 marc: should this requirements document be at a stable uri? 17:13:20 michael: would like to do this, but still need CVS access 17:13:32 dan: we're going through this process as quickly as we can 17:13:54 ...this is supposed to be more of a live draft 17:14:01 ...comments on requirements? 17:14:25 paolo: would like to know more about section 4, is that the place where the new changes will be placed? 17:14:30 michael: yes 17:14:58 ...as we go through this, all of section 3 will be moved, but section 4 will replace it 17:15:16 ...confusing to mix old and new requirements 17:15:40 dan: paolo, were you asking if there was a way to link old and new requirements? 17:16:00 paolo: yes, and in the old one there was explanation 17:16:30 michael: was mainly trying to capture what we agreed on rather than include a lot of text 17:16:42 ...new ones link back to old ones 17:17:13 zakim, who is noisy? 17:17:24 marc, listening for 10 seconds I could not identify any sounds 17:17:40 dan: this approach reduces confusion about which are the new requirements, when we finish the old requirements I will ask if there's anything that didn't get captured, then it will be safe to remove old requirements. 17:17:50 topic: requirements 17:18:07 dan: start with R21 17:18:55 michael: status is that there was mostly agreement that it was out of scope, was not sure what the scenario was that Eric Johansson had in mind. 17:19:53 dan: recommend drawing a line through this requirement and letting Eric raise concerns 17:20:11 zakim, nick bringert is Bjorn_Bringert 17:20:11 ok, burn, I now associate bringert with Bjorn_Bringert 17:20:22 bjorn: i think Eric agreed that it was out of scope 17:20:30 dan: he may want to continue discussion 17:20:38 dan: R6 17:21:08 michael: we already covered this in discussion of R27 17:21:17 for the minutes, also note that we explicitly had consensus on the call to remove R21 17:21:46 dan: does anyone object to removing R6? 17:21:50 (no objections) 17:21:59 topic: R17 17:22:29 michael: little discussion, but some bleeded over from R18 17:22:50 ...should have requirement that API for recognition should not introduce unneeded latency 17:23:15 s/dan: R6/topic:R6 17:23:34 dan: any more comments on R17? 17:23:55 my connection dropped, dialing again 17:23:56 ...any objections to Michael's wording to replace R17 17:24:25 michael: two requirements, one that Bjorn proposed and one that Michael proposed 17:25:13 ...first one is that applications can start processing captured audio right away and second says that applications can introduce unneeded latency 17:25:37 s/can introduce/should not/ 17:25:50 dan: any objections to replacing original R17 with these two requirements? 17:25:59 (no objections) 17:26:01 "Implementations should be allowed to start processing captured audio before the capture completes." and "The API to do recognition should not introduce unneeded latency." 17:26:13 topic: R18 17:26:14 (w3c doesn't answer the phone), no objection from me 17:26:37 bjorn, just try again. zakim does this sometimes 17:27:23 dan: we acknowledge no objections to R17 (including from bjorn) 17:28:00 ...we will do what we can on R18, will get started while waiting for bjorn 17:28:29 + +44.122.546.aadd 17:28:31 dan: any objections? 17:29:08 zakim, aadd is Bjorn_Bringert 17:29:08 +Bjorn_Bringert; got it 17:29:54 Implementations should be allowed to start playing back synthesized speech before the complete results of the speech synthesis request are available. 17:30:25 dan: any objections to adding this requirement? 17:30:47 michael: other proposals that may have come out of this discussion. 17:31:21 ... other issues about what user agent must allow, codecs, not sure how to tackle 17:31:38 dan: there are others where we're closed to consensus, so let's start with those 17:32:00 ...some of the ones proposed by Milan were addressing other wording on email list 17:32:20 milan: first two, about writing and reading to the audio buffer is handled 17:32:42 bjorn: we might have consensus on the requirement about unneeded latency 17:32:49 ...for TTS 17:33:15 marc: we agree with this 17:33:49 michael: that takes care of the first 3, and the others aren't needed 17:34:10 ...then we had user agents must not filter results from the application 17:34:41 no, not "the others aren't needed" -- instead, the next 3 are not needed 17:35:02 ... through filtering results 17:35:17 bjorn: the next two are that the ua must not interfere with result or timing 17:35:42 oops, i meant through passing parameters 17:35:58 yes, we are now discussing results and timing 17:36:17 bjorn: you might have some results that are suspicious 17:36:27 ..that the ua might want to filter 17:36:39 marc: what speaks in favor of these requirements? 17:37:23 milan: concerned that ua think that it knows best about the speech interaction and deleting something that was part of the protocol between the speech resource and the web application 17:38:14 michael: api needs to be extensible so that additional functionality is supported? 17:39:09 milan: actually wants to make sure that if an EMMA result is sent to the web app the ua must not change it because it doesn't know what it is? also wants events fired by the speech resource to make it into the user space. 17:39:22 bjorn: what kind of event are you talking about? 17:39:35 actually, that was olli 17:39:42 (not bjorn) 17:39:59 Yeah, I think ealier too about supspicious results was also olli 17:40:14 milan: if we have a "start of speech" event, as long as our API is flexible enough we can add new things 17:40:31 s/bjorn: what/olli: what 17:41:12 s/bjorn: you/olli: you 17:42:13 milan: API should be flexible so that new events don't break apps 17:42:47 bjorn: how about if we say that it should be possible to add new information to speech recognition results 17:43:16 ...then for events we would have to require that speech server specific events would be able to be returned to the web app 17:44:01 milan: web applications need to be able to continue to run even if they aren't expecting them 17:44:15 bjorn: can't expect that app will never crash 17:44:34 ...so speech server should be able to return implementation-specific events 17:44:55 ...two new requirements 17:45:32 new requirement 1: speech recognition implementations should be allowed to add implementation specific information to speech recognition results 17:46:29 new requirement 2: speech recognition implementations should be allowed to fire implementation specific events 17:47:41 milan: previously we had discussed that there was a particular ordering of events, if we just have events being fired in between those events, we might destabilize applications 17:48:24 bjorn: the apps that don't care about events won't register for them. and if we say "a before b" that doesn't mean that there's anything in between. 17:49:11 milan: if speech server doesn't generate "start of speech" does the ua insert it? 17:49:21 bjorn: we haven't defined events 17:49:25 s/that doesn't mean that there's anything/that doesn't mean that there's not anything/ 17:49:32 -Paolo_Baggia 17:49:41 michael: it could be the ua or the speech resource that generates some of these events 17:50:18 dan: any objections to adding these two requirements as stated on IRC 17:50:44 milan: are we going to allow TTS to fire events? these just talk about recognition. 17:50:53 +Paolo_Baggia 17:50:56 marc: it would make sense to have that? 17:51:33 michael: rather than changing R2 to speech resources would it make sense to have new requirement for TTS events? 17:51:49 bjorn: other use cases for TTS events other than "mark"? 17:52:11 marc: yes, "mark" is too coarse-grained for lip synchronization 17:52:26 new requirement 3: speech synthesis implementations should be allowed to fire implementation specific events 17:52:55 dan: sounds like we have agreement on these three 17:53:07 michael: back to R18 17:53:30 dan: should be split into two, first is exactly as worded 17:54:21 michael: i.e. requires support for remote services, e.g. by HTTP 1.1 17:54:45 dan: yes, second is that speech services and ua may negotiate on other protocols 17:55:13 dan: first one is mandatory support for HTTP 1.1 17:55:30 michael: does HTTP 1.1 include https? 17:55:59 marc: this seems like a low-level technical requirement, should not discuss now 17:56:30 dan: two issues -- we use a generic term "communication" but we might need to distinguish control and media 17:56:59 ...at IETF there is a lot of discussion for using HTTP for streaming, when should you use that or not 17:57:27 ...we might not want to define these protocols, just pick from what's available, but may not want to pick now 17:57:47 marc: would like to settle on some existing protocol, but now is too early 17:58:04 michael: likes rewording with "such as" 17:58:30 olli: likes "lowest common denominator" 17:58:45 that wasn't me 17:58:57 that was robert 17:59:13 s/olli: likes/robert: likes 18:01:00 marc: requirement should be: the communication between the us and the speech server must allow for some lowest common denominator like HTTP 1.1 18:01:07 s/us/ua 18:01:43 dan: we want to require a lowest common denominator protocol 18:02:14 bjorn: we want to avoid mismatch between browser and speech server communication 18:02:18 i said "require a mandatory-to-support communication protocol, such protocol TBD" 18:03:45 michael: the communication between the us and the speech server must require a mandatory-to-support lowest common denominator such as HTTP 1.1, TBD 18:03:51 s/us/ua 18:05:07 dan: the reason for this is that four months from now it might be construed as requiring HTTP 1.1 18:05:21 michael: do we have agreement on this? 18:06:01 ...ok what about the second sentence? 18:07:02 dan: we could write a requirement but may not end up considering it as something that we need to do now 18:07:20 ...we don't want to prevent negotiation in the future 18:07:56 robert: negotiation sounds like a runtime handshake but what we really want is the freedom to use something else if something better shows up 18:08:55 ...in telephony negotiation is important, but in web apps you just ask for what you want and if it doesn't work, try something else 18:09:05 dan: this concept is important to capture 18:09:13 michael: requirement text? 18:09:23 what i had proposed initially was: "UAs and speech services may negotiate on use of other protocols for communication." 18:10:34 ... another wording: "UAs and speech services may agree to use alternate protocols for communication." 18:10:41 dan: agreement? 18:10:49 (no objections) 18:10:58 agree with another wording 18:12:18 marc: another one for R18, should we have one for implementation data for TTS, a mirror image to the new one we added for ASR 18:12:31 robert: we added that at the f2f 18:12:40 michael: two more 18:12:59 ...require ua's to expose an API for local speech services 18:13:13 bjorn: i don't see this as necessary 18:13:20 michael: i agree 18:13:35 dan: do you think that whatever API we define should be sufficient? 18:14:05 bjorn: the ua is free to talk to any speech service whenever it wants 18:14:14 my question is whether even in the local case that the single api we are defining should be required 18:14:26 ... and that UAs can optimize the behavior 18:14:46 or are you saying that the UA can do anything it wants any way it wants when the resource is local? 18:17:17 bjorn: as long as the specification is specified, if the web app requests a local speech service but it's not available, what happens 18:17:17 -Paolo_Baggia 18:18:14 milan: we have a stereotype that a local speech service is embedded, but it could be a plug in. so if the app requests a local service, that should be honored. 18:18:20 bjorn: what if it's not there? 18:18:52 milan: if it's a plugin you might ask the user if they want to install a plugin 18:19:34 bjorn: to make that work we would have to specify a plugin language 18:20:07 milan: agree, but if a local service is requested it should be used 18:20:38 michael: isn't this an example of the web app requesting an alternate speech service? 18:20:53 milan: don't expect to require downloading of code 18:21:06 RobertBrown has joined #htmlspeech 18:21:13 bjorn: how could this work without downloading code? 18:21:29 milan: we don't need to specify, just say that there has to be some mechanism 18:22:15 bjorn: what if we say that web app could point to a local service, but we shouldn't say that the ua has to try to install a local service 18:22:54 ??: what if some vendor only makes plugins for, i.e. IE 18:23:06 s/??/olli/ 18:23:06 that was olli 18:23:41 michael: we should have a requirement that speech services can specify a local service 18:23:56 new req: Speech services that can be specified by web apps must include local speech services. 18:24:59 dan: if we agree on this requirement, if someone wants to propose additional requirements we can discuss 18:26:16 michael: if the ua refuses resource it must inform web app (we already have this one) 18:26:31 so we have agreed to add this new requirement 18:26:38 ...don't distinguish between network and local 18:26:56 milan: what does "default speech resource"? 18:27:25 we are now discussing the related FPR9 and 10 to understand whether we need to add anything else 18:27:32 michael: in the future the author should be able to just ask for speech service without specifying one 18:28:09 bjorn: a browser must provide speech service, it doesn't have to build its own 18:28:24 I may not agree with the requirement 18:28:33 dan: it sounds like we do have agreement with this requirement 18:28:35 I need to think about it a bit 18:29:09 Who is smaug? 18:29:13 michael: can we remove R18 and discuss codecs later 18:29:14 smaug is Olli 18:29:26 bringert_ has joined #htmlspeech 18:29:44 dan: we could keep R18 just for that last point 18:30:12 dan: milan's concerns about R18 haven't been addressed 18:30:27 -Olli_Pettay 18:30:29 -Dan_Druta 18:30:31 -Milan_Young 18:30:32 -Bjorn_Bringert.a 18:30:33 -Marc_Schroeder 18:30:43 ..next call next week 18:31:15 rrsagent, who was present? 18:31:15 I'm logging. Sorry, nothing found for 'who was present' 18:31:21 -Robert_Brown 18:31:28 rrsagent, present? 18:31:28 I'm logging. Sorry, nothing found for 'present' 18:31:38 present? 18:32:04 zakim, list participants? 18:32:04 I don't understand your question, ddahl. 18:32:15 zakim, list participants 18:32:15 As of this point the attendees have been Debbie_Dahl, Olli_Pettay, +39.335.766.aaaa, Marc_Schroeder, Dan_Burnett, Michael_Bodell, Paolo_Baggia, Milan_Young, +1.760.705.aabb, 18:32:18 ... Bjorn_Bringert, +1.425.391.aacc, Robert_Brown, Dan_Druta, +44.122.546.aadd 18:32:45 rrsagent, make logs public 18:32:52 rrsagent, format minutes 18:32:52 I have made the request to generate http://www.w3.org/2010/11/11-htmlspeech-minutes.html ddahl 18:38:44 bringert has joined #htmlspeech 18:43:27 http://dev.w3.org/2002/scribe/scribedoc.htm 18:45:35 http://www.w3.org/2010/11/11-htmlspeech-irc.txt 18:49:20 -Michael_Bodell 18:49:22 -Debbie_Dahl 18:49:25 -Dan_Burnett 18:49:29 ddahl has left #htmlspeech 18:54:26 disconnecting the lone participant, Bjorn_Bringert, in INC_(HTMLSPEECH)12:00PM 18:54:29 INC_(HTMLSPEECH)12:00PM has ended 18:54:34 Attendees were Debbie_Dahl, Olli_Pettay, +39.335.766.aaaa, Marc_Schroeder, Dan_Burnett, Michael_Bodell, Paolo_Baggia, Milan_Young, +1.760.705.aabb, Bjorn_Bringert, +1.425.391.aacc, 18:54:37 ... Robert_Brown, Dan_Druta, +44.122.546.aadd 20:39:57 Zakim has left #htmlspeech