16:03:57 RRSAgent has joined #apa 16:03:58 logging to https://www.w3.org/2021/10/26-apa-irc 16:03:59 RRSAgent, make logs Public 16:04:01 please title this meeting ("meeting: ..."), jamesn 16:04:20 present+ 16:04:22 present+ 16:04:24 present+ 16:04:29 Scribe: Joshue108 16:04:31 NeilS has joined #apa 16:04:33 present+ 16:04:35 present+ 16:04:36 aaronlev has joined #apa 16:04:38 jasonjgw has joined #apa 16:05:05 meeting: APA & ARIA: The Future of Accessibility APIs 16:05:23 present+ 16:05:24 SamKanta has joined #apa 16:05:34 present+ 16:05:42 present+ 16:06:53 16:08:07 hmm 16:08:22 hEn!2xGg!u 16:08:28 try that 16:08:35 or 16:08:36 4613145760 16:08:54 it gives two for some reason 16:09:13 bkardell_ has joined #apa 16:09:17 Matthew_Atkinson has joined #apa 16:09:48 s/hmm/ 16:09:56 s/hEn!2xGg!u/ 16:10:01 s/4613145760/ 16:10:12 cyns has joined #apa 16:15:56 present+ 16:16:14 SteveNoble has joined #apa 16:16:27 present+ 16:16:57 TOPIC: Pronunciation Spec Discussion 16:17:06 mhakkinen has joined #apa 16:17:19 JS: Thanks for joining - lets look at this 16:17:32 Bridge differences from engineering perspectives. 16:17:40 Can Mark or Irfan kick this off? 16:17:53 So we can share perspectives etc? 16:18:01 q+ 16:18:04 Single vs multiple attributes.. 16:18:10 ack Paul 16:18:30 https://www.w3.org/TR/spoken-html/ 16:18:36 PG: The goal is create authoring capabilities in HTML 16:18:51 We have identified a gap in specs and APIs 16:19:00 This is augmentation of AX Tree 16:19:17 there are two candidates - one is a single attribute tbd 16:19:35 Also there is a multi attribute approach data-ssml currently 16:19:49 Currently tech based values 16:20:20 q+ to ask about tag review 16:20:25 Irf: We need to find a way to expose this 16:20:41 JS: The reason it is prefixed data dash is as this is defined in HTML 16:21:01 Once we have an implementation, then we go to the HTML group, and ask for a reserved prefix 16:21:06 We are a way off that. 16:21:26 But we need to get POC built etc. Make sure it works. 16:21:39 Then we can get reserved prefix etc 16:21:49 ack br 16:21:54 ack bk 16:21:54 bkardell_, you wanted to ask about tag review 16:22:07 BK: Is the spoken HTML idea reviewed by TAG? 16:22:12 Seems like a good idea. 16:22:32 JS: We did that last year - we heard from them, don't ask the parser to change 16:22:58 Our current approach is within the scope of current parsing capability 16:23:13 BK: I've seen two diff interpretations around the use of these attributes. 16:23:17 Where can we discuss? 16:23:26 JS: Happy to discuss now. 16:23:48 BK: has heard different interpretations of this 16:23:55 I think data attributes are fine 16:24:22 Some feel strongly that for things that are standard, that isn't appropriate 16:24:27 Can we open an issue? 16:24:29 JS: Yes 16:24:42 It may be on the HTML spec - we are following their guidelines. 16:24:56 To drive consistent TTS out put in various envs. 16:25:26 Matthew mentioned approach in personalisation , is using data- to drive it over there. 16:25:35 To drive personalization 16:25:45 JS: You can't get to a W3C REC using data- 16:25:51 CR is as far as it will go. 16:25:59 Thats the sandbox for data- 16:26:21 keep implementations to allay cross site concerns 16:26:34 JS: 16:26:59 16:27:03 JS: Does that help? 16:27:23 BK: Thats not new but just sharing counter interpretation 16:27:38 This seems like a good use case to begin discussion 16:27:58 JS: Mentions other specs using this approach 16:28:22 Do we have a preference, is the crux here? 16:28:25 q? 16:28:41 JS: Others ? 16:28:56 JS: Dave Tseng, how does that sound? 16:29:06 Is multiple attribute preferable? 16:29:50 Paul: Did mention an affinity for the multi attribute approach. There is no corollary for JSON as a value. 16:30:18 JS: That is one view from one AT , which is fine. 16:30:20 q? 16:30:40 The difference for AT is around approach 16:31:08 The group is more interested in JSON, as it is a single target, selector - info is picked up, the AX can abstract that, augment and provide info 16:31:26 +q 16:31:47 JS: The direct read group may be different - things that power our speech recognition devices. 16:32:02 q+ 16:32:14 GlenG: We do not have a problem parsing the HTML 16:32:32 Making fewer calls, from an AT angle, is good - esp if noisy. 16:32:38 q? 16:32:41 JSON, is good so we can get it all at once 16:32:58 ack mhakkinen 16:33:03 In a single attribute we can do that. 16:33:03 ack m 16:33:06 q+ 16:33:11 MK: Mentions read aloud tools 16:33:25 Text Help have a preference for single attribute 16:33:39 MS has immersive reader capabilities 16:33:44 ack cyns 16:33:56 How would they use pronunciation q.. 16:34:11 CS: Jumping out of the A11y APIs seems like a big step. 16:34:14 How did that happen? 16:34:31 s/pronunciation q../pronunciation cues/ 16:34:35 JS: We see use cases that provide benefit for AT that doesn't use the AX tree. 16:34:36 ack tink 16:34:40 s/Some feel strongly that for things that are standard,/Some have expressed to me subtler interpretations that data-* is for something more narrow, I'd like to see if I can get them to share discussion there 16:35:09 q+ 16:35:12 LW: From those who train and teach - while the JSON attribute is unfamiliar - adding more attributes could be confusing 16:35:38 We have spent years explaining around applied semantics, and the ground work of understanding the A11y tree 16:35:49 ack SteveNoble 16:35:51 And it could be confusing for adoption with a different approach 16:36:23 SN: Pearson has an implementation of the single attribute approach - just to channel what Paul G says.. 16:36:37 16:37:01 Sniffing and selecting tons of attributes in the DOM will be worse, that teasing out JSON via a processor 16:37:01 The GitHub comment being referenced is https://github.com/w3c/pronunciation/issues/86#issue-904400398 16:37:22 When a TTS player has to rip through the data the round trip is brutal 16:37:43 Jemma has joined #apa 16:37:53 present+ 16:38:05 q+ 16:38:15 SN: Performance will be impacted in the read aloud environment, it's a concern. 16:38:20 rrsagent, make minutes 16:38:20 I have made the request to generate https://www.w3.org/2021/10/26-apa-minutes.html Jemma 16:38:26 ack paulg 16:38:40 s/comment being referenced/SteveNoble mentioned/ 16:39:39 PG: With ARIA live the A11y tree gets updated - there is a concern about dynamic content - when there is a special region, the browser has to catch up 16:39:47 scribe: Matthew_Atkinson 16:39:50 present+ 16:39:55 Matthew_Atkinson: present+ 16:39:57 scribe: Matthew_Atkinson 16:40:05 q? 16:40:35 cyns: aaronlev/David Tseng: any thoughts? 16:41:17 q+ 16:41:19 aaronlev: Concerns around how these things will be misused by authors (c.f. live regions). What is the ideal markup that we would want? 16:41:28 q- 16:41:42 q+ to ask about css pronunciation 16:42:03 janina: The ideal would be to make SSML a native citizen of HTML. Concern around it not being possible to change validators (not necessarily parsers as mentioned above). 16:42:08 q+ 16:42:13 aaronlev: How about providing a separate SSML file? 16:42:37 janina/becky: Don't think that was considered. 16:42:59 aaronlev: Want to consider: what's the potential for misuse. Also: platform AX APIs can often be extended to provide more information. 16:43:36 PaulG: Made a note about homonym attacks (in the document). 16:43:56 q? 16:44:13 q+ 16:44:15 q+ to echo that authors WILL use things if they are available - just because they can 16:44:43 aaronlev: Concern around empowering authors to give the users a bad experience (c.f. ARIA can be misused in this way). Interested to hear Glen Gordon [who's on the line]'s thoughts on this. Examples such as inconsistencies inter-site or intra-site. 16:45:09 aaronlev: Using a single element could increase the risk that the AT can't present things consistently to users? 16:45:12 ack cyn 16:45:12 cyns, you wanted to ask about css pronunciation 16:45:29 q+ 16:45:57 cyns: Is there a relationship between this and the pronounciation functionality being proposed for CSS, or are they different use cases. Also concern around author misuse: remember when everyone made all the fonts small and light gray? Worried that a lot of things will get sped up. 16:46:17 We covered CSS Speech gap analysis here https://w3c.github.io/pronunciation/gap-analysis/ 16:46:41 janina: I think the CSS work is orthogonal to what we're trying to do; we did a gap analysis [URL above] that may have more info. 16:46:42 section #3 16:46:54 could we ++ tink in the queue before me? 16:47:11 sure will do that bkardell 16:47:37 janina: Possibility for misuse: anything that allows extra functionality could be mis-used. Need to be aware of it. But this allows us to do things such as mixing languages within a book (such as historical text). 16:48:34 Can someone drop a link to the use cases in here? 16:48:36 janina: This also helps confuse wind/wind and tear/tear. Much opportunity for improvement. It's not a proposal that the entire document needs to be marked up for pronounciation. In most cases TTS engines will do reasonably well. 16:49:00 https://w3c.github.io/pronunciation/use-cases/ 16:50:00 ack janina 16:50:27 q+ AT and voice assistants could "learn" from authors 16:50:30 q-- 16:50:31 jamesn: Worried about over-use and mis-use. Not sure how we counter this. Yes, it is necessary in certain cases. Not a screen reader user, so unsure if this is a problem: company names. Can see people wanting to put this into their company name. Is it a problem if a compan's site pronounces it correctly, but everywhere else on the web it's incorrect? 16:50:35 ack Jamesn 16:50:35 jamesn, you wanted to echo that authors WILL use things if they are available - just because they can 16:50:42 ack tink 16:51:11 q+ to comment "learning" from authors 16:51:27 tink: To answer your question jamesn, I would find it useful to hear the canonical company name pronounciation. Can get too used to how the AT pronounces it. 16:51:54 tink: CSS Speech is catering for a specific set of use cases: it's trying to make the auditory experience less tedious. 16:52:22 tink: Yes it can be misued: HTML, ARIA, XML are all misused, but think we can mitigate against this, but not stop it. 16:52:56 tink: For now the CSS Speech media type isn't supported by UAs, sadly. But, different use cases. 16:52:57 ack bkardell_ 16:54:10 bkardell_: The use cases are different, but problems are similar in that we need to affiliate nodes with values. Presumably wouldn't have one SSML document for the entire page, nor hundreds/thousands (would be very slow to load from network). 16:55:07 aaronlev: Was spitballing; though we can load a hundred images for a document. Can we put element IDs in SSML documents? The idea is mainly to avoid adding noise to the markup of the document. 16:55:30 bkardell_: Whilst the single JSON attribute could be ugly, can also see the benefit of keeping the info together. 16:55:57 q? 16:56:01 bkardell_: Brought this up in the MathML meeting as well, but we could polyfill something like this with existing technologies? Then authors wouldn't need to create the cumbersome JSON attributes. 16:56:26 ack PaulG 16:56:26 PaulG, you wanted to comment "learning" from authors 16:56:26 aaronlev: This feels a bit like [CSS] background images; it's changing the presentation as opposed to the semantics? 16:57:16 PaulG: For linking, we did talk briefly about linking external resources (for the next stage of the spec). If SSML came to the document as a first-class citizen like SVG we would look into that. 16:58:06 PaulG: Performance: TTS uses a lot of heuristics to determine pronounciation. Reducing the need for heuristics may mitigate some performance hit. 16:58:28 q+ 16:58:44 PaulG: Voice assistants might start to learn correct pronounciations e.g. for company names from their official sites. 16:59:07 ack mhakkinen 16:59:11 janina: e.g. Versailles is pronounced differently based on location. 16:59:14 interesting point PaulG 17:00:00 mhakkinen: We have a lot of need for pronounciation in education (ref Pearson's work discussed before). We have looked for a standard solution, e.g. PLS. Pronounciation Lexicon Specification [scribe note: PaulG mentioned this just above]. 17:00:28 q? 17:01:02 mhakkinen: We want screen readers, read-aloud tools, etc. to benefit. Another example: pharmaceutical producs. And another: television/film/movie program guides (character names, actor names, etc.) 17:01:26 janina: We wanted to discuss this near-term problem but didn't intend to take the whole time for this; will summarize. 17:01:46 q+ 17:02:09 janina: Hearing from aaronlev that browsers aren't expected to be a blocker as to which approach is taken. Need more feedback from AT vendors. Is that a reasonable summary? 17:02:18 q? 17:02:52 ack cyns 17:02:54 aaronlev: We _can_ implement anything; we still would need to look carefully at the proposal. We'd want good markup, good API support, good AT support; an end-to-end plan. Doesn't sound like all options have been looked at yet. 17:03:39 cyns: Have a similar view to aaronlev. The single-attribute approach feels counterintuitive for authors. It doesn't feel very HTML-like. Concerned about readability. 17:03:44 q+ 17:03:56 q+ 17:04:12 ack mhakkinen 17:04:17 aaronlev: JSON can be hard to read. 17:04:33 cyns: In general, it is OK but as an attribute value it is hard to read. 17:05:18 q+ one of the goals of markup is to be human readable 17:05:32 mhakkinen: From an authoring tool perspective, authors don't necessarily need to see the output HTML. We have tools already that allow authors to provide pronounciation hints that are intuitive to use. We need a standard way for ATs and others to consume it. 17:05:53 ack tink 17:06:03 tink: Is the idea with the single attribute that the JSON will be in the HTML code, or some external file that will be linked? 17:06:24 PaulG: Our current implementations/experimentations have the attribute value embedded in the HTML. 17:06:28 tink: How about an external file? 17:07:08 PaulG: We've had discussions about this before; have not yet found/developed method to do external linking. 17:07:32 jcraig has joined #apa 17:07:42 tink: Providing common rules is very much like CSS and could be of benefit here. 17:07:56 PaulG: Agree; would be great to have first-class SSML support. 17:08:18 cyns: Concerns around readability; if it's an external file this is less so. Could this just use CSS? 17:08:38 q+ to point out that external file would violate the AT privacy guidelines from web platform design principles that Leonie helped author 17:09:04 ack jcraig 17:09:04 jcraig, you wanted to point out that external file would violate the AT privacy guidelines from web platform design principles that Leonie helped author 17:09:11 bkardell_: There are efforts ongoing to allow authors to create CSS-like languages. (c.f. Houdini) 17:10:04 q+ to say that pronunciation could be used by other things besides AT 17:10:07 but it isn't really AT specific, it would apply to many speech agents 17:10:09 jcraig: The web platform design principles mention the importance of making AT _not_ detectable. Would be good to have SSML in the document, but requesting an external file would be detectable. 17:10:26 ack cyns 17:10:26 cyns, you wanted to say that pronunciation could be used by other things besides AT 17:10:46 cyns: I think use cases for this extend beyond AT, so not sure this would be useful for fingerprinting. Don't want to end up with what looks like inline CSS. 17:10:47 "hey read this" is a thing I use all the time - those would be indistinguishable 17:10:50 q+ 17:10:55 q+ 17:11:18 janina: Referencing external files could be helpful to avoid repetition. 17:11:20 ack aaronlev 17:12:02 aaronlev: Not sure if proposed, but: for the use case where changing the name of a product/address/company, sounds like we could use a dictionary. Problem: every time that name/phrase/word is announced you'd have to wrap its markup. 17:12:37 PaulG: We discussed this. Some tags like prossidy or voice can control an entire block. Others like pauses weren't there originally, so need an extra , with single or multi-attribute. 17:12:46 indistinguishable depends on many factors of entropy... client + accessed this other file + other factors might equal reasonable certainty of AT... FWIW, I think pronunciation rules are necessary. Just trying to point out the complications wrt that particular design principle 17:13:09 PaulG: The single attribute would, at first, encourage authors to summarize an entire block of text all at once, thus making it hard to update the pronounciation if the text changes. 17:13:19 PaulG: Would thus need help for developers to keep those in sync 17:13:42 PaulG: If everything is chopped up (multi-attribute), as a developer I think this would be easier, espeically for hand-coding devs. Interested as to others' views. 17:13:44 ack jamesn 17:14:25 jamesn: Replying to jcraig around detection. We _could_ require browsers to always fetch these files (is an additional complication, but could be managed). 17:15:01 jcraig: Absolutely agree that pronounication rules need to be defined in some format; just wanted to raise the issue. Has AX API design implications. 17:15:25 embeddacble as a css-like would work too and no extra fetch 17:15:31 q+ 17:15:39 janina: Seems everyone's agreed on the _need_ but we are still unsure as to single/multiple attributes, and there is the second-order question of external file. 17:16:04 q+ to ask if l10n/i18n was discussed in this context earlier 17:16:14 q+ 17:16:28 Joanmarie: If ATs want a single attribute, but authors want multiple attributes, or vice-versa, the implementation could be to take all the single attributes and parse them all together. 17:16:44 Joanmarie: We should consider what's best for authors, as a result. 17:16:46 ack aaronlev 17:16:48 q+ to mention l10n both with languages as well as with TTS capabilities 17:17:04 qv? 17:17:34 aaronlev: I feel there are many proposals that haven't been made yet, so should continue offline. But for the dictionary resource proposal: this could be something the AT fetches itself (circumventing the privacy issues; allowing caching across sites/domains). 17:17:55 aaronlev: Seems odd to me that we're going to be saying how to pronounce things, but only in one place. 17:18:09 "pronunciation" is only for phonemes. There are many more aural expressions from SSML that this spec would allow for. 17:18:16 aaronlev: ...where it'd be more useful if that was everywhere. 17:18:26 ack jcraig 17:18:26 jcraig, you wanted to ask if l10n/i18n was discussed in this context earlier and to mention l10n both with languages as well as with TTS capabilities 17:18:35 Q? 17:18:43 jcraig: aaronlev: are you implying there's a need for a global registry? 17:18:54 aaronlev: Not sure, but worth looking into. Consistency is important. 17:19:52 q+ 17:20:00 jcraig: Has l10n and i18n been discussed? E.g. Homonyms, in different languages/locales. Also different TTS voices may be able to pronounce Spanish and English, but not Chinese. Has any of the rules discussion covered this? 17:20:03 ack janina 17:20:33 janina: We have discussed those naunaces and the need to disambiguate them. The problem is that defining what the correct pronounciation is will change (e.g. wind/wind). 17:20:41 s/Homonyms, in different languages/locales. /Homonyms may be pronounced differently in languages/locales. / 17:21:27 [ scribe note: jcraig wished to raise having been delayed in joining, so missed some prior discussion ] 17:21:38 close/close is a more common homonym in UI in English 17:21:46 janina: Another example is English, but at different times in history, as proncounciations evolve. 17:22:10 ack PaulG 17:23:01 q+ 17:23:17 PaulG: A dictionary would be limited to phonemes. We have an example that's wider than this [Vincent Price reading The Raven]; covering audio "performance". 17:24:07 PaulG: Devs are guided towards specifying the language of the document, and the TTS does the rest. But there is contextual info (such as location) that might impact accent, vernacular, local place names, and that's part of what we're aiming to provide. 17:24:39 q+ 17:25:14 PaulG: Voice packs being able to support different proncounciations is another issue that we would need to resolve as an industry, but isn't something we can solve in the spec. Some pre-reading, or meta tags could be added to encourage assistants/AT to load specific voice packs/TTS capabilities to ensure a good experience for the user. 17:25:14 ack janina 17:25:39 janina: Maybe the voice packs issue is a metadata issue. 17:26:28 janina: Want to revisit Joanmarie's suggestion, as that could give us a path forward. If authoring is easier in multi-attributes, as long as the UAs can expose what the ATs need, that could address this. We should explore this. 17:27:15 janina: My concern is if we were to have conflicting views accross UAs. Joanmarie's suggested approach could help us address the UA-AT aspect. Does that sound good? 17:27:37 +1 to glen 17:27:50 +1 to glen 17:27:55 Glen: I don't think authors will unanimously agree on whether single-, or multi-attribute approach is easier. 17:28:11 q+ 17:28:17 jcraig: +1; depends on tooling 17:28:24 janina: I think we have to presume tooling. 17:28:39 cyns: Still thinking readability is important. 17:29:36 ack bkardell_ 17:30:29 bkardell_: There should be some experimentation, particularly with a CSS-like solution. There was discussion of polyfills in last year's TPAC? How are they getting the SSML to the AT? 17:30:39 q+ 17:30:45 ack SteveNoble 17:31:20 SteveNoble: Authoring: as mhakkinen said, the people authoring this stuff every day are using authoring tools. 17:32:28 +1 to general philosophical view that readability is important, though I am not an implementation expert in this field! 17:32:51 SteveNoble: [demonstrates some content that has been marked up for proncounciation] 17:33:59 ... The authoring tool identifies this as "alternate text for text-to-speech" that allows users to highlight a word, e.g. melancholy, and provide alternate text, e.g. melancollie that the system turns into SSML. 17:34:23 q+ to demo something similar from the iOS VoiceOver settings 17:34:30 q+ to say that a wysiwyg editors don't address my concerns about human readability of markup. You shouldn't have to use a special tool to write or read markup 17:34:41 ... There are also tables of words and "how to spell these phonetically in the system" 17:34:42 AndroUser has joined #apa 17:34:56 q+ to say that to me this looks like the kind of overuse I fear 17:35:06 ... e.g. dinosaur may be expressed as dinosore 17:35:43 ... Authors may be creating 1,000 SSML fragments per week (though they don't know it as SSML) to correct the way that the TTS pronounces things. 17:36:08 ... [Compares this to creating MathML with a WYSIWYG editor] 17:36:26 ack mhakkinen 17:36:55 mhakkinen: To echo what SteveNoble said, for classrom materials, many states have specific guidelines on pronounciation and we've had to spend time tweaking text so that it'll be pronounced with the right sort of pronounciation or pausing. 17:37:08 mhakkinen: We've tried to do this without authors having to learn SSML. 17:37:33 mhakkinen: We've had to create hacks to support whatever sorts of AT/read-aloud we are using at delivery time. We have not necesarily been able to get this into screen readers. 17:37:56 mhakkinen: E.g. if we altered the way the screen reader pronounced things, this could really confuse Braille users. 17:38:08 mhakkinen: We don't think this is a challenge for authors with the correct tooling. 17:38:26 ack jcraig 17:38:26 jcraig, you wanted to demo something similar from the iOS VoiceOver settings 17:39:04 jcraig: [Demonstrates VoiceOver Settings > Speech > Pronounciation for some of our names] 17:39:33 q+ 17:40:03 jcraig: You can speak how you want the term to be pronounced, and the device will interpret this and offer options (that it reads back) from which you can choose. 17:41:18 jcraig: Users can do this accross the system. Perhaps this could be exposed through WebKit. 17:41:35 ack cyns 17:41:35 cyns, you wanted to say that a wysiwyg editors don't address my concerns about human readability of markup. You shouldn't have to use a special tool to write or read markup 17:41:56 cyns: Special tools shouldn't be needed to make markup readable. 17:42:30 cyns: Have you looked at using a polyfil to pull all of the info into the AX API's description field? 17:42:36 thanks cyns that was the q I was asking too - your articulation was better 17:42:54 s/AX API/Accessibility API/ 17:42:55 mhakkinen: We've not done anything specifically for screen readers; our use cases are wider (e.g. read aloud tools). 17:43:03 mhakkinen: We prefer a standards-based approach. 17:43:32 The authoring standard also allows for scenarios like kiosks where an individual's AT/voice assistance solution may not be integrated with the content 17:43:44 SteveNoble: Our support internally is for our own TTS system and read and write extension. TextHelp is another vendor that supports SSML (single-attribute). 17:43:49 q? 17:44:24 ack me 17:44:24 jamesn, you wanted to say that to me this looks like the kind of overuse I fear 17:44:26 ack jamesn 17:44:29 s/Perhaps this could be exposed through WebKit./Perhaps this could be exposed through WebKit. I don't have a strong preference for whether that's via and attribute, or in a dictionary defined in a page script block or external resource. / 17:44:35 mhakkinen: Some years ago we prototyped a custom element that allows you to specify pronounciation and a Braiile label, but this didn't solve the problem of getting the screen reader to direct content specifically to TTS vs Braille. 17:45:36 ack jasonjgw 17:45:41 jamesn: Can see the publishing appraoch SteveNoble demo'd working when you have control over the TTS, but this standards approach is much more general. This doesn't seem like an appropriate use case for the wider web. 17:46:18 +q to agree with ETS comments that any polyfill implementable today may help speech users, but would be harmful to braille users. the standards approach takes longer but is the right path. 17:46:54 q? 17:47:19 jasonjgw: Trying to maximize author convenience and ACKing that this will differ across authors. The ability to define information globally and at the individual text element level seems to have got agreement. There's some flexibility on the UA side as to how it's represented in the markup and it seems possible to tailor the deliverey via the AX API for ATs that will maximize efficiency there. 17:48:15 jasonjgw: This has some parallels to the work NeilS demo'd at TPAC last week about how to provide disambiguating information on MathML. They're considering the same problems wrt how to specify the markup side and the deliverey side. We should aim to produce a simlar approach in both cases. 17:48:17 q+ 17:48:41 s/AX API/Accessibility API/g 17:48:47 ack me 17:48:47 jcraig, you wanted to agree with ETS comments that any polyfill implementable today may help speech users, but would be harmful to braille users. the standards approach takes 17:48:50 ... longer but is the right path. 17:48:51 ack jcraig 17:48:52 jasonjgw: That might help the discussion along. There were broader issues raised in the agenda, but these seem to have specific parallels across the work of different groups. 17:48:57 I agree it is very hard for me to not see the interrelationship here -- they might not be the exact same thing, but they certainly seem to have some intersection of concerns 17:49:25 jcraig: Wanted to +1 the ETS comments: bending existing pronounciation rules in specific contexts would be harmful to Braille in the general context. 17:50:05 jcraig: The standards approach is the right approach for our use-cases; don't see a polyfill approach working. 17:50:08 ack NeilS 17:50:37 q+ to respond 17:50:47 NeilS: Our (MathML)'s question was: if we're polyfilling, what's the target? Can't use aria-label as would negatively affect Braille. There is no target in the AX API. Seems we have to add something. 17:50:54 ack jcraig 17:50:54 jcraig, you wanted to respond 17:51:56 jcraig: +1; MathJax polyfil is a good example as it degrades the user experience when the platform has wider features (such as conversion to Nemeth Braille, which is bypassed by the Polyfill). 17:52:23 becky: Does anyone want to provide a summary, or next steps? 17:53:07 s/AX API/Accessibility API/g 17:53:30 bkardell_: The single-attribute version is not pretty, but if we could figure out how to plumb that down so that it could be used for polyfills/idea experimentation, we could always sugar on top of it (e.g. like with CSS, you can use inline style attributes, but normal humans authoring HTML wouldn't—we could have a similar abstraction). 17:53:44 present+ 17:54:23 q+ to ask if we can have the broader api discussion in the second session 17:54:32 q+ 17:54:52 ack cyns 17:54:52 cyns, you wanted to ask if we can have the broader api discussion in the second session 17:55:06 cyns: Could we have the broader AAPI discussion in the second session? 17:56:02 ack jcraig 17:56:09 mhakkinen: Helpful discussion; lots for the TF to consider. 17:57:23 jcraig: Is there a subset of spoken presentation in HTML draft that you'd recommend (some things such as prossidy have been mentioned as out of scope). In order to get this on an implementation schedule, suggest cutting it down and agreeing on any non-contraversial aspects, as could then get implementations behind runtime flags. 17:58:38 janina: One thing holding us back is to decide on the either/or with respect to attributes. 17:59:21 jcraig: Could include a dictionary inside a