17:07:50 RRSAgent has joined #media 17:07:50 logging to http://www.w3.org/2009/11/01-media-irc 17:07:55 http://lists.whatwg.org/pipermail/whatwg-whatwg.org/2009-July/021125.html 17:07:56 chaals has joined #media 17:08:06 rrsagent, make log public 17:08:18 Chair: Dave and John 17:08:36 ScribeNick: chaals 17:08:51 Meeting: Accessibility of Media Elements in HTML 5 Gathering 17:09:02 Scribe: Chaals 17:11:03 dsinger has joined #media 17:11:22 rrsagent, make logs public 17:11:42 dsinger has changed the topic to: Making Excessible Displays Imminently Available 17:17:01 dsinger has changed the topic to: Media Experts Define Innovative Accessibility 17:22:46 chaals has changed the topic to: Moderately Excessive Debating Idiocy: Accessibility 17:23:13 fsasaki has joined #media 17:34:20 agenda+ Introductions 17:34:25 agenda? 17:34:34 Zakim has joined #media 17:34:38 agenda+ Introductions 17:34:49 agenda+ set agenda... 17:35:17 agenda+ Changes for HTML5? 17:35:21 agenda+ Changes for CSS? 17:35:48 dsinger has joined #media 17:35:53 joakim has joined #media 17:35:56 agenda+ What's next - Review of decisions and actions 17:36:04 dsinger has left #media 17:36:09 agenda+ Any Other Business 17:36:46 dsinger has joined #media 17:40:54 soap has joined #media 17:42:53 SCain has joined #media 17:45:12 wondering where a few people are: silvia, judy, and a few others 17:45:58 MichaelC has joined #media 17:46:22 ChrisL2 has joined #media 17:47:38 chaals has changed the topic to: Accessible video media discussion thing 17:48:42 eric_carlson has joined #media 17:49:15 mattmay has joined #media 17:49:46 zakim, agendum 1 17:49:46 I don't understand 'agendum 1', chaals 17:49:54 zakim, next agendum 17:49:54 agendum 1. "Introductions" taken up [from chaals] 17:50:04 Topic: Introductions 17:50:36 JF: I am John Foliot. I want to show the Captioning service from Stanford at some point. 17:50:47 ... work on accessibiltiy services at Stanford 17:51:10 JS: Janina Sajka, chair WAI Protocols and Formats. Want to present some thughts about need for controls API 17:51:39 SC: Sally Cain, RNIB. Member of W3C PF group. Echoing what Janina said, want to talk about audio description 17:52:26 AT: oops from IBM lab Tokyo. We have been working on this stuff for a decade. Am a general chair for W4A. Want to indtroduce our audio description stuff. 17:52:55 s/oops/Hiro/ 17:53:17 MM: Matt May, Adobe. Want to talk about what we learned from accessibility in Flash, and the authoring tools context 17:53:59 MC: Michael Cooper, work as WAI staff. We're interested in W3C technology having accessibility baked in. I am intersted in using existing stuff rather than new things where effective. 17:55:02 Frank: Microsoft, here to listen 17:55:45 KH: Ken Herrington, google. Work on captioning systems - would like to show a bit of what we have done, and have tried to avoid standards work for ages (but failed now ;) ) 17:56:14 Joakim: researched at Ericsson, everyday job is on indexing (photo tagging, media services, ...) 17:56:25 ... co-chair of media annotation group at W3C 17:56:49 ... want to talk about what we have done at Ericsson. 17:57:33 FS: Felix Sasaki, from University of Applied Sciences in Potsdam. Teaching metadata and in media annotation workshop. Not presenting in particular 17:57:57 Marisa: DAISY Consortium developer (we make Digital Talking Book standards) 17:58:04 ... here to learn 17:58:35 ... can present a bit about DAISY and possibilities with HTML5 17:58:57 DB: Dick Bolterman, co-chair of SYMM (group at W3C doing SMIL) 17:59:27 ... I have been working on this for 14 decades :( Researcher at CWI in Amsterdam, interested in authoring systems for multimedia presentations. 17:59:59 ... would like to talk about what SMIL has done, maybe demo a captioning system we have for YouTube and some seperate caption streams allowing 3rd-party personalisation 18:00:10 IH: Ian Hickson, Google, HTML5 editor. 18:00:57 CL: Chris Lilley, W3C multimedia and SVG group, etc. Want to make sure that whatever we can do will be usable in SVG as well as HTML. Interested in i18n question - make sure you can have different languages. 18:01:51 PLH: Philippe le Hegaret, W3C. Responsible for HTML, and video, within W3C. Late-arriving participant in timed text working group. Hoping to get that work finished, have a demo of timed text with HTML5. 18:01:55 s/multimedia and/Hypertext CG cochair, CSS and/ 18:02:04 ... Didn't hear anyone wanting to present the current state in HTML5 18:03:01 JF: Also representing Pierre-Antoince Champain from Lyris (France) who ar working on this. 18:03:34 EC: Eric Carlson, Apple. Mostly responsible for engineering HTML5 media elements in Webki 18:04:47 CMN: Chaals, Opera. In charge of standards, hope not to present anything but interested in i18n and use of accessibility methods across different technologies withut reinventing wheels 18:05:19 DS: Dave Singer, Apple, head of multimedia standards. Interested in building up a framework that gets better accessibility over time. 18:06:29 SP: Sylvia Pfeiffer, work part time for Mozilla, trying to figure out how to get accessibility into multimedia (including looking at karaoke and various other time-based systems). Have some demo stuff to show, but want to review the requirements we have... 18:07:02 fo has joined #media 18:07:50 GF: Geoff Freed, NCAM, want to talk about captions and audio description in HTML5 18:09:14 JB: Judy Brewer, head of WAI at W3C. Interested in how the options for accessible media affect the user, especially when there are multiple options. 18:10:12 DS: Please do not speak over each other, speak clearly and slowly so interpreters and scribe can follow. 18:11:42 Topic: agenda bashing 18:11:54 DB: SMIL current experience 18:12:43 [we get: DAISY, Geoff, Google, Timed Text, Stanford Captioning, Silvia, Matt] 18:13:07 DS: go like this: 18:13:18 1. Geoff - about multimedia 18:13:24 2. John, Standford 18:13:34 3. Ken, Google stuff 18:13:47 4. Marisa, Daisy 18:14:00 James Craig rolls in from Apple 18:14:24 5. Silvia, stuff she has done 18:14:30 6. Matt (Flash) 18:14:43 7. Dick - SMIL-based captioning 18:15:05 8. Philippe, Timed Text 18:15:28 2.5. Pierre's video 18:15:56 geoff: how do we do your presentation? 18:16:12 Topic: Geoff Freed 18:16:20 GF: Want to talk a little about my concerns. 18:16:39 ... we also did some Javascript real-time captioning 18:17:30 GF: We have been playing around at NCAM with some captioning stuff that might work with video as is in HTML5, using real-time captioning. 18:18:07 ... we have been testing by stripping captions from a broadcast stream and embedding them in a page using javascript. We have a way to change channels - e.g. to have live captioning for an event like this 18:18:26 ... or it could be used as a way to stream caption data over different channels. 18:18:43 ... not the ideal way, it involves some outside work 18:19:08 ... would like to see something like a caption element rather than having to inject stuff with JS because that is probably not the most efficient way. 18:19:21 ... We have a demo, I will send a screenshot 18:20:15 Judy has joined #media 18:21:22 Topic: John Foliot 18:22:36 s/Foliot/Foliot and Stanford workflow for Captioning/ 18:23:36 JF: My role is assisting content providers to get accessible content online. Video has been a problem for some time. 18:24:10 ... Feedback from people with Video was that they found it expensive and difficult to actually make it happen when they were just staff doing something else, not video experts. 18:24:27 ... We made a workflow that allows staff to get captioned video online more or less by magic. 18:24:38 ... We set up a user account, and they can upload a media file. 18:25:02 silvia has joined #media 18:25:04 http://captiontool.stanford.edu/public/UploadFile.aspx 18:25:42 ... We have contracted with professional transcription companies (auto-transcription was not accurate enough for us) to do transcripts. 18:26:14 ... $5/minute to get 24-hour turn-around, $1.50 / minute for 5-day turnaround - $90 per hour of video. 18:26:57 ... System allows us to use multiple transcription services. We have created some custom dictionaries (e.g. people's name, technical terms we use, and so on) 18:27:27 ... Content owners can also add special terms if they use something odd, to improve accuracy. 18:28:00 ... If you already have a transcript you can upload that instead. (We then do the remaining work for free) 18:28:51 JF: Upload file, and we generate multiple formats - FLV, MP4, MP3 (which is sent to transcript company). 18:29:26 ... email is sent to transcription company when we have the file, and one to content producer so they can start putting their content online even if they haven't yet received the captions. 18:29:57 ... When transcription is done it is returned by the transcription company into the web interface. 18:30:23 ... Then we do some automatic timestamp generation to turn transcript into various formats. 18:30:58 ... User gets an email saying they have their captions, and we give them some code to copy that incorporates the captions into the web. 18:32:01 ... This is not quite shrink-wrapped and painless, but it is pretty close. You still have to shift some files yourself from server to server, but we are working on automating these steps too. 18:32:31 ... We rolled out the system at the end of the summer, have some users now and are talking to campus content creators to roll it out widely. 18:33:21 JF: Scalability of production is important. Stanford makes maybe 200 hours of video / week, and a bunch of that has archival value. So we need to be able to run it simply, and scalably. 18:33:45 SP: You mentioned a company that does timesamping 18:34:01 JF: Docsoft http://docsoft.com 18:34:05 SP: Open source? 18:34:52 JF: Nope. Shrinkwrap product based on Dragon. Speech reco not close enough for datamining, but good enough to automate timestamping completely. 18:37:06 JF: We focus on the datamining aspect of this as much as the accessibility 18:37:17 JC: Have you experimented with the UI in the video player for this? 18:37:39 JF: Not at this point. Also talking to reelsurfer about this. We are looking at it. 18:38:05 SP: Linking directly - should mention that there is a W3C Media Fragments group, looking at standards to directly address into a time with a URI 18:39:38 JF: We see people putting stuff onto YouTube - and that will use our caption materials if they are there. We are giving people the ability to do this stuff... 18:39:56 ACTION: John to provide a pointer into this 18:40:20 Please let us have your sample code for embedding, so we can see what people are using today 18:40:48 s/pointer into this/link that shows us some of how the stanford system works and looks for users/ 18:41:30 Topic: Pierre Antoine 18:41:37 s/Lyris/Liris/ 18:45:47 Video is presented (slideset and commentary) 18:47:48 Wonder what "speaker diarization" is ? 18:50:49 SP: Tracking who is speaking when 18:52:30 Topic: Ken Harrenstien, Google 18:53:08 http://www.youtube.com/watch?v=QRS8MkLhQmM 18:54:22 (Note that the video *relies* on the captions) 18:56:03 KH: Video is spoken and signed in different langauges. Trying t oclarify it isn't for a handful of poor deaf people, but for everyone. 18:56:15 (Luckily everyone here gets that) 18:56:25 PLH: Are the captions burned into the file or seperate 18:56:32 KH: Seperate. I will explain that... 18:56:54 ... A bit about numbers. Google has a very large number of videos, and everything has to work at scale. 18:57:28 ... Numbers are importa, going back to what John was saying, we show there is value for a lot of poeple. Now we have numbers like how many people are using captions on You Tube. 18:57:40 ... those numbers can drive the adoption of captioning. 18:58:07 ... we offer video searching of caption data. 18:58:34 (demo of actually searching, digging into a particular term's occurence in a video) 18:59:34 PLH: Is there anything in the usage numbers about how much people use video and if captioning has an influence on it. 18:59:50 KH: Don't have numbers to give, but they are good enough that I get more support in Google than I used to :) 19:00:05 ... want to let people see the numbers for their own video. 19:00:37 [Shows translation of captioning] 19:01:07 PLH: Speech to text, or caption as a source? 19:01:13 janina has joined #media 19:01:14 KH: Taking the captions as a source. 19:01:42 SP: Has Google included the offset search for YouTube into the Google engine. 19:01:52 KH: Yep. That was what I just showed. 19:03:00 SP: Captions can be used to improve search results. Is Google using them for indexing videos? 19:03:02 KH: Yes 19:03:23 Hiro has joined #media 19:03:30 DB: If I post a video, and someone else wants to caption later, how does that work? Can anybody caption anybody else's video? 19:03:38 ... can I restrict who uses the content? 19:03:55 KH: There are several websites that allow this. You can do it with or without permission. 19:04:06 DB: So I effectively make a new copy of the base video? 19:04:22 KH: Sort of depends on how you do it. For YouTube you ned to be the owner - or contact the owner. 19:04:38 JF: There are other 3rd prty tools that pull tools and transcript from seperate places. 19:05:17 ... The value of captions gets attention when people see the translation of captions 19:05:26 KH: Behind the scenes... 19:06:11 http://video.google.com/timedtext?v=aC_7NzXAJNI&lang=en 19:06:30 jcraig has joined #media 19:06:31 SP: You use a new XML format instead of SRT - for some particular reason? 19:06:55 KH: We control it. SRT isn't always so sweet.... 19:07:50 ... we actually produce various formats (incl. srt among others). It's easy to add new ones, so we don't care which format people have. 19:08:35 http://video.google.com/timedtext?v=aC_7NzXAJNI&lang=en&tlang=ru is a live-generated translation 19:11:14 Topic: Multi-channel caption streamer. 19:11:18 [screenshot] 19:11:33 Caption text is stripped from the broadcast and injected in real time using Javascript. 19:11:47 various UI controls on top. 19:11:47 --> http://lists.w3.org/Archives/Public/public-tt/2009Oct/0014.html NCAM real-time javascript captions 19:12:24 Transcript button can also give you a full log. 19:13:48 Break - 17 minutes. 19:14:31 present: Judy_Brewer, Dick_Bolterman, Sally_Cain, Eric_Carlson, James_Craig, Michael_Cooper, Marisa_DeMeglio, John_Foliot, Geoff_Freed, Ken_Harrenstien, Philippe_le_Hégaret, Ian_Hickson, Chris_Lilley, Charles_McCathieNevile, Matt_May, Frank_Olivier, Silvia_Pfeiffer, Janina_Sajka, Felix_Sasaki, David_Singer, Joakim_Söderberg, Hironobu_Takagi 19:15:21 I have made the request to generate http://www.w3.org/2009/11/01-media-minutes.html MichaelC 19:34:31 ScribeNick: ChrisL2 19:34:47 [Please do not use camera flash] 19:34:53 Marisa gives a presentation 19:35:06 rrsagent, here 19:35:06 See http://www.w3.org/2009/11/01-media-irc#T19-35-06 19:35:12 rrsagent, draft minutes 19:35:12 I have made the request to generate http://www.w3.org/2009/11/01-media-minutes.html ChrisL2 19:36:01 John introduces a guest account for the Stanford facility to experiment with the transcription service 19:40:02 Topic: Marisa DeMeglio, DAISY 19:40:25 [Marisi introduces DAISY with a demo] 19:40:29 http://captiontool.stanford.edu/ user: W3C_HTML5 pass: Accessible! 19:41:38 DAISY uses SMIL to synchronise the audio with the html text 19:42:19 MD: DAISY 4 being defined adds forms and video. Has an authoring and a distribution side. Can HTML5 be used for distribution 19:44:21 MD: Native browser support for audio and video is good, but we se barriers in html5, for example the extensibility. How to de add sidebars for example. Needs more roles than are built in. 19:44:38 MD: Want to avoid downgrading the content to represent it 19:45:42 JS: DAISY did a demo with two filmings of Henry V, scene by scene sync for comparative cinematography plys avatar for lip readers. Still interest in multiple synced media levels? 19:45:53 MD: Will be able to answer that next week 19:46:49 JS: Not clear if it fits in HTML 5 or 5.1 or 6 19:47:11 MD: Interested to use SMIL and HTML5 together to get that synchronisation 19:47:24 ... also forms and annotations 19:47:34 ... constrained by implementations 19:48:33 present+ Victor_Tsaran 19:48:51 Topic: Silvia_Pfeiffer, Demo and Proposals.... and Everything 19:50:32 SP: Demos with three Firefox builds - standard, with screenreader, and custom patched to add native accessibility support 19:51:23 SP: Need for signing captions for hard of hearing and deaf 19:51:30 ... audio descriptions 19:51:57 ... textual transcripts for screen reader or braille 19:53:43 SP: All of these create multiple content tracks. Multiple audio, text, and video tracks for a composite media eleent 19:53:56 ... need a standard javaScript interface to control this 19:54:08 ... and coneg reduces download to just the items of interest 19:54:56 SP: Text tracks are special and much more valuable outside the multimedia container. Easier to search index etc more easily than burried in media 19:55:18 ... aids editability, crowdsourching, CMS integration 19:55:42 ... sharable 19:56:24 Proposes an itext element with @lang, @type, @charset @src @category and @display 19:57:03 https://wiki.mozilla.org/Accessibility/HTML5_captions 19:57:48 sent to whatwg and html5 list, not much discussion there, but feedback directly on the wiki 19:59:46 [demo with video and selectable captions, subtitles] 20:00:23 SP: Issue is that all timed text tracks are treated identically. So next proposal identifies some categories, not all supported in the implementation 20:00:37 ... but these are the categories in use based on 10 years of collecting them 20:01:00 [demo with compiled firefox, rather than using scripting] 20:02:42 (prettier interface compared to the scripted one, same functionality, captions are transparent background now like proper subtitles 20:03:00 CL: Question on encodings vs display 20:03:12 PLH: Why is this an issue, Mozilla does this 20:03:58 SP: Uses HTML existing code for layout 20:04:06 PLH: default captioning language 20:05:02 SP: display="auto" selects an auto language negortiation. other options are none and force 20:05:36 ... user can override author choice, but author shoule be abel to express their design too 20:06:13 ... second proposal has a grouping element itextlist to express common categories etc rather than repeating them 20:06:40 https://wiki.mozilla.org/Accessibility/HTML5_captions_v2 20:07:10 dom interface allos to see if a given itext element is active 20:07:30 PLH: Are you generating events? 20:07:50 SP: Yes, onenter and onleave for new caption blocks or segments, so can listen for that 20:08:15 PLH: charset because srt ddoes not indicate the charset? 20:08:37 SP: yes. Some formats dont provide this 20:08:55 SP: charset is optional, some formats self describing and don't need it 20:09:04 ... no registered mime type for srt 20:10:05 [discussion on where SRT is defined] 20:10:06 http://srt-subtitles.com/ 20:10:49 SP: currentText api shows the currently displayed text, so a script can manipulate the text or display it 20:11:02 ... also a currenttime interface 20:11:35 ... works the same for external text tracks of for ones in the media container 20:11:59 JC: So can do searching or access text that will be displayed later 20:13:05 [demo of v2 using scripting implementation, changing language on the fly] 20:13:38 [demo of firefox with a screen reader, firevox] 20:16:00 [demo hgad audio discriptions for vision impaired, and text to speech audio description. uses aria.] 20:16:13 aria active region 20:17:01 SC: Will that use the defualt screenreader, or ontly the one in the browser 20:17:05 SP: the default 20:17:14 CN: How do you add audio ? 20:17:56 SP: needs native support in the browser with a n interface so its the same for internal and external sources ... audo and video needs to be synchronised 20:18:14 ... dynamic composition on server is recommended way to do that 20:18:48 CMN: Issue with therird party signing track, third party has no access to server where the video lives 20:19:10 SP: text is special, but audo could be treated like that too. To be discussed 20:19:14 ... needs a spec 20:19:35 CMN: Signed video is similarly special to subtitle text 20:20:40 JC: Can the audio pause the video for long descriptive text (for time to read it or have it spoken) 20:21:14 FO: If the text is inside the container and there is external text too what happens? 20:21:30 (we agree that the proiority needs to be defined) 20:21:43 FO: User needs to decide, no one size fits all solution 20:22:00 SP: yes we need the flexibility there and have the api to make it workable 20:22:10 JS: Resource discovery 20:22:36 CMN: So need to think about an API that finds internal as wel las external tracksa dn treat them uniformly 20:23:07 HT: Can itext eleents be added or changed dynamically? 20:23:09 SP: yes 20:23:40 S\[SP demonstrates a video hosting site with ogg audio, video, and subtitle./captioning/transcript support] 20:24:04 (all done in HTML5 and script) 20:24:48 http://oggify.com/ under development 20:25:20 SP: Like youtube except with open source and open formats and script plus HTML5 20:25:40 FS: Where does the timing info come from? 20:25:56 SP: From the subtitle file, they all have start and end times 20:26:40 JS: Nested structural navigation is important. chapters, sections etc 20:27:05 ... access to next scrne next act would be good 20:27:21 SP: Titled text tracks have DVD-style chapter markers 20:27:41 .. linear though, not hierarchical due to limitations of flat file 20:28:08 MD: Yes DAISY infers structure from heading levels 20:28:41 SP: Complext to bring in generic HTML and then display it o anothert HTL stream .. security issues 20:29:16 ... media fragments wg is specifying how to jump to named offsets as well as time offsets 20:29:23 ... not finished yet 20:29:50 JS; Direct access also for bookmarking as well as flipping through 20:30:47 SP: Chapter markers and structure exposes the structural content of the video, for speed reading among others. can do it with URIs so bookmarkable and can be in the history 20:31:36 Topic: Matt_May, Adobe, Flash experience 20:32:15 [Matt talks about history of accessible captioning in Flash] 20:32:54 MM: Two minimal requirements, flash support in video since Flash 6 and later the ability to insert cue points to associate captions with the video, in Flash 8 20:33:30 .. several attempts to crreate captioning, but they were unsynced so unsuccessful. Result was lack of adoption and a thousand hacks to try and do it 20:33:42 ... cue points got us closer, reliable feature 20:34:31 ... starting with flash 8, reliable caption sync but no standard way to do it. usually embedded int ehc ontsainer so hard coded fonts, done in actionscript. Content buried inside script 20:35:09 MM: Came to realisation that inside-only was a naive approach, looked for alternatives. In flash 9 we supported timed text dfxp 20:35:38 ... can assocate flv_playback_caption component, takes an external timed text file for captions 20:35:50 ... used an existing standard, tt was there 20:36:07 ... not re-re-re-re inventing the whee; 20:36:24 ... third parties can build captions and authoring software 20:36:48 ... hopeful that other formats adopt dfxp as well 20:37:56 MM: Breaking out the captions, as Sylvia and Ken mentioned, is important for any standard. Cam embed but thats just the first step. Only download the required captions. Allosw third parties to add their own content. Crowdsourcing captions 20:38:23 ... dealing with third parties adding captions later 20:39:16 MM: For html5, also important to have captions *inline* in the html document itself. Not complex to add 20:40:14 CL: SMIL also found a need to have text inline as an option 20:40:18 DB: One option 20:40:46 SP: most of my demos are at http://www.annodex.net/~silvia/itext/ 20:41:14 MM: Everyone is familiar with magpie 20:41:18 [we aren't] 20:42:04 MM: Shows this is well tried territory, so HTML5 should be able t use existing authoring tools and workflows to do this 20:42:15 MAGpie is a captioning tool from NCAM 20:42:23 MM: So please consider existing solutions 20:42:40 http://ncam.wgbh.org/webaccess/magpie/ 20:43:24 JC: Many content providers of video have no idea how captioning works, as other people do it 20:43:58 M: Using captions in AJAX, inserting vie athe DOM in real time. Issues of scrolling. 20:44:10 [lunch break] 20:50:41 Laura has joined #media 21:22:45 present+ Doug_Schepers 21:33:10 Hiro has joined #media 21:33:39 SCain has joined #media 21:35:55 t 21:36:57 Dick Bulterman (sp?) will talk about SMIL 21:37:15 ScribeNick: sylvia 21:37:34 Topic: Dick_Bolterman, CWI, SMIL Text 21:37:40 ScribeNick: silvia 21:37:41 s/Bo/Bu/ 21:38:11 chaals: I will scribe *for a bit* :-) 21:39:18 Scribe: Silvia 21:40:45 "Supporting Accessible Content" - lessons from SMIL 1/2/3/ 21:40:51 co-chair of SMIL working group 21:40:59 shepazu has joined #media 21:41:06 head of distributed & interactive systems group at CWI 21:41:16 interest for long time in working with multimedia 21:41:26 temporal & spatial synchronisation 21:41:39 take-home message: a11y isn't about a particular media format 21:41:48 it is about supporting selectivity among peer encodings 21:42:30 e.g. different encoding of same content for different situations, e.g. when driving/reading/conference 21:42:45 it is about a coordination mechanism to manage selection of emdia streams 21:42:59 s/emdia/media/ 21:43:18 difficulty is that they change over time 21:44:09 it is about providing 'situational accessibility' support 21:44:15 nobody wants special-purpose tools 21:44:18 --> http://www.w3.org/2009/Talks/1031-html5-video-plh/Overview.xhtml plh's slides 21:44:31 we should solve the problem for everybody 21:44:58 we need to make the temporal and spatial synchronisation explicit to be able to do the complex things 21:45:08 what is accessible content? 21:45:35 what kind of things would need to be done with a video object 21:45:46 it could be a svg object or another object, too 21:45:51 - want to add subtitles 21:46:03 - want to add captions and labels 21:46:28 labels are a cheap and simple way to communicate what is being visible in the video 21:46:42 "you are about to see my son play an instrument" 21:46:52 - line art & graphics 21:47:19 it would be nice to have a uniform model for all types of objects 21:47:23 - audio descriptions 21:47:49 - semantic discriminators 21:48:03 people want things at different levels of detail at different times 21:49:01 Some experiences from our SMIL work: 21:49:10 - not all encodings will be produce by the same party 21:49:39 e.g. even while Disney owns the video, they may not be the ones to create the captions 21:49:45 - not all content will be co-located 21:50:00 e.g. a video may be on one server, but content enrichments will be on many different servers 21:50:24 if you want highly synchronised audio and video, they practically have to be in the same file 21:50:58 the network delays can easily add up to make it impossible to synchronise them 21:51:06 but you can create some things on different servers 21:51:17 - there may be complex content dependencies 21:51:30 - each piece of content may not be aware of the complete presentation 21:51:45 *SMIL Support for Alternative Cotnent* 21:51:55 SMIL 1.0 in 1996 21:52:17 : selection based on system test attributes 21:52:30 (language, bitrate, captions) 21:52:42 support alternative selection of parallel tracks 21:53:13 demo of MIT Prof. Lewen (sp?) 21:53:41 in SMIL: 21:53:50