Well-deployed technologies

An increasing share of mobile devices relies on touch-based interactions. While the traditional interactions recognized in the Web platform (keyboard, mouse input) can still be applied in this context, a more specific handling of touch-based input is a critical aspect of creating well-adapted user experiences, which Touch Events in the DOM (Document Object Model) enable.

The Vibration API lets mobile developers take advantage of haptic feedback to create new forms of interactions (e.g. in games).

Mobile devices follow their users everywhere, and many mobile users rely on them to remind them or notify them of events, such as messages: the Web Notifications specification enables that feature in the Web environment.

The User Agent Accessibility Guidelines (UAAG) 2.0 note defines principles and guidelines for user agents to design an accessible user agent interface and communicate with assistive technologies. The supporting document UAAG 2.0 Reference explains the intent and best practices of UAAG 2.0 success criteria, and lists numerous examples for each of them. Examples that are directly targeted at mobile devices are summarized in the Mobile Accessibility Examples from UAAG 2.0 Reference.

Following the Web Content Accessibility Guidelines (WCAG) 2.1 will make content accessible to a wider range of people with disabilities. The 2.1 revision adds new success criteria and guidelines to version 2.0, including new criteria that have a specific resonance in mobile contexts, such as the Pointer Gestures, Target Size and Orientation criteria.

Web content developers may benefit from authoring tools that follow the Authoring Tool Accessibility Guidelines (ATAG) 2.0 standard, which provides guidelines for designing Web content authoring tools that are both more accessible to authors with disabilities and that help design content that conforms to WCAG.

The Mobile Accessibility note explains how WCAG and other accessibility guidelines can be applied to mobile Web applications, as well as to native applications and hybrid applciations using Web components inside native applications.

The Accessible Rich Internet Applications (WAI-ARIA) 1.1 standard provides an ontology of roles, states, and properties that define the semantics of user interface elements and that can be used to improve the accessibility and interoperability of Web content and applications. The Core Accessibility API Mappings 1.1 standard describes how user agents should expose these semantics to accessibility APIs — unfortunately, mobile platforms do not yet have fully comprehensive accessibility API mappings and these mappings are only meaningful on desktop platforms for now.

FeatureSpecification / GroupMaturityCurrent implementations
Select browsers…
Touch-based interactionsTouch Events
Web Events Working Group
REC
VibrationVibration API (Second Edition)
Device and Sensors Working Group
REC
NotificationWeb Notifications
Web Notification Working Group
REC
AccessibilityUser Agent Accessibility Guidelines (UAAG) 2.0
User Agent Accessibility Guidelines Working Group
NOTE

N/A

UAAG 2.0 Reference: Explanations, Examples, and Resources for User Agent Accessibility Guidelines 2.0
User Agent Accessibility Guidelines Working Group
NOTE

N/A

Web Content Accessibility Guidelines (WCAG) 2.1
Accessibility Guidelines Working Group
PR

N/A

Authoring Tool Accessibility Guidelines (ATAG) 2.0
Authoring Tool Accessibility Guidelines Working Group
REC

N/A

Mobile Accessibility: How WCAG 2.0 and Other W3C/WAI Guidelines Apply to Mobile
Accessibility Guidelines Working Group
User Agent Accessibility Guidelines Working Group
WD

N/A

Accessible Rich Internet Applications (WAI-ARIA) 1.1
Accessible Rich Internet Applications Working Group
REC
Core Accessibility API Mappings 1.1
Accessible Rich Internet Applications Working Group
REC

N/A

Technologies in progress

The Pointer Events Working Group has made good progress on an alternative approach to handle user input, Pointer Events, that allows to handle mouse, touch and pen events under a single model. It provides a complementary and more unified approach to the currently more widely deployed Touch Events.

In particular, the CSS property touch-action that lets filter gesture events on elements is gaining traction beyond implementations of Pointer Events.

The early proposal for an Input Device capabilities API would provide information about a given “mouse” event comes from a touch-capable device.

As more and more content gets rendered as long scrollable lists, more and more logic is attached to scrolling events, and the quality of the user experience of these actions is highly dependent on their performances. The CSSOM View Module determines when scrolling events get fired, and let developers specify the type of scrolling behavior they want.

The proposed work on CSS Scroll Snap Points adds greater ability to control the behavior of panning and scrolling by defining points to which an app view would snap when the user moves through the page.

The CSS will-change property is also available to indicate to browsers that a given part of the page will be soon scrolled to and should be pre-rendered.

The Push API makes it possible for server-side notifications to alert the user, even if the browser is not running.

Whether users are speaking commands to their apps or working with them through non-haptic interactions, they risk seeing the screens turned off automatically by their devices screensaver. An early proposal for a Wake Lock API would let developers signal the needs to keep the screen up in these circumstances.

FeatureSpecification / GroupMaturityCurrent implementations
Select browsers…
Touch-based interactionsPointer Events
Pointer Events Working Group
WD
touch-action in Pointer Events
Pointer Events Working Group
WD
Input Device Capabilities
Web Platform Incubator Community Group
LS

N/A

Smooth scrollingCSSOM View Module
CSS Working Group
WD

N/A

CSS Scroll Snap Module Level 1
CSS Working Group
CR
CSS Will Change Module Level 1
CSS Working Group
CR
NotificationPush API
Web Platform Working Group
WD
Screen wakeWake Lock API
Device and Sensors Working Group
CR

N/A

Exploratory work

Mobile devices, and mobile phones in particular, are also in many cases well-suited to be used through voice-interactions; the Speech API Community Group developed a JavaScript API to enable interactions with a Web page through spoken commands. Speech synthesis is well supported across browsers. Support for speech recognition is still underway.

The Input Method Editor (IME) API provides Web applications with scripted access to an IME (input-method editor) associated with a hosting user agent. Editorial support is required for this specification to move forward.

FeatureSpecification / GroupImplementation intents
Select browsers…
Speech-based interactionsSpeech synthesis in Web Speech API
Speech API Community Group
Speech recognition in Web Speech API
Speech API Community Group
Input methodInput Method Editor API
Web Platform Working Group

Discontinued features

Intent-based events
As the Web reaches new devices, and as devices gain new user interactions mechanisms, it seems useful to allow Web developers to react to a more abstract set of user interactions: instead of having to work in terms of “click”, “key press”, or “touch event”, being able to react to an “undo” command, or a “next page” command independently of how the user instructed it to the device. The IndieUI Events specification was an attempt to address this need. The work has been discontinued for now, due to lack of support from would-be implementers.