Generic Sensor API

W3C Working Draft,

This version:
https://www.w3.org/TR/2016/WD-generic-sensor-20160830/
Latest published version:
https://www.w3.org/TR/generic-sensor/
Editor's Draft:
https://w3c.github.io/sensors/
Previous Versions:
https://www.w3.org/TR/2016/WD-generic-sensor-20160324/
https://www.w3.org/TR/2015/WD-generic-sensor-20151015/
Version History:
https://github.com/w3c/sensors/commits/gh-pages/index.bs
Feedback:
public-device-apis@w3.org with subject line “[generic-sensor] … message topic …” (archives)
Issue Tracking:
GitHub
Level 1 Issues
Level 2 Issues
Editors:
(Intel Corporation)
Rick Waldron (jQuery Foundation)
Bug Reports:
via the w3c/sensors repository on GitHub
Test Suite:
web-platform-tests on GitHub

Abstract

This specification defines a framework for exposing sensor data to the Open Web Platform in a consistent way. It does so by defining a blueprint for writing specifications of concrete sensors along with an abstract Sensor interface that can be extended to accommodate different sensor types.

Status of this document

This section describes the status of this document at the time of its publication. Other documents may supersede this document. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at https://www.w3.org/TR/.

This document was published by the Device and Sensors Working Group as a Working Draft. This document is intended to become a W3C Recommendation. The Device and Sensors Working Group is seeking wide review on this document.

If you wish to make comments regarding this document, please send them to public-device-apis@w3.org (subscribe, archives). When sending e-mail, please put the text “generic-sensor” in the subject, preferably like this: “[generic-sensor] …summary of comment…”. All comments are welcome.

Publication as a Working Draft does not imply endorsement by the W3C Membership. This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.

This document was produced by a group operating under the 5 February 2004 W3C Patent Policy. W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.

This document is governed by the 1 September 2015 W3C Process Document.

1. Introduction

Increasingly, sensor data is used in application development to enable new use cases such as geolocation, counting steps or head-tracking. This is especially true on mobile devices where new sensors are added regularly.

Exposing sensor data to the Web has so far been both slow-paced and ad-hoc. Few sensors are already exposed to the Web. When they are, it is often in ways that limit their possible use cases (for example by exposing abstractions that are too high-level and which don’t perform well enough). APIs also vary greatly from one sensor to the next which increases the cognitive burden of Web application developers and slows development.

The goal of the Generic Sensor API is to promote consistency across sensor APIs, enable advanced use cases thanks to performant low-level APIs, and increase the pace at which new sensors can be exposed to the Web by simplifying the specification and implementation processes.

2. Scope

This section is non-normative.

The scope of this specification is currently limited to specifying primitives which enable expose data from local sensors.

Exposing remote sensors or sensors found on personal area networks is out of scope. As work in these areas mature, it is possible that common, lower-level primitives be found, in which case this specification will be updated accordingly. This should have little to no effects on implementations, however.

This specification also does not currently expose a sensor discovery API. This is because the limited number of sensors currently available to User Agents does not warrant such an API. Using feature detection, such as described in §4 A note on Feature Detection of Hardware Features, is good enough for now. A subsequent version of this specification might specify such an API, and the current API has been designed with this in mind.

3. Background

This section is non-normative.

The Generic Sensor API is designed to make the most common use cases straightforward while still enabling more complex use cases.

Most devices deployed today do not carry more than one sensor of each sensor types. This shouldn’t come as a surprise since use cases for more than a sensor of a given type are rare and generally limited to specific sensor types such as proximity sensors.

The API therefore makes it easy to interact with the device’s default (and often unique) sensor for each type simply by instantiating the corresponding Sensor subclass.

Indeed, without specific information identifying a particular sensor of a given type, the default sensor is chosen.

Listening to geolocation changes:
let sensor = new GeolocationSensor({ accuracy: "high" });

sensor.onchange = function(event) {
    var coords = [event.reading.latitude, event.reading.longitude];
    updateMap(null, coords, event.reading.accuracy);
};

sensor.onerror = function(error) {
    updateMap(error);
};

Note: extension to this specification may choose not to define a default sensor when doing so wouldn’t make sense. For example, it might be difficult to agree on an obvious default sensor for proximity sensors.

In cases where multiple sensors of the same type may coexist on the same device, specification extension will have to define ways to uniquely identify each one.

For example checking the pressure of the left rear tire:
let sensor = new DirectTirePressureSensor({ position: "rear", side: "left" });
sensor.onchange = event => console.log(event.reading.pressure);

4. A note on Feature Detection of Hardware Features

This section is non-normative.

Feature detection is an established Web development best practice. Resources on the topic are plentiful on and offline and the purpose of this section is not to discuss it further, but rather to put it in the context of detecting hardware-dependent features.

Consider the below feature detection examples:

if (typeof Gyroscope === "function") {
    // run in circles...
}
        
if ("ProximitySensor" in window) {
    // watch out!
}
        
if (window.AmbientLightSensor) {
    // go dark...
}
        
// etc.

All of these tell you something about the presence and possible characteristics of an API. They do not tell you anything, however, about whether that API is actually connected to a real hardware sensor, whether that sensor works, if its still connected, or even whether the user is going to allow you to access it. Note you can check the latter using the Permissions API [PERMISSIONS].

In an ideal world, information about the underlying status would be available upfront. The problem with this is twofold. First, getting this information out of the hardware is costly, in both performance and battery time, and would sit in the critical path. Secondly, the status of the underlying hardware can evolve over time. The user can revoke permission, the connection to the sensor be severed, the operating system may decide to limit sensor usage below a certain battery threshold, etc.

Therefore, an effective strategy is to combine feature detection, which checks whether an API for the sought-after sensor actually exists, and defensive programming which includes:

  1. checking for error thrown when instantiating a Sensor object,

  2. listening to errors emitted by it,

  3. handling all of the above graciously so that the user’s experience is enhanced by the possible usage of a sensor, not degraded by its absence.

try { // No need to feature detect thanks to try..catch block.
    let sensor = new GeolocationSensor({});
    sensor.start();
    sensor.onerror = error => gracefullyDegrade(error);
    sensor.onchange = event => updatePosition(event.reading.coords);
} catch(error) {
    gracefullyDegrade(error);
}

5. Security and privacy considerations

Privacy risks can arise when sensors are used with each other, in combination with other functionality, or when used over time, specifically with the risk of correlation of data and user identification through fingerprinting. Web application developers using these JavaScript APIs should consider how this information might be correlated with other information and the privacy risks that might be created. The potential risks of collection of such data over a longer period of time should also be considered.

Variations in sensor readings as well as event firing rates offer the possibility of fingerprinting to identify users. User agents may reduce the risk by limiting event rates available to web application developers.

Note: do we really want this mitigation strategy?

If the same JavaScript code using the API can be used simultaneously in different window contexts on the same device it may be possible for that code to correlate the user across those two contexts, creating unanticipated tracking mechanisms.

User agents should consider providing the user an indication of when the sensor is used and allowing the user to disable it.

Web application developers that use sensors should perform a privacy assessment of their application taking all aspects of their application into consideration.

Ability to detect a full working set of sensors on a device can form an identifier and could be used to fingerprinting.

A combination of selected sensors can potentially be used to form an out of band communication channel between devices.

Sensors can potentially be used in cross-device linking and tracking of a user.

5.1. Browsing Context

Sensor readings must only be available in the top-level browsing context to avoid the privacy risk of sharing the information defined in this specification (and specifications extending it) with contexts unfamiliar to the user. For example, a mobile device will only fire the event on the active tab, and not on the background tabs or within iframes.

5.2. Secure Context

Sensor readings are explicitly flagged by the Secure Contexts specification [POWERFUL-FEATURES] as a high-value target for network attackers. As such, sensor readings must only be available within a secure context.

5.3. Obtaining Explicit User Permission

Permission to access sensor readings must be obtained through the Permissions API [PERMISSIONS].

6. Concepts

A sensor measures different physical quantities and provide corresponding raw sensor readings which are a source of information about the user and their environment.

Known, predictable discrepancies between raw sensor readings and the corresponding physical quantities being measured are corrected through calibration.

Known but unpredictable discrepancies need to be addressed dynamically through a process called sensor fusion.

Calibrated raw sensor readings are referred to as sensor readings, whether or not they have undergone sensor fusion.

Different sensor types measure different physical quantities such as temperature, air pressure, heart-rate, or luminosity.

For the purpose of this specification we distinguish between high-level and low-level sensor types.

Sensor types which are characterized by their implementation are referred to as low-level sensors. For example a Gyroscope is a low-level sensor type.

Sensors named after their readings, regardless of the implementation, are said to be high-level sensors. For instance, geolocation sensors provide information about the user’s location, but the precise means by which this data is obtained is purposefully left opaque (it could come from a GPS chip, network cell triangulation, wifi networks, etc. or any combination of the above) and depends on various, implementation-specific heuristics. High-level sensors are generally the fruits of applying algorithms to low-level sensors—for example, a pedometer can be built using only the output of a gyroscope—or of sensor fusion.

That said, the distinction between high-level and low-level sensor types is somewhat arbitrary and the line between the two is often blurred. For instance, a barometer, which measures air pressure, would be considered low-level for most common purposes, even though it is the product of the sensor fusion of resistive piezo-electric pressure and temperature sensors. Exposing the sensors that compose it would serve no practical purpose; who cares about the temperature of a piezo-electric sensor? A pressure-altimeter would probably fall in the same category, while a nondescript altimeter—which could get its data from either a barometer or a GPS signal—would clearly be categorized as a high-level sensor type.

Because the distinction is somewhat blurry, extensions to this specification (see §10 Extensibility) are encouraged to provide domain-specific definitions of high-level and low-level sensors for the given sensor types they are targeting.

Sensor readings from different sensor types can be combined together through a process called sensor fusion. This process provides higher-level or more accurate data (often at the cost of increased latency). For example, the readings of a three-axis magnetometer needs to be combined with the readings of an accelerometer to provide a correct bearing.

Smart sensors and sensor hubs have built-in compute resources which allow them to carry out calibration and sensor fusion at the hardware level, freeing up CPU resources and lowering battery consumption in the process.

But sensor fusion can also be carried out in software. This is particularly useful when performance requirements can only be met by relying on application-specific data. For example, head tracking for virtual or augmented reality applications, requires extremely low latency to avoid causing motion sickness. That low-latency is best provided by using the raw output of a gyroscope, and waiting for quick rotational movements of the head to compensate for drift.

Note: sensors created through sensor fusion are sometimes called virtual or synthetic sensors. However, the specification doesn’t make any practical differences between them, preferring instead to differentiate sensors as to whether they describe the kind of readings produced--these are high-level sensors—or how the sensor is implemented (low-level sensors).

Sensors have different reporting modes. When sensor readings are reported at regular intervals, at an ajustable frequency measured in hertz (Hz), the reporting mode is said to be periodic. When it is only reported upon measurable change, the sensor is said to be in auto reporting mode.

Auto reporting mode can give the user agent more latitude to carry out power- or CPU-saving strategies and should generally be favored. Periodic reporting mode, on the other hand, allows a much more fine-grained approach and is essential for use cases with, for example, low latency requirements.

Note: reporting mode is distinct from, but related to, sensor readings acquisition. If sensors are polled at regular interval, as is generally the case, reporting mode can be either periodic or auto. However, when the underlying implementation itself only provides sensor readings when it measures change, perhaps because is is relying on smart sensors or a sensor hubs, the reporting mode cannot be periodic, as that would require data inference.

7. Model

7.1. Sensor Type

A sensor type has one or many associated sensors.

A sensor type has an associated Sensor subclass.

A sensor type has an associated SensorReading subclass. Attributes of the SensorReading subclass subclass that hold sensor readings must be readonly.

A sensor type may have a default sensor.

A sensor type has an associated set of supported reporting modes, which cannot be empty and must be picked from the following: auto and periodic.

If a sensor type has more than one sensor, it must have a set of associated identifying parameters to select the right sensor to associate to each new Sensor objects.

A sensor type has an associated abstract operation to construct a SensorReading object which takes the sensor readings emitted by the sensor and returns an initialized SensorReading object which uses the sensor type’s associated SensorReading subclass.

7.2. Sensor

A sensor has an associated set of activated Sensor objects. This set is initially empty.

A sensor has an associated current reading, which can initially be null or be set to a SensorReading object, that was cached in a user-agent-specific way.

Note: there are additional privacy concerns when using cached readings which predate either navigating to resources in the current origin, or being granted permission to access the sensor.

A sensor is said to support periodic reporting mode if its associated sensor type’s supported reporting modes contains the periodic reporting mode.

A sensor has an associated reporting flag which is initially unset.

A sensor has an associated current reporting mode which is initially undefined.

A sensor has an associated current polling frequency which is initially null.

A sensor has an associated abstract operation to retrieve its permission which takes a Sensor object as input and returns a permission and, eventually, its associated PermissionDescriptor.

8. API

8.1. The Sensor Interface

interface Sensor : EventTarget {
  readonly attribute SensorState state;
  readonly attribute SensorReading? reading;
  void start();
  void stop();
  attribute EventHandler onchange;
  attribute EventHandler onstatechange;
  attribute EventHandler onerror;
};

dictionary SensorOptions {
  double? frequency;
};

enum SensorState {
  "idle",
  "activating",
  "active",
  "errored"
};

A Sensor object has an associated sensor.

A Sensor object has an associated state, which is one of "idle", "activating", "active", and "errored". It is initially "idle".

A Sensor object has an associated desired frequency. It is initially null.

Each Sensor object has a task source called a sensor task source, initially empty. A sensor task source can be enabled or disabled, and is initially enabled. When enabled, the event loop must use it as one of its task sources. The task source for the tasks mentioned in this specification is the sensor task source.

When the visibility state of the Document in the top-level browsing context changes, let current_visibility_state be the result of running the steps to determine the visibility state of the Document. If current_visibility_state is "visible", enable the sensor task source, otherwise, disable it.

Note: user agents are encouraged to stop sensor polling when sensor task sources are disabled in order to save battery.

8.1.1. Sensor.state

The state attribute represents a Sensor object’s state.

8.1.2. Sensor.reading

When a Sensor's state is "active", its reading attribute must always point to the current reading whatever the frequency so that the reading attribute of two instances of the same Sensor interface associated with the same sensor hold the same SensorReading during a single event loop turn.

8.1.3. Sensor.start()

The start() method must run these steps or their equivalent:
  1. If sensor_instance’s state is neither "idle" nor "errored",

    1. throw an "InvalidStateError" exception and abort these steps.

  2. Invoke the update state algorithm passing sensor_instance and "activating" as the arguments.

  3. Run these sub-steps in parallel:

    1. Let state be the result of invoking the Request Sensor Access abstract operation, passing it sensor_instance as argument.

    2. If state is "granted",

      1. Invoke Register a Sensor passing it sensor_instance as argument.

      2. Abort these sub-steps.

    3. If state is "denied",

      1. let e be the result of creating an exception named NotAllowedError.

      2. Invoke the Handle Errors abstract operation, passing it e and sensor_instance as arguments.

      3. Abort these sub-steps.

  4. return undefined.

8.1.4. Sensor.stop()

The stop() method must run these steps or their equivalent:
  1. If sensor_instance’s state is either "idle" or "errored", then

    1. throw an "InvalidStateError" exception and abort these steps.

  2. Set sensor_instance’s reading to null.

  3. Invoke the update state algorithm passing sensor_instance and "idle" as the arguments.

  4. Run these sub-steps in parallel:

    1. Invoke Unregister a Sensor passing it sensor_instance as argument.

  5. return undefined.

8.1.5. Sensor.onerror

8.1.6. Sensor.onchange

8.1.7. Sensor.onstatechange

8.1.8. Event handlers

The following are the event handlers (and their corresponding event handler event types) that must be supported as attributes by the objects implementing the Sensor interface:

event handler event handler event type
onchange change
onstatechange statechange
onerror error

8.2. The SensorReading Interface

interface SensorReading {
  readonly attribute DOMHighResTimeStamp timeStamp;
};

A SensorReading represents the state of a sensor at a given point in time.

8.2.1. SensorReading.timeStamp

Returns a timestamp of the time at which the reading was obtained from the sensor expressed in milliseconds that passed since the time origin.

<https://github.com/w3c/sensors/issues/101>

8.3. The SensorReadingEvent Interface

[Constructor(DOMString type, SensorReadingEventInit eventInitDict)]
interface SensorReadingEvent : Event {
  readonly attribute SensorReading reading;
};

dictionary SensorReadingEventInit : EventInit {
  SensorReading reading;
};

8.3.1. SensorReadingEvent.reading

The reading attribute references the current reading at the time of event dispatch.

8.4. The SensorErrorEvent Interface

[Constructor(DOMString type, SensorErrorEventInit errorEventInitDict)]
interface SensorErrorEvent : Event {
  readonly attribute Error error;
};

dictionary SensorErrorEventInit : EventInit {
  Error error;
};

8.4.1. SensorErrorEvent.error

9. Abstract Operations

9.1. Construct Sensor Object

input

options, a SensorOptions object.

output

sensor_instance, a Sensor object.

  1. If the incumbent settings object is not a secure context, then:

    1. throw a SecurityError.

  2. If the browsing context is not a top-level browsing context, then:

    1. throw a SecurityError.

  3. Otherwise, if identifying parameters in options are set, then:

    1. If these identifying parameters allow a unique sensor to be identified, then:

      1. let sensor be that sensor,

      2. let sensor_instance be a new Sensor object,

      3. associate sensor_instance with sensor.

    2. Otherwise, throw a TypeError.

  4. Otherwise, if a default sensor exists for this sensor type:

    1. let sensor be that sensor,

    2. let sensor_instance be a new Sensor object,

    3. associate sensor_instance with sensor.

  5. Otherwise, throw a TypeError.

  6. Let frequency be the value of the frequency member of options.

  7. If frequency is not null,

    1. If sensor supports periodic reporting mode,

      1. Set sensor_instance’s desired frequency to frequency.

    2. Otherwise,

      1. throw a TypeError.

  8. Set sensor_instance’s reading attribute to null.

  9. Set sensor_instance’s state to "idle".

  10. return sensor_instance.

9.2. Register a Sensor

input

sensor_instance, a Sensor object.

output

None

  1. Let sensor be the sensor associated with sensor_instance.

  2. Add sensor_instance to sensor’s set of activated Sensor objects.

  3. Invoke the Set Sensor Settings abstract operation, passing it sensor as argument.

  4. Let current_reading be sensor’s associated current reading.

  5. If current_reading is not null and sensor_instance’s state is still "activating", then

    1. invoke the Update Reading operation, passing it sensor_instance and current_reading as arguments.

9.3. Unregister a Sensor

input

sensor_instance, a Sensor object.

output

None

  1. Let sensor be the sensor associated with sensor_instance.

  2. Remove sensor_instance from sensor’s set of activated Sensor objects.

  3. If sensor’s set of activated Sensor objects is empty,

    1. Set current reporting mode to undefined.

    2. Set current polling frequency to null.

    3. Update the user-agent-specific way in which sensor readings are obtained from sensor to no longer provide readings.

    4. Abort these steps.

  4. Invoke the Set Sensor Settings abstract operation, passing it sensor as argument.

9.4. Set Sensor Settings

input

sensor, a sensor.

output

None

  1. Let settings_changed be false.

  2. Let mode be the result of invoking the find current reporting mode of a sensor abstract operation, with sensor as argument.

  3. If mode is different from sensor’s current reporting mode,

    1. set settings_changed to true.

    2. Set current reporting mode to mode.

  4. Let frequency be the result of invoking the Find the polling frequency of a Sensor abstract operation, with sensor as argument.

  5. If frequency is different from sensor’s current polling frequency,

    1. set settings_changed to true.

    2. Set current polling frequency to frequency.

  6. If settings_changed is true

    1. Invoke the Observe a Sensor abstract operation, passing it sensor as argument.

9.5. Observe a Sensor

input

sensor, a sensor.

output

None

  1. If sensor’s current reporting mode is periodic,

    1. let frequency be the current polling frequency, capped by the upper and lower bounds of the underlying hardware.

    2. Invoke the Update Current Reading abstract operation at a frequency of frequency passing it sensor and the latest sensor reading as arguments.

      Note: there is not guarantee that the current polling frequency can be respected. The actual frequency can be calculated using SensorReading timeStamp attributes.

  2. If the current reporting mode is set to auto,

    1. the user-agent can decide on the best reporting strategy for this particular sensor and sensor type.

    2. The user-agent must invoke the Update Current Reading abstract operation at to report fresh readings, passing it sensor and the latest sensor reading as arguments.

9.6. Find Current Reporting Mode of a Sensor

input

sensor, a sensor.

output

mode, a reporting mode.

  1. let mode be auto.

  2. For each sensor_instance in sensor’s set of activated Sensor objects:

    1. if sensor_instance’s frequency is not null,

      1. set mode to periodic

  3. return mode.

9.7. Find the polling frequency of a Sensor

input

sensor, a sensor.

output

frequency, a frequency.

  1. let frequency be null.

  2. For each sensor_instance in sensor’s set of activated Sensor objects:

    1. let f be sensor_instance’s frequency.

    2. if f is not null and f is greater than frequency,

      1. set frequency to f.

  3. return frequency.

9.8. Update State

input

sensor_instance, a Sensor object.

state, a Sensor object’s state.

output

None

  1. Set sensor_instance’s state to state.

  2. Queue a task to fire a simple event named statechange at sensor_instance.

9.9. Update Current Reading

input

sensor, a sensor.

reading, a sensor reading.

output

None

  1. If sensor’s reporting flag is set,

    1. abort these steps.

  2. Set sensor’s reporting flag.

  3. Let reading_instance be the result of invoking sensor type’s associated Construct SensorReading Object operation, passing it reading as argument.

  4. Set sensor’s current reading to reading_instance.

  5. For each sensor_instance in sensor’s set of activated Sensor objects:

    1. Invoke the update reading algorithm passing sensor_instance and reading_instance as arguments.

  6. Unset sensor’s reporting flag.

9.10. Update Reading

input

sensor_instance, a Sensor object.

reading, a SensorReading object.

output

None

  1. Set sensor_instance’s reading to reading.

  2. If sensor_instance’s state is "activating":

    1. Invoke the update state algorithm passing sensor_instance and "active" as the arguments.

  3. Create an event e that uses the SensorReadingEvent interface, with the event type reading, which does not bubble, is not cancelable, is trusted, and has no default action.

  4. Let the reading attribute of e be initialized to reading.

  5. Queue a task to dispatch e at sensor_instance.

9.11. Handle Errors

input

sensor_instance, a Sensor object.

error, an Error object.

output

None

  1. Set sensor_instance’s reading to null.

  2. Invoke the update state algorithm passing sensor_instance and "errored" as the arguments.

  3. Create an event e that uses the SensorErrorEvent interface, with the event type error, which does not bubble, is not cancelable, is trusted, and has no default action.

  4. Let the error attribute of e be initialized to error.

  5. Queue a task to dispatch e at sensor_instance.

9.12. Request Sensor Access

input

sensor_instance, a Sensor object.

output

state, a permission state.

  1. Let sensor be the sensor associated with sensor_instance.

  2. let permission be the result of invoking the abstract operation retrieve the sensor permission associated with sensor, passing it sensor_instance as argument.

  3. Let state be the result of retrieving the permission state for permission.

  4. If state is "prompt",

    1. prompt the user in a user-agent-specific manner for permission to provide access to sensor.

    2. If permission is granted,

      1. set state to "granted",

    3. else,

      1. set state to "denied".

  5. return state.

10. Extensibility

This section is non-normative.

Its purpose is to describe how this specification can be extended to specify APIs for different sensor types.

Extension specifications are encouraged to focus on a single sensor type, exposing both high and low level as appropriate.

10.1. Naming

Sensor interfaces for low-level sensors should be named after their associated sensor. So for example, the interface associated with a gyroscope should be simply named Gyroscope. Sensor interfaces for high-level sensors should be named by combining the physical quantity the sensor measures with the "Sensor" suffix. For example, a sensor measuring the distance at which an object is from it may see its associated interface called ProximitySensor.

Attributes of the SensorReading subclass that hold sensor readings values should be named after the full name of these values. For example, the TemperatureSensorReading interface should hold the sensor reading’s value in a temperature attribute (and not a value or temp attribute). A good starting point for naming are the Quantities, Units, Dimensions and Data Types Ontologies [QUDT].

10.2. Unit

Extension specification must specify the unit of sensor readings.

As per the Technical Architecure Group’s (TAG) API Design Principles [API-DESIGN-PRINCIPLES], all time measurement should be in milliseconds. All other units should be specified using, in order of preference, and with the exception of temperature (for which Celsius should be favored over Kelvin), the International System of Units (SI), SI derived units, and Non-SI units accepted for use with the SI, as described in the SI Brochure [SI].

10.3. Exposing High-Level vs. Low-Level Sensors

So far, specifications exposing sensors to the Web platform have focused on high-level sensors APIs. [GEOLOCATION-API] [ORIENTATION-EVENT]

This was a reasonable approach for a number of reasons. Indeed, high-level sensors:

However, an increasing number of use cases such as virtual and augmented reality require low-level access to sensors, most notably for performance reasons.

Providing low-level access enables Web application developers to leverage domain-specific constraints and design more performant systems.

Following the precepts of the Extensible Web Manifesto [EXTENNNNSIBLE], extension specifications should focus primarily on exposing low-level sensor APIs, but should also expose high-level APIs when they are clear benefits in doing so.

10.4. When is Enabling Multiple Sensors of the Same Type Not the Right Choice?

TODO: provide guidance on when to:

10.5. Definition Requirements

At least the following definitions must be specified in each extension specification:

An extension specification may specify the following defintions:

10.6. Extending the Permission API

Provide guidance on how to extend the Permission API [PERMISSIONS] for each sensor types.

<https://github.com/w3c/sensors/issues/22>

10.7. Example WebIDL

Here’s example WebIDL for a possible extension of this specification for proximity sensors.

[Constructor(optional ProximitySensorOptions proximitySensorOptions)]
interface ProximitySensor : Sensor {
  readonly attribute ProximitySensorReading? reading;
};

interface ProximitySensorReading : SensorReading {
    readonly attribute unrestricted double distance;
};

dictionary ProximitySensorOptions : SensorOptions {
    double? min = -Infinity;
    double? max = Infinity;
    ProximitySensorPosition? position;
    ProximitySensorDirection? direction;
};
    
enum ProximitySensorPosition {
    "top-left",
    "top",
    "top-right",
    "middle-left",
    "middle",
    "middle-right",
    "bottom-left",
    "bottom",
    "bottom-right"
};

enum ProximitySensorDirection {
    "front",
    "rear",
    "left",
    "right",
    "top",
    "bottom"
};

11. Acknowledgements

First and foremost, I would like to thank Anssi Kostiainen for his continuous and dedicates support and input throughout the development of this specification.

Special thanks to Rick Waldron for driving the discussion around a generic sensor API design for the Web, sketching the original API on which this is based, providing implementation feedback from his work on Johnny-Five, and continuous input during the development of this specification.

Special thanks to Boris Smus, Tim Volodine, and Rich Tibbett for their initial work on exposing sensors to the Web with consistency.

Thanks to Anne van Kesteren for his tireless help both in person and through IRC.

Thanks to Domenic Denicola and Jake Archibald for their help.

Thanks also to Frederick Hirsch and Dominique Hazaël-Massieux (via the HTML5Apps project) for both their administrative help and technical input.

Thanks to Tab Atkins for making Bikeshed and taking the time to explain its subtelties.

The following people have greatly contributed to this specification through extensive discussions on GitHub: Anssi Kostiainen, Boris Smus, chaals, Claes Nilsson, Dave Raggett, David Mark Clements, Domenic Denicola, Dominique Hazaël-Massieux (via the HTML5Apps project), Francesco Iovine, Frederick Hirsch, gmandyam, Jafar Husain, Johannes Hund, Kris Kowal, Marcos Caceres, Marijn Kruisselbrink, Mark Foltz, Mats Wichmann, Matthew Podwysocki, pablochacin, Remy Sharp, Rich Tibbett, Rick Waldron, Rijubrata Bhaumik, robman, Sean T. McBeth, smaug----, Tab Atkins Jr., Virginie Galindo, zenparsing, and Zoltan Kis.

We’d also like to thank Anssi Kostiainen, Dominique Hazaël-Massieux, Erik Wilde, and Michael[tm] Smith for their editorial input.

Conformance

Document conventions

Conformance requirements are expressed with a combination of descriptive assertions and RFC 2119 terminology. The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in the normative parts of this document are to be interpreted as described in RFC 2119. However, for readability, these words do not appear in all uppercase letters in this specification.

All of the text of this specification is normative except sections explicitly marked as non-normative, examples, and notes. [RFC2119]

Examples in this specification are introduced with the words "for example" or are set apart from the normative text with class="example", like this:

This is an example of an informative example.

Because this document doesn’t itself define APIs for specific sensor typesthat is the role of extensions to this specification—all examples are inevitably (wishful) fabrications. Although all of the sensors used a examples would be great candidates for building atop the Generic Sensor API, their inclusion in this document does not imply that the relevant Working Groups are planning to do so.

Informative notes begin with the word "Note" and are set apart from the normative text with class="note", like this:

Note, this is an informative note.

Conformant Algorithms

Requirements phrased in the imperative as part of algorithms (such as "strip any leading space characters" or "return false and abort these steps") are to be interpreted with the meaning of the key word ("must", "should", "may", etc) used in introducing the algorithm.

Conformance requirements phrased as algorithms or specific steps can be implemented in any manner, so long as the end result is equivalent. In particular, the algorithms defined in this specification are intended to be easy to understand and are not intended to be performant. Implementers are encouraged to optimize.

Conformance Classes

A conformant user agent must implement all the requirements listed in this specification that are applicable to user agents.

Index

Terms defined by this specification

Terms defined by reference

References

Normative References

[HR-TIME-2]
Ilya Grigorik; James Simonsen; Jatinder Mann. High Resolution Time Level 2. 20 July 2016. WD. URL: https://www.w3.org/TR/hr-time-2/
[HTML]
Ian Hickson. HTML Standard. Living Standard. URL: https://html.spec.whatwg.org/multipage/
[PAGE-VISIBILITY]
Jatinder Mann; Arvind Jain. Page Visibility (Second Edition). 29 October 2013. REC. URL: https://www.w3.org/TR/page-visibility/
[PERMISSIONS]
Mounir Lamouri; Marcos Caceres. The Permissions API. 7 April 2015. WD. URL: https://www.w3.org/TR/permissions/
[POWERFUL-FEATURES]
Mike West. Secure Contexts. 19 July 2016. WD. URL: https://www.w3.org/TR/secure-contexts/
[RFC2119]
S. Bradner. Key words for use in RFCs to Indicate Requirement Levels. March 1997. Best Current Practice. URL: https://tools.ietf.org/html/rfc2119
[WebIDL-1]
Cameron McCormack; Boris Zbarsky. WebIDL Level 1. 8 March 2016. CR. URL: https://www.w3.org/TR/WebIDL-1/
[WHATWG-DOM]
Anne van Kesteren. DOM Standard. Living Standard. URL: https://dom.spec.whatwg.org/

Informative References

[API-DESIGN-PRINCIPLES]
Domenic Denicola. API Design Principles. 29 December 2015. URL: https://w3ctag.github.io/design-principles/
[EXTENNNNSIBLE]
The Extensible Web Manifesto. 10 June 2013. URL: https://extensiblewebmanifesto.org/
[GEOLOCATION-API]
Andrei Popescu. Geolocation API Specification. 28 May 2015. PER. URL: https://www.w3.org/TR/geolocation-API/
[ORIENTATION-EVENT]
Rich Tibbett; et al. DeviceOrientation Event Specification. 18 August 2016. CR. URL: https://www.w3.org/TR/orientation-event/
[QUDT]
Ralph Hodgson; et al. QUDT - Quantities, Units, Dimensions and Data Types Ontologies. 18 March 2014. URL: http://www.qudt.org/
[SI]
SI Brochure: The International System of Units (SI), 8th edition. 2014. URL: http://www.bipm.org/en/publications/si-brochure/

IDL Index

interface Sensor : EventTarget {
  readonly attribute SensorState state;
  readonly attribute SensorReading? reading;
  void start();
  void stop();
  attribute EventHandler onchange;
  attribute EventHandler onstatechange;
  attribute EventHandler onerror;
};

dictionary SensorOptions {
  double? frequency;
};

enum SensorState {
  "idle",
  "activating",
  "active",
  "errored"
};

interface SensorReading {
  readonly attribute DOMHighResTimeStamp timeStamp;
};

[Constructor(DOMString type, SensorReadingEventInit eventInitDict)]
interface SensorReadingEvent : Event {
  readonly attribute SensorReading reading;
};

dictionary SensorReadingEventInit : EventInit {
  SensorReading reading;
};

[Constructor(DOMString type, SensorErrorEventInit errorEventInitDict)]
interface SensorErrorEvent : Event {
  readonly attribute Error error;
};

dictionary SensorErrorEventInit : EventInit {
  Error error;
};