RE: EME test reports

On clearkey-mp4-playback-temporary-waitingforkey.html:  I don’t know the history, but the test report tool is written to ignore tests that result in all timeouts.  In fact, in its normal version it ignores tests with only 1 pass/fail result.  I’ve worked around that for our currently published reports.

If I count tests that all timeout as fails, the “complete failure” group jumps dramatically:


·       The modified report is:  Completely failed files: 54; Completely failed subtests: 45; Failure level: 45/294 (15.31%)

·       Vs. online which is:  Completely failed files: 50; Completely failed subtests: 10; Failure level: 10/299 (3.34%)

We clearly need to better understand what’s causing the timeouts.

Jerry

From: David Dorwin [mailto:ddorwin@google.com]
Sent: Monday, August 29, 2016 1:18 PM
To: Mark Watson <watsonm@netflix.com>
Cc: Jerry Smith (WPT) <jdsmith@microsoft.com>; public-html-media@w3.org; Philippe Le Hégaret <plh@w3.org>; Paul Cotton <Paul.Cotton@microsoft.com>; Matthew Wolenetz <wolenetz@google.com>
Subject: Re: EME test reports

A couple notes on the results:

  *   clearkey-mp4-playback-temporary-waitingforkey.html times out on all three browsers but doesn't appear in complete-fails.html or less-than-2.html. Is this a bug in the report generation tool?
  *   drm-keystatuses.html fails or times out on all browsers. We may want to review that test.

On Mon, Aug 29, 2016 at 12:44 PM, Mark Watson <watsonm@netflix.com<mailto:watsonm@netflix.com>> wrote:
I think we should replace the key system name in the test name with either 'drm' or 'clearKey'. Since each browser supports only a single DRM (that we can test), then we know the keysystem that is being tested in each case.

...Mark

On Aug 29, 2016, at 12:36 PM, Jerry Smith (WPT) <jdsmith@microsoft.com<mailto:jdsmith@microsoft.com>> wrote:
One thing we should probably resolve:  Including key system strings in the subtest name is creating misleading test gaps vs. our 2 passing implementation requirement.  The wptreport tool treats these as separate tests.  All com.microsoft.playready subtests are flagged currently as not having two passing implementations.

As a simple experiment, I edited the JSONs to pull the key system names from the drm subtest names.  That results in:


-          Test files: 105; Total subtests: 257

-          Test files without 2 passes: 29; Subtests without 2 passes: 49; Failure level: 49/257 (19.07%)

-          Completely failed files: 29; Completely failed subtests: 8; Failure level: 8/257 (3.11%)

Vs. online (with key system names in subtest names);


-          Test files: 105; Total subtests: 299

-          Test files without 2 passes: 50; Subtests without 2 passes: 91; Failure level: 91/299 (30.43%)

-          Completely failed files: 50; Completely failed subtests: 10; Failure level: 10/299 (3.34%)

That means:


-          Listing the key system in the subtest names results in 42 additional subtests.

-          Most of these have more than one passing implementation presently.

-          2 subtests (drm-keystatuses-multiple-sessions) report as complete fail with key system in the subtest name, but have one passing implementation when key system is not.

Options to handle this are:


-          Remove the keysystem names from the subtests, but lose visibility into keysystem specific failures.

-          Leave them in and post those results to the website, but prepare manual tallies when we submit PR for review.

Jerry

Received on Monday, 29 August 2016 21:17:41 UTC