19:02:57 RRSAgent has joined #aria-at 19:03:01 logging to https://www.w3.org/2025/09/04-aria-at-irc 19:03:01 RRSAgent, make logs Public 19:03:02 please title this meeting ("meeting: ..."), jugglinmike 19:03:17 meeting: ARIA and Assistive Technologies Community Group Weekly Teleconference 19:03:19 present+ jugglinmike 19:03:21 scribe+ jugglinmike 19:03:26 present+ Carmen 19:03:33 present+ dean 19:03:37 present+ ChrisCuellar 19:03:46 present+ Matt_King 19:03:54 present+ mfairchild_ 19:04:00 present+ kelly 19:04:05 present+ howard-e 19:04:22 Topic: Review agenda and next meeting dates 19:04:37 https://github.com/w3c/aria-at/wiki/September-4%2C-2025-Agenda#js-repo-pjax-container 19:04:55 Matt_King: Requests for changes to agenda? 19:05:16 Carmen: I had a question about AT support with the bots, and another question from ChrisCuellar 19:05:42 Matt_King: We can definitely make room for both of those; we'll tee them up for after the agenda items that we've already planned 19:05:45 Carmen: Thank you! 19:05:56 Matt_King: Next AT Driver Subgroup meeting: Monday September 8 19:06:02 Matt_King: Next CG meeting: Wednesday September 10 19:07:12 Topic: Current status 19:07:23 Matt_King: The big thing this week is that the "accordion" plan advanced 19:07:34 Matt_King: That was a lot of work resolving conflicts! Thank you dean for your part in that 19:07:44 Matt_King: This bumps us up to 17 plans in "Candidate Review" 19:07:58 Matt_King: And we have seven plans in "draft review". Four of those are on today's agenda 19:08:32 howard-e has joined #aria-at 19:08:36 present+ 19:09:16 Matt_King: Assuming we make good progress on these over the next week or so--I'm hoping that these seven (plus one more that's approach "draft review") will be able to move forward 19:09:36 Matt_King: There was a JavaScript error with that eighth plan, though; I made a comment, and I'm hoping we can work it out 19:09:49 howard-e: Yes, I saw that, and I've responded with a suggestion for a fix 19:10:06 howard-e: We shouldn't be setting the ARIA property; we should be setting the HTML property 19:10:15 Matt_King: Ah, right. Why didn't I catch that? 19:10:48 howard-e: I also made another note in another pull request. This may be for a separate discussion, but the pattern says that something is required, but it isn't mentioned in the example 19:11:05 Matt_King: Ah, yes. Even in the references, it's not really a test of aria-checked; it's a test of HTML checked 19:11:20 Matt_King: It sounds like maybe I should go back to the APG documentation because the APG documentation might not be reflecting that fact 19:11:25 Matt_King: That's a good call-out. Thank you 19:11:53 Matt_King: Anyway, that's where we're at right now. It's very possible that we could reach 25 test plans by the end of the month 19:12:13 Topic: What was solution to VO accordion conflicts? 19:12:33 Matt_King: We landed accordion, and we discussed some of the issues last week, but we didn't discuss them in the context of a GitHub issue 19:12:54 Matt_King: According to the minutes from last week, dean was going to check on a different macBook that he has with a fresh install 19:13:22 Matt_King: Ultimately, I don't know how this conflict was resolved. You were using the "b" key to navigate backwards, and instead of it doing "quick nav", it was doing autocomplete 19:13:39 Matt_King: I wanted to create a record here because I'm not sure it will be the last time we experience this problem. How did we solve it? 19:14:43 dean: In a follow-up e-mail, I confirmed it was doing exactly the same thing in a fresh mac. If you start the test with "quick nav" on, and you go to that field and type the letter "b", it demonstrates the expected behavior. If you toggle quick nav, first, then it demonstrates the behavior I originally reported 19:15:11 Matt_King: It is a confusing issue. If "quick nav" is on by default, then setting focus on a field is having a side-effect of disabling quick nav 19:15:27 dean: I'm guessing that's what it does, because you can toggle it on and off, and then it will work, too 19:16:16 Matt_King: I don't know if Apple would consider this a bug or not. It's marginal--whether this is in-scope for this project 19:16:33 dean: I think that it behaves correctly. I think the test is not really correct 19:16:50 dean: Because I think that when you're in a form field and you type "shift+b", you should get a capital B 19:17:13 Matt_King: Well if quick nav is indeed turned on, then the point of having quick nav on is that it does NOT type a "b" 19:17:26 dean: Okay, so then it is a problem with quick nav, then 19:17:38 Matt_King: The setup script does not touch any of the VoiceOver settings 19:18:00 Matt_King: We'll stick with the way that IsaDC handled it. I'm going to make a note to myself to ask James Craig or someone else at Apple about this 19:19:08 Matt_King: Maybe I'll phrase it like this: "When focus is set in a text field, should quick nav functions work if quick nav is on? Or does setting focus temporarily override the quick nav setting?" 19:19:37 Matt_King: Great; that's very helpful to know. Thank you 19:19:51 Matt_King: If Apple says this is a bug, then we might want to change this to document an unexpected side-effect 19:20:04 dean: Yeah, we could do that It's only one test as I recall 19:20:12 Topic: Running Disclosure of Answers to Frequently Asked Questions test plan 19:20:27 Matt_King: There are 0 conflicts right now. JAWS and NVDA are complete 19:20:45 Matt_King: But the active testers are absent today, so there's actually not much to talk about right now 19:20:52 Topic: Running Switch test plan 19:21:08 Matt_King: We'll skip this due to inattendance, as well 19:21:19 Topic: Running test plan for Switch Example Using HTML Button 19:21:42 Matt_King: We have NVDA and JAWS responses available, ready for a volunteer to assign verdicts 19:21:49 dean: I had a question about my work here 19:22:15 dean: If the test is operating a switch is off, after doing "ctrl+space", it says "no actions available", would it be untestable? 19:22:26 Matt_King: It doesn't sound untestable--it just didn't work 19:22:52 dean: So we just record the output that we got and fail it 19:23:10 s/"ctrl/"ctrl+option/ 19:23:53 Matt_King: Yes 19:24:03 dean: I can pick up the NVDA work, as well. 19:26:52 Topic: Running test plan for Tabs with Automatic Activation 19:27:05 Matt_King: None of the assigned Testers are present today, so we'll skip this, as well 19:27:31 Topic: Bot AT Version updates 19:27:37 Carmen: I have two questions 19:27:55 Carmen: First: should we update NVDA Bot and VoiceOver Bot to the latest version? 19:28:18 Carmen: The second is more about process: who should give the green light for updating the bots? 19:28:40 Matt_King: Can we currently pick a bot version? Or do we only support one version at a time 19:29:03 Carmen: You can choose, but only versions that we had previously working 19:29:15 Matt_King: In general, we always want to move to the latest version when it becomes public 19:29:27 ChrisCuellar: In that case, I think it's still a process question. 19:29:48 ChrisCuellar: At least on the NVDA- and VoiceOver- side (for now), we will probably notice when there is a new version sooner than the CG 19:30:24 ChrisCuellar: Should we flag it, bring it to the CG, get approval to update, and only then update? Or should we just go ahead and update it on our own and let everyone know? 19:31:01 Matt_King: I'm trying to think of any scenarios where updating the bot could be disruptive or cause problems 19:31:14 Matt_King: When we have plans in the test queue, and we do a new run for a new version... 19:31:29 Matt_King: I think that in general, due to the way we use the bots, that it's not going to be a problem 19:31:44 Matt_King: I don't see anything wrong with Bocoup being proactive with new releases and just informing the CG 19:32:16 ChrisCuellar: Before we started running our own macOS images--that was the case. With GitHub, they would just update the image without warning. Now, we have more control. And it's a little different for each bot 19:32:34 ChrisCuellar: But even JAWS is going to let us know about new reaeaes 19:32:43 s/reaeaes/releases/ 19:33:10 ChrisCuellar: historically, with VoiceOver, we didn't know when it was upgrading--at least, not with major releases. We just had to keep an eye on GitHub's releases 19:33:26 ChrisCuellar: So far, it wasn't been a problem that we didn't exactly know which version was running 19:33:36 ChrisCuellar: But this leads me to what I think is a larger product question 19:34:01 ChrisCuellar: I was wondering if it would be better to move to a UI where an admin or tester could just select an available version of the bot to run 19:34:22 ChrisCuellar: Right now, there's some logic in the back-end that tries to match the best bot to the AT version that was selected for the test plan run 19:34:36 ChrisCuellar: ...but I'm wondering if it might be easier/simpler to just let the tester select the version that they want to use 19:34:55 Matt_King: Yeah, though there will probably always be cases where the tester wants to use an older version 19:35:26 Matt_King: It's pretty cool if we have those images there, now, and we can just let people do that 19:35:55 ChrisCuellar: I think with JAWS, we could request that the JAWS version is a flag that gets deployed on GitHub when we're running the test 19:36:15 Matt_King: So, for JAWS and NVDA, is it a script that builds the image? 19:36:26 ChrisCuellar: I think with JAWS, we just have a link to the latest build 19:36:33 Matt_King: So we're not storing a binding? 19:36:40 ChrisCuellar: Not currently 19:37:26 Matt_King: I think that might be an improvement. It feels like storing a binding that we use on-demand makes the infrastructure a little more robust (compared with relying on a link that could change) 19:37:31 ChrisCuellar: Definitely 19:37:57 ChrisCuellar: We have some changes to the bot-running UI coming up soon. We can revisit how much granularity we allow to admins and testers when they select a bot 19:38:40 https://github.com/w3c/aria-at-app/issues/1467 19:39:05 Topic: Add information about conflicts to the test plan status column for automated re-runs 19:39:08 github: https://github.com/w3c/aria-at-app/issues/1467 19:39:38 ChrisCuellar: This is kind of a gnarly product terminology question that we've been discussing a bit on the issue itself 19:40:56 ChrisCuellar: After an automated re-run is completed when there's a new bot version--right now, due to the way that the bots and testers are rendered on the test queue, it's really confusing. Whenever there are response conflicts between an automated re-run and the previous report--in those situations, we record the mismatch between the AT outputs, and we leave the verdict unassigned 19:41:23 ChrisCuellar: So the new run of the test plan is in a state where there are a certain number of unassigned verdicts. A human tester needs to go in and verify, maybe run the tests manually to confirm, etc. 19:41:42 ChrisCuellar: When that happens, it's very hard to find exactly how many tests in a test plan are in a certain state 19:42:19 Matt_King: Basically, we have two runs of a test plan, a prior report that's already been published (that we're comparing to) and a new run that we're comparing to that prior report 19:42:41 Matt_King: ...and we're comparing what the bot got for a response from the AT in the new one to what's in the prior report (whether from humans or bots) 19:43:02 Matt_King: This is very similar to what we do when we have two humans running it. We compare responses from person "a" and person "b" 19:43:33 Matt_King: The way we've done that, to date, is we've actually looked at verdict mismatches, only. When there are four conflicts, it means that there are four verdicts that don't match 19:43:50 Matt_King: We don't surface in the status the mismatches in the response. 19:44:09 Matt_King: That's where I think we're getting bogged down. We're talking about two different kinds of conflicts, but they have the same effect. 19:44:26 Matt_King: I'm hoping we can simplify things. We have four conflicting responses or four conflicting verdicts 19:45:00 ChrisCuellar: You were proposing to add a new kind of conflict terminology to describe the case where you're taking over an automated re-run and there are unassigned verdicts due to differences 19:45:33 ChrisCuellar: We were a little concerned about exactly what you just said. When we report conflicts in the "Status" column right now, it links to a conflict resolution screen, and there's a whole process 19:46:17 ChrisCuellar: So rather than adding a new type of conflict and muddying the meaning of the "status" column, I thought it might be cleaner to surface all ofthe stats that we currently surface for bot runs--to surface those for all human testers, too 19:46:37 ChrisCuellar: The statistic about "how many verdicts have been assigned" is what we're trying to chase right now, I think 19:46:49 ChrisCuellar: The human tester only has the number of tests complete. 19:47:02 ChrisCuellar: I don't remember the historical reasons why there are differences in how we report progress 19:47:30 Matt_King: In the very beginning, humans were always recording both the responses and verdicts at the same time. Historically, it was not a thing 19:48:31 ChrisCuellar: What's confusing about the UI right now is that when you re-assign a bot run to a human tester, it's hard to know how many conflicts you have to resolve at that point. You only know how many tests have been submitted, which is sort of a different number 19:48:56 Matt_King: It's really helpful to know if there were any differences in responses (and if so, how many). That tells you the magnitude of work that you have to do 19:49:12 Matt_King: It's the number of responses that tells you the amount of work that you have to do--not the number of verdicts 19:49:33 Matt_King: Once you know you have the responses settled, then rendering the verdicts is almost trivial 19:50:15 ChrisCuellar: To your point about now knowing how many responses just didn't match--I wonder if we keep that also for human testers, and that we also report there how many responses were recorded but mismatched from the previous run--maybe that's a solution 19:50:37 ChrisCuellar: Essentially, I was suggesting that we keep the same stats for all testers whether they are a bot or a human because that seems more informative 19:51:42 Matt_King: I think if we're always talking about responses and verdicts (which is always applicable whether you are talking about bots or humans), then you could report on missing verdicts. We could say "X or Y verdicts complete" or "X of Y conflicts compete" 19:52:01 Matt_King: The idea of "X conflicting verdicts" and "X conflicting responses" would then make more sense 19:52:09 s/conflicts complete/responses complete/ 19:52:17 Matt_King: then we're always reporting the same two stats 19:52:22 Matt_King: The number of tests is not relevant 19:52:36 ChrisCuellar: Yeah, I was thinking that, too. The percent complete really reflects the assignment 19:52:48 Matt_King: We can get rid of that noise. We don't care about the number of tests 19:53:02 ChrisCuellar: To me, the "percent complete" was always very confusing 19:54:03 Matt_King: Let's figure out new content for the "testers" column where we only talk about responses and verdicts for both humans and bots. You could probably consolidate 19:54:29 Matt_King: It could be a little more compact--we could figure out a wording that's more concise 19:54:39 ChrisCuellar: Maybe it's better to always use percentages 19:55:00 Matt_King: Yeah, maybe. But we always want to show counts for conflicts because that's a statistic where the number truly matters 19:55:42 Matt_King: Maybe we put something like, "Progress: 40% of responses, X% of verdicts" Always put responses first because that's what you record, first 19:55:51 Matt_King: You might be able to make it pretty concise that way 19:56:06 ChrisCuellar: We'll brainstorm it a bit, but I hear you: make it concise, aim for shorter 19:56:28 ChrisCuellar: If we can get that progress statement as concise as possible for each tester, I think that's the goal 19:56:58 Matt_King: I would still want to go back to this concept of reporting the number of conflicting responses and reporting the number of conflicting verdicts 19:57:15 Matt_King: You could shorten it: "Conflicts: 0" or "Conflicts: 4 responses, 8 verdicts" 19:57:20 Matt_King: That would be pretty short 19:58:28 ChrisCuellar: It's still a problem, if we show it in the status--only one of them should be a link. 19:58:50 Matt_King: The whole text could be one link. "Conflicts: 4 responses, 8 verdicts" would be one link that leads to the conflicts page 20:00:38 ChrisCuellar: These aren't conflicts in the way that we typically think about them--we'll have a row with only one tester 20:00:44 Matt_King: Ah, I see. In that case, you're right 20:01:08 Matt_King: Now, I'm actually thinking that the percentage in the "testers" column is less helpful because the absolute numbers are better 20:01:47 ChrisCuellar: I do feel like the whole review process for an automated re-run is a different enough workflow to warrant a different screen/UI. I think trying to show-horn all these concepts into the test queue as we know it is maybe a problem. 20:02:02 ChrisCuellar: I think that's why we have all this confusion right now 20:02:19 Matt_King: We did it this way because wanted to limit the amount of new UI to develop 20:02:31 Matt_King: If we can solve this with content instead of code, I'd like to do that 20:03:19 Matt_King: I can revisit this issue with my new understanding of the problem (that is: the conflicts for automated re-runs being substantively different) 20:03:32 Matt_King: I'll put this on my to-do list to make a comment here 20:03:47 Matt_King: It sounds like we still made some good progress in terms of how we're thinking about this problem, thogh 20:03:51 s/thogh/though/ 20:03:56 ChrisCuellar: Agreed 20:04:54 Zakim, end the meeting 20:04:54 As of this point the attendees have been jugglinmike, Carmen, dean, ChrisCuellar, Matt_King, mfairchild_, kelly, howard-e 20:04:57 RRSAgent, please draft minutes 20:04:58 I have made the request to generate https://www.w3.org/2025/09/04-aria-at-minutes.html Zakim 20:05:06 I am happy to have been of service, jugglinmike; please remember to excuse RRSAgent. Goodbye 20:05:06 Zakim has left #aria-at 20:08:57 RRSAgent, leave 20:08:57 I see no action items