Outline plan is to run some sort of Beta Trial in Boston.
...
Users get notified when they have been in contact with an infected person, with few false positives, and few false negatives.
Notifications are timely, relative to the contact tracing interview that identified the “points of concern”
The app provides clear information to users about what has been detected, and what steps they should take
The app provides a high-quality user experience: slick, attractive, usable.
The app does not cause frustration
The app does not inconvenience (battery usage, data costs, unhelpful notifications, other problems)
The user trusts the app.
The app user is very privacy conscious - match behavior with expectations
Asymptomatic vs Symptomatic users - contact tracing impact
Diagnosed users:
The contact tracing experience is clear, straightforward, and informative.
Contact tracing based on data from the app is superior to contact tracing without the app.
The user continues to have a positive experience after the contact tracing is complete.
Is he able to identify where he got the infection from
Is he able to ask his family/friends before sharing any details with contact tracers
Contact tracer
The contact tracing process is clear and straightforward
It is straightforward to publish points of concern in Safe Places.
Contact tracing based on data from the app is superior to contact tracing without the app.
It is straightforward to redact data to meet a user’s privacy needs.
Does it reduce the load on contact tracer considerably - able to do more no of patients (how many more? in the same time?
Does it accelerate publication of data significantly?
...
The app supports contact tracing efforts
The app helps to reduce the spread of COVID-19
Does it mean health facility requiring less staff?
…and how might we measure it?
Objective | Measurement | Implementation |
---|---|---|
Users get notified when they have been in contact with an infected person, with few false positives, and few false negatives. | What % of notifications did / did not seem to match an exposure? | Daily report from all experiment participants, reporting screenshots of any notifications, and their own view of when/where they might have been. One day later, we publish comentary describing the “points of concern” locations, published the previous day. Participant then follows up with their classification of events as true positive / false positive / false negative, with explanations. |
Notifications are timely, relative to the contact tracing interview that identified the “points of concern” | Time lag between completion of contact tracing interview, and notifications. | For each point of concern published, we track the time of the contact tracing interview (real or imagined) that generated it. Participants sharing screenshots of notifications also indicate the arrival time of the notification. |
The app provides clear information to users about what has been detected, and what steps they should take | Qualitative user input | Covered by a standard set of questions that the participant answrs for each notification they receive. |
The app provides a high-quality user experience: slick, attractive, usable. | Participant feedback via survey | Participant survey after 2d, 1 week, 2 weeks? |
The app does not cause frustration | Participant feedback via survey | Participant survey after 2d, 1 week, 2 weeks? |
The app does not inconvenience (battery usage, data costs, unhelpful notifications, other problems) | Participant feedback via survey | Participant survey after 2d, 1 week, 2 weeks? |
The user trusts the app. | Participant feedback via survey | Participant survey after 2d, 1 week, 2 weeks? |
The app user is very privacy conscious - match behavior with expectations | Participant feedback via survey | Participant survey after 2d, 1 week, 2 weeks? |
Asymptomatic vs Symptomatic users - contact tracing impact | Deepti gulati pahwa - I didn’t understnd this one… | |
DIagnosed users… | ||
The contact tracing experience is clear, straightforward, and informative. | Participant feedback via survey | Survey after contact tracing interview |
Contact tracing based on data from the app is superior to contact tracing without the app. | Compare surveys of participants interviewed having used, or not used, the app. | Some participants have the app & are contact traced Some participants are contact traced without having installed the app. Both fill in the same survey questions. |
The user continues to have a positive experience after the contact tracing is complete. | Follow-up survey | Everyone asked to install the app after contact tracign interview (if they didn’t have it alreadu). Specific survey 3d after contact tracing interview |
Is he able to identify where he got the infection from | Not sure about this as a thing to try to measure… | I am not sure that our narative for fictional contact tracing events needs to include a fictional point of origin for the infection - especially in the case where the user does not have the app. |
typically be 3-4 days prior to symptoms, test & contact trace. | ||
Is he able to ask his family/friends before sharing any details with contact tracers | I am not sure we have designed for this… | May be better addressed to Design team in the first instance, rather than trying to answer this in Beta trial? |
Contact tracer | ||
The contact tracing process is clear and straightforward | Survey with contact tracer | Survey after each contact tracing interview & general survey after completing several of them. This should cover the cases with & without the App. |
It is straightforward to publish points of concern in Safe Places. | Survey with contact tracer | Include in contact tracer survey |
Contact tracing based on data from the app is superior to contact tracing without the app.
| Survey with contact tracer. | Ask explicit questions on this. But also compare scores between the two types of contact trace exercise. We could also do some contact trace experiemtns where the particiant has been running the app, but does not make use of it. We can then review the data here afterwards with the contact tracer & participant to determine whether any significant data points were missed. |
It is straightforward to redact data to meet a user’s privacy needs. | Survey with contact tracer. | Include in contact tracer survey |
Does it reduce the load on contact tracer considerably - able to do more no of patients (how many more? in the same time? | Measure duration of contact trace interview & any follow-up work. | Compare interviews with & without the app. |
Does it accelerate publication of data significantly? | Measure time from interview to publication of data with & without the app. | Record time lag from contact trace interview completing to (a) data being published, and (b) participants getting notifications from that data. |
Health authority | ||
The app supports contact tracing efforts | Interview with HA administrators. | After some number of contact tracing interviews have been completed, we allow HA administrators to conduct their own interviews of contact tracers and participants, before responding to our survey. |
The app helps to reduce the spread of COVID-19 | Interview with HA administrators. | This interview should be informed by data & analysis of points above. |
Does it mean health facility requiring less staff? | Interview with HA administrators. | This interview should be informed by data & analysis of points above. |
Diagnostics
It’s not enough to determine that we are falling short on some goal. We need information that will allow us to understand and rectify the cause of the problem. We expect the following diagnostics will be useful
...
Review, feedback & sign-off of this plan
Resolve unanswered questions re: target participant group & desired numbers
Recruit particpants
Recruit contact tracers
Get special build of the App from Dev with appropriate feature flags as required
Agree with Dev whether to include Firebase / Crashlytics
Define “daily reporting” ask for participants, and supporting technology
Define who will receive daily reports from participants, and what analysis they will do on them
Define what the daily register of points of concern looks like, and how it is published
Design & test procedures for pushing points of concern that don’t come from contact tracing interviews
Set up supporting systems for contact tracing & publishing to take place (Safe Places instance)
Set schedule for contact tracing interviews,
Define participant surveys post-contact tracing (immediate & 3 days later)
Define contact tracer survey post-contact tracing (immediate & 3 days later)
Processes & tech to administer post-contact tracing surveys
Define who will collate & analyze info from contact tracing surveys
Overall onboarding brief for participants
Overall onboarding brief for contact tracers
Identify Safe Paths volunteers to stand in for contact tracers if we don’t have enough.
Plan for follow-up with Health Authority admins (giving them access to participating contact tracers & participants).
Participation agreement for participants? (probably needed since PII shared)
Participation agreement for contact tracers? (maybe not needed since no PII shared)
Create overall project plan (covering all items above + whatever else) and identify a PM to run this.
Establish target dates for program, up to a first report with real data from the Beta trial.
Define governance plan for this program: regular reviews of whether we are achieving the goals we set out to achieve, any other issues.
Define & implement data retention policies.
Define person at Path Check responsible for us acting responsibly with participants data.
Previous Design Notes
These notes were recorded in an early draft of this article, and may still contain useful ideas and insights, so I am preserving them here…
...
Drive predictable volumes of “contact trace” interviews
Tech validation
Collect complete location data from non-infected patients to assess what matched & what didn’t
test location contexts form wide area
test moving objects with moving paths?
test location related interactive behaviors between multiple mobile users and objects
Location services on and off between different test users - impact
Social system validation
Get feedback from the person who participated in the interview
Get feedback from non-infected patients as to who matched a given location.
Volunteer Base to be used (question):
Harvard students - putting them at risk?
Health officials working daily - also with actual COVID patients ( Mayo Clinic staff, or something similar)
Fedex/ logistic company delivery people - as they move around city.
Small geographical area to consider - (Boston? Or smaller - to ensure paths are crossed often.
Tech Aspects to consider
Run detailed analytics / diagnostics on individual phones: firebase, crashlytics etc.
Collect daily? location logs from every phone & check for reliability of logging.
Get qualitative feedback from individuals
Set up lots of infections, and get a much higher rate of notifications than we could do otherwise.
Push up scale / number of infections/ data points to download.
What features we want to test - only crossed Paths?
Check the difference in mapping when we use bluetooth only vs / bluetooth and GPS? - possibility MVP2 - is ready to test?
Contact tracers perspective - safe places - new
Risks related to experimentation: Product dependencies - must be in place before we start
...
How do we measure “effectiveness” ? Define KPIs to make this prototype/ simulation successful ( Tech as well as non-tech)
User accounts of their movements vs. what the location trails tell us vs. what notifications triggered? Use this info to try to account for numbers of false negatives, true positives & false positives…
Epidemiological view of effectiveness - independent review of stuff in previous bullet?
User feedback on messaging - how seriously do they take it? Do they know what to do? Other feedback?
Contact tracer feedback.
MoP feedback on contact tracing experience
60% penetration vs 20% penetration of app for contact tracing to work. (is there a way to test that through simulation.
How many people?
Depends how active / mobile / engaged people are going to be.
Directed interactions only need a small number of people (< 10) to be able to do some effective stuff, (controlled experiment)
Larger groups of people enable more non-directed learnings, much more unexpected stuff will start to happen as we get to 50+ people.
Much more tech/automation needed to process from 50+ people than from 5-10 people - with 5-10 lots could be manual.
Suspect we should aim for ~10 people for a week, then grow by ~20 people/week so we are at 50 people after 3 weeks. Not obviously going to get lots more benefit from scaling above 50 people, and will become increasingly challenging to organize….
2 groups -
10 people - directed learnings with pre-determined KPIs
They should follow plans, and on purpose cross paths to validate our idea of false positives and false negatives. (public transportation, cafe, office building)
50 people - 2nd group - non- directed - get unexpected feedback. (extra feedback on what info they trust - and how important is privacy to them at all levels. Develop a questionnaire) non-technological part.
Use cases for directed learnings -
Cafes
Grocery stores
Public transportation - false positives/ negatives possibility
Office buildings - floors
Diffrence in logging - wifi vs 3g vs 4g
User journey aspects to consideration:
App perspective - user experience
IOS versions (different) vs android versions - range of phone devices to be checked. Select and test a set of popular mobile devices
User behavior in surroundings
Demographics behaviour - old age 60+, parents and kids, essential workers
If social distancing is maintained - people practicing, people not practicing.
Mobility impacts - on feet, bicycle, by car, by metro/train, bus/tram
Contact tracer/interview:
Dashboard - what helps (through interviews generate features)