There is little doubt regarding the usefulness and practicality of 
digital phenotyping to measure patient health and behavior. Traditional 
clinical scales––the current gold standards for primary endpoints––can 
be cumbersome and demanding. Digital measurements purport to solve this 
problem. But the questions of their reliability as accurate assessments 
remain.
Ultimately, both digital and traditional measures are based on a 
common understanding: Disease often manifests itself in observable 
behavior; behavior that can be measured. Traditional clinical scales do 
this through clinician observation. Digital measurements do this through
 a combination of computational tools that mimic and in many cases 
enhance what a clinician would observe.
Of course, none of this is novel. Researchers have been developing 
digital measurement tools for decades. Computer vision-based coding of 
facial expressions showed reliability in the 1990s. Same goes for 
ubiquitously used software to measure vocal acoustics. From the very 
beginning, our goal has simply been to replicate the reliability of 
these measurements on the AiCure platform.
And that’s what we did. We scoured the scientific literature for 
digital measurements shown to be related to clinical functioning and 
replicated those methods so we could make those same measurements using 
our own platform. All of a sudden, we realized we were in sole 
possession of a unique codebase for reliable measurement of a multimodal
 set of facial and vocal digital biomarkers.
And so we put it all back in the public domain.
Folks have been confused by our decision to do this––but to us, it’s 
the only way forward. We believe that trust for novel measures should be
 built in the public domain, through peer-review, by the scientific and 
medical community. But this process has been remarkably slow, partially 
because methods for digital phenotyping have historically been disparate
 and difficult to access. 
By making our code open source, we are increasing the accessibility of these methods to more researchers and our partners in industry. Consequently, we can all contribute to these methods, we can unrestrictedly comment on the reliability and validity of these methods, and we can collectively move closer towards adoption of these methods in patient care and clinical research.
I invite you all to check out our code on GitHub, read through the list of biomarkers that can be calculated through it, and watch a recording of the webinar we hosted when we launched the software in the fall of 2020. And as always, please reach out if you want to chat.
On March 3, Anzar will present with Colin Sauder PhD, Director, Clinical Scientist at Karuna Therapeutics, on “Moving Digital Biomarkers from the Future into the Present,” at SCOPE Virtual 2021. Click here to learn more.
1. Lien, J. J., Kanade, T., Cohn, J. F., & Li, C. C. (1998, April). Automated facial expression recognition based on FACS action units. In Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition (pp. 390-395). IEEE.
2. Boersma, P. (1993, March). Accurate short-term analysis of the fundamental frequency and the harmonics-to-noise ratio of a sampled sound. In Proceedings of the institute of phonetic sciences (Vol. 17, No. 1193, pp. 97-110).