Abstract
Objective
Examine the associations between smartphone keystroke dynamics and cognitive functioning among persons with multiple sclerosis (MS).
Methods
Sixteen persons with MS with no self-reported upper extremity or typing difficulties and 10 healthy controls (HCs) completed six weeks of remote monitoring of their keystroke dynamics (i.e., how they typed on their smartphone keyboards). They also completed a comprehensive neuropsychological assessment and symptom ratings about fatigue, depression, and anxiety at baseline.
Results
A total of 1,335,787 keystrokes were collected, which were part of 30,968 typing sessions. The MS group typed slower (P < .001) and more variably (P = .032) than the HC group. Faster typing speed was associated with better performance on measures of processing speed (P = .016), attention (P = .022), and executive functioning (cognitive flexibility: P = .029; behavioral inhibition: P = .002; verbal fluency: P = .039), as well as less severe impact from fatigue (P < .001) and less severe anxiety symptoms (P = .007). Those with better cognitive functioning and less severe symptoms showed a stronger correlation between the use of backspace and autocorrection events (P < .001).
Conclusion
Typing speed may be sensitive to cognitive functions subserved by the frontal–subcortical brain circuits. Individuals with better cognitive functioning and less severe symptoms may be better at monitoring their typing errors. Keystroke dynamics have the potential to be used as an unobtrusive remote monitoring method for real-life cognitive functioning among persons with MS, which may improve the detection of relapses, evaluate treatment efficacy, and track disability progression.
Introduction
Multiple sclerosis (MS) is a progressive, demyelinating disease of the central nervous system. The most common disease course of MS is relapsing-remitting, which is characterized by episodes of relapses (i.e., sudden worsening of neurological functioning lasting for at least 24 h) followed by periods of remission (i.e., neurological functioning returning to baseline). However, even among persons with relapsing-remitting MS, levels of disability increase over time, and most individuals eventually convert to a progressive disease course. 1
Cognitive decline may indicate increased disease activity (e.g., clinical relapse) 2 and disability progression 3 among persons with MS. Therefore, routine monitoring of cognition is a vital component of comprehensive MS care. 4 Unfortunately, the current gold standard approach of evaluating cognitive functioning—neuropsychological assessment—is time-consuming (i.e., taking hours) and prone to practice effects (i.e., performing better on repeated cognitive assessments due to familiarity with the test materials), 5 which makes it impractical to be conducted more than once per year. Thus, there is a need to develop continuous monitoring methods that are sensitive to real-time changes in cognition (e.g., during a relapse), which may complement traditional neuropsychological exams and lead to more timely interventions.
Digital phenotyping is an emerging field that aims to characterize real-life internal state and behavior at the individual level through digital devices such as smartphones. Using digital phenotyping, researchers have been able to objectively quantify real-life behaviors and situations, such as social isolation (based on call and text logs, global positional system [GPS]), physical activity and sleep (through accelerometer, gyroscope, screen on/off status), and situational contexts (using cameras, GPS).6,7 Keystroke dynamics (i.e., how individuals type on smartphone keyboards) is a promising digital phenotyping approach and may be used to infer real-life cognitive functioning continuously and unobtrusively. Keystroke dynamics (e.g., slower typing speed, typing inconsistency) have been associated with motor functioning among individuals with Parkinson's8,9 and Huntington's 10 diseases (see recent meta-analysis 11 on the associations between keystroke dynamics and fine motor functions) and depressive and manic symptoms among individuals with bipolar disorder.12,13 In MS, keystroke dynamics have been linked with performances on motor and cognitive tests, brain lesions, and neurological disability.14,15
Although these studies show keystroke dynamics as a potential clinical tool for remote monitoring, there has not been a comprehensive investigation into the associations between keystroke dynamics and various cognitive domains. Existing keystroke studies in MS have only used a brief cognitive screener measuring processing speed.14,15 Studies in populations with cognitive impairment as a primary feature (e.g., mild cognitive impairment and dementia) focus on classification between clinical and control groups16,17; thus, associations between keystroke features and specific cognitive domains remain unknown. The goal of the current study was to provide proof-of-concept evidence that keystroke dynamics have the potential to be used as an unobtrusive remote monitoring method for real-life cognitive functioning. Specifically, the study explored differences in keystroke dynamics between persons with MS and healthy controls (HCs) and associations between keystroke dynamics and a range of cognitive and symptom measures.
Methods
Participants
Adult participants with MS and HCs were recruited from six parent studies with comprehensive neuropsychological assessment. Across these studies, exclusion criteria were any neurological or neurodevelopmental conditions other than MS, the presence of MS relapse symptoms within a month prior to study participation, and use of drugs and medications that may affect cognitive functioning. One of the studies was a cognitive rehabilitation study, so it only included individuals with MS who scored below cutoff on an episodic memory measure. An additional exclusion criterion for the current study was severe upper extremity motor difficulties that interfered with typing. Initially, only individuals who owned an iPhone were recruited because the study app (BiAffect) was only available for iOS devices. Subsequently, an Android version of the app was developed, and individuals with Android devices were also recruited.
Procedures
The study was approved by the Kessler Foundation Institutional Review Board. Since the current investigation used baseline neuropsychological test data that were collected for the parent studies, participants only needed to complete six weeks of remote keystroke monitoring for the current study. Informed consent was obtained either verbally by phone or via an online survey using Research Electronic Data Capture (REDCap) tools.18,19 Participants were instructed on downloading and using the BiAffect app on their personal smartphone. For iPhone users, the app was available for download from the Apple app store. For Android users, they were emailed an apk file, which they used to install the app on their phones. There were no active tasks that participants needed to complete; participants were told to simply type on their phones as they would normally type for six weeks. Using participants’ personal smartphones, instead of study phones, ensured that we were capturing their typical typing patterns. To protect their privacy, participants were told that the app only tracked the metadata associated with their typing (such as typing speed and errors) and not the content of what they typed (e.g., words, sentences). They were also informed that the app would not access any of their personal information on their phones. Keystroke data were monitored weekly by study personnel. If there were missing data for several days, study personnel would call the participant to troubleshoot any technical issues, and data collection time would be lengthened to make up for the missing days with the permission of the participant.
Smartphone app
Keystroke dynamics was monitored using the BiAffect app. For more details on the app, refer to previous publications on the current20,21 and previous pilot12,13,22 versions of the app. In brief, BiAffect consists of a custom virtual keyboard that temporarily replaces the native smartphone keyboard. The app records each keystroke, with information on keypress type (e.g., alphanumeric, backspace, autocorrection), timestamp, relative distance between two consecutive keystrokes, distance between the keystroke and the center of the keyboard, and accelerometer. Keystrokes were organized into typing sessions. Operationally, one typing session was initiated when the keyboard was activated and terminated after 6 s of inactivity, or at the time of keyboard deactivation. The 6-s threshold was implemented to avoid collecting very long sessions with just the accelerometer but no keypresses, which could drain the participant's phone battery. Latency between consecutive keystrokes was calculated as interkey delay, and median interkey delay within each session was used to represent typing speed for that session (median instead of mean was used because interkey delay was positively skewed). We only included transitions between alphanumeric characters when computing session-level median interkey delay to accurately characterize typical typing speed (latencies may be larger when using backspace, punctuations, and special characters). For typing speed variability within the session, we calculated the median absolute deviation (MAD) of interkey delay. We also derived rates of autocorrection events and use of backspaces within each session to signify the frequency of typing errors and self-monitoring of typing errors, respectively. See Figure 1 for an illustration of keystroke metrics.

Keystroke metrics. Panel A illustrates the frequency distribution of interkey delay (latency between consecutive keypresses) during transitions between alphanumeric characters in a sample typing session, with the blue dotted line denoting the median. Given the skewness of interkey delay, the median value is used to represent typing speed for each typing session. Median absolute deviation is used to represent variability in typing speed. Panel B illustrates rates of autocorrection and backspace in a sample typing session. In multilevel models, rates of autocorrection and backspace are represented with the total number of autocorrection events and backspaces used as the outcomes and offset by a log-transformed character count for each typing session.
In addition, we evaluated whether each typing session was conducted using two hands or one hand only based on our previously validated algorithm. 20 To determine this, interkey distance (distance from one keystroke to the next on the keyboard) was regressed on interkey delay. If the coefficient was positive and statistically significant (p < 0.05), the typing session was labeled as one-handed. The intuition was that when typing with one hand, it would take longer to transition between keystrokes that were further away from each other on the keyboard. The rest of the sessions were labeled as two-handed because it would be quicker to press keys on opposite ends of the keyboard with two hands. To ensure stable estimates, only typing sessions with at least 20 keystrokes were used to derive the handedness measure.
Neuropsychological and symptom rating measures
Since we used neuropsychological test and symptom rating data from six parent studies, the studies varied slightly in the measures used. Cognitive domains assessed by neuropsychological tests included processing speed (Symbol Digit Modalities Test [SDMT] 23 ), attention and working memory (digit span24,25), cognitive flexibility (Trail-Making Test [TMT] 26 ), behavioral inhibition (Delis-Kaplan Executive Function System [D-KEFS] 27 Color-Word Interference Test [CWIT]), verbal fluency (Controlled Oral Word Association Test 28 /animal fluency 29 or D-KEFS 27 Verbal Fluency Test), and episodic memory (California Verbal Learning Test-II, 30 Rey Auditory Verbal Learning Test, 31 or Hopkins Verbal Learning Test-Revised 32 ). Symptom rating scales included the Modified Fatigue Impact Scale (MFIS), 33 Chicago Multiscale Depression Inventory (CMDI), 34 and State-Trait Anxiety Inventory (STAI). 35 In order to compare different measures for each domain across studies, demographically adjusted z scores were calculated based on published norms. To reduce the total number of comparisons, z-scores for different parts of the same test were averaged into a composite (e.g., immediate and delayed recall of an episodic memory measure); we verified that findings were similar when individual components were run in separate models. Supplemental Table 1 lists the measures used in each of the six parent studies. Supplemental Table 2 provides additional details on each measure, including test versions and data transformations.
Statistical analysis
All data analyses were conducted in R 36 version 4.1.3. Group differences (MS vs. HC) in demographics were evaluated using Welch's t test for continuous variables and Pearson's chi-squared test for categorical variables (or Fisher's exact tests for count data for comparisons with expected frequency of less than five). Multilevel models (MLMs) were used to model keystroke dynamics within typing sessions (level 1) which were nested within subjects (level 2). Supplemental Table 3 details all MLM model specifications. Because median interkey delay yielded non-normal residuals, we used a Gamma distribution as it is more suitable for nonzero, positively skewed reaction time-type data. 37 A Gaussian distribution was used for interkey delay MAD because it contained zero values, so we were unable to use Gamma distribution. For rates of autocorrection events and use of backspaces, we performed Poisson generalized MLMs, with the count of autocorrection/backspace events per session as dependent variables and offset with the log-transformed character count per session (in order to signify proportions).
Group differences in keystroke dynamics were determined using a level-2 diagnostic group fixed effect. Associations between keystroke dynamics and neuropsychological test and symptom rating scores across both MS and HC groups were assessed with level 2 neuropsychological test/symptom scores as fixed effects (each score in separate model). Since the number of autocorrection events was dependent on typing speed (faster speed, more autocorrection events), we examined interactions between the main predictor (group, neuropsychological test, or symptom rating) and median interkey delay in predicting autocorrection events. To elucidate how individuals monitored their typing errors, we also tested the interaction between rate of autocorrection events and each main predictor (group, neuropsychological test, or symptom rating) in predicting number of backspaces used. All MLMs included a random intercept for the subject and adjusted for age, sex, years of formal education, operating system, and use of one or two hands as fixed effects when feasible (some covariates were removed for some models due to failed model convergence). Since a portion of our subjects (n = 6) underwent remote instead of in-person neuropsychological assessment, we examined whether remote test administration moderated these associations (see Supplemental Table 4).
Results
Thirty-three participants were recruited. Seven dropped out prematurely due to difficulties with using the BiAffect keyboard. Specifically, they noted that the autocorrection algorithm of the BiAffect keyboard was less accurate than the native smartphone keyboard. This resulted in a final sample of 16 persons with MS and 10 HCs. There were no demographic differences between those who dropped out and those who remained in the study (P > .05). A total of 1,335,787 keystrokes were collected, which were part of 30,968 typing sessions. Mean number of days with recorded keystrokes was 37 days (no significant differences between groups, P > .05). The ICCs were 65%, 33%, and 37% for median interkey delay, rate of autocorrection event, and rate of backspace used, respectively, which justified the use of MLMs. Lower median interkey delay (faster typing speed) was associated with more autocorrection events (estimate = −1.34, P < .001) and greater use of backspace (−0.12, P < .01). More autocorrection events were associated with less frequent backspace used (estimate = −0.96, P < .001).
Differences between MS and HC groups
Table 1 summarizes demographic, clinical, and keystroke dynamics for MS and HC groups. There were no significant demographic differences between groups (P > .05). A higher proportion of typing sessions were performed with one hand (as opposed to two hands) in the MS group, compared to the HC group (38% in MS vs. 19% in HC; χ2(1) = 542.70, P < .001). Adjusting for demographics, operating system, and typing with one or two hands, the MS group typed slower (longer median interkey delay; estimate = −1.45, P < .001) and more variably (higher interkey delay MAD; estimate = 0.11, P = .032) within sessions compared to the HC group (see Figure 2). The HC group had more autocorrection events than the MS group, but this was evident only among fast typers (group × median interkey delay interaction estimate = 1.82, P < .001; see Figure 2). The negative association between autocorrection and backspace rates was steeper for HC than MS group (group × rate of autocorrection events interaction estimate = 1.02, P = .002; see Figure 2).

Group differences in keystroke dynamics. Panel A displays differences between MS and HC groups in median interkey delay (typing speed); the MS group typed significantly slower. Panel B displays moderating effect of median interkey delay on number of autocorrection events (offset with number of characters per session); interkey delay is split into lower, median, and upper quartiles for visualization. The HC group had more autocorrection events than the MS group, but only among those who typed faster. Panel C displays the moderating effect of diagnostic group on the relationship between backspace and autocorrection rates; the relationship was steeper among HC than MS. Plots represent predicted values from multilevel models, and error bars/bands represent 95% confidence intervals. MS: multiple sclerosis; HC: healthy control.
Summary of demographic, clinical, and keystroke dynamics.
Note. Welch's two-sample t test and Pearson's chi-squared test (or Fisher's exact tests for count data for comparisons with expected frequency <5) were used to examine group differences in demographic variables. MLM with keystroke sessions nested within persons were used to examine group differences in keystroke dynamics, adjusted for age, sex, years of formal education, operating system, and use of one or two hands (see Supplemental Table 3 for more details).
MS: multiple sclerosis; HC: healthy control; SD: standard deviation; MLM: generalized multilevel model.
*Statistically significant difference between MS and HC groups.
Associations between keystroke dynamics and baseline neuropsychological test performance and symptom ratings
Table 2 summarizes associations between median interkey delay per session and baseline neuropsychological test performance and symptom rating. Models were adjusted for demographics, operating system, and typing with one or two hands. Higher median interkey delay (slower typing speed) was associated with worse performance on the SDMT (estimate = .34, P = .016), digit span (estimate = .62, P = .022), TMT composite (estimate = .35, P = .029), CWIT (estimate = 1.27, P = .002), and verbal fluency composite (estimate = .46, P = .039) scores (see Figure 3). Median interkey delay was not significantly associated with the episodic memory composite score (P > .05). Regarding symptom ratings, higher median interkey delay was associated with more severe impact from fatigue (MFIS; estimate = −1.87, P < .001) and more severe anxiety symptoms (STAI composite; estimate = −.74, P = .007). There was also a trend of higher median interkey delay correlating with more severe depressive symptoms (CMDI; estimate = −.33, P = .090). Sensitivity analyses did not reveal moderating effects from remote neuropsychological test administration (P > .05; see Supplemental Table 4).

Associations between median interkey delay and neuropsychological performance/symptom rating. Plots represent predicted values from multilevel models, and error bands represent 95% confidence intervals. Faster typing speed was associated with better performance on neuropsychological tests (e.g., SDMT and verbal fluency in Panels A and B and less severe symptom ratings (e.g., MFIS in Panel C). SDMT: Symbol Digit Modalities Test; MFIS: Modified Fatigue Impact Scale.
Associations between median interkey delay (typing speed) per session and baseline neuropsychological test performance and symptom ratings.
Note. Estimates were from generalized multilevel models with Gamma distributions. Analyses were adjusted for age, sex, years of formal education, operating system, and use of one or two hands (see Supplemental Table 3 for more details). All neuropsychological and symptom scores are standardized z scores based on demographics (see Supplemental Table 2 for more information).
SDMT: Symbol Digit Modalities Test; TMT: Trail-Making Test; CWIT: Color-Word Interference; MFIS: Modified Fatigue Impact Scale; CMDI: Chicago Multiscale Depression Inventory; STAI: State-Trait Anxiety Inventory.
*Statistically significant associations.
There was a significant moderating effect of median interkey delay on the associations between rates of autocorrection and neuropsychological test performance/symptom rating (see Table 3 and Figure 4). Among fast typers (lower median interkey delay), more autocorrection events (more tying errors) were associated with better performance on the SDMT (interaction estimate = −.34, P < .001), digit span (interaction estimate = −.62, P < .001), TMT composite (interaction estimate = −.75, P < .001), CWIT (interaction estimate = −1.60, P < .001), and verbal fluency composite (interaction estimate = −1.07, P < .001) scores. Fewer autocorrection events were associated with better performance on the episodic memory composite score, and this relationship was stronger among those who typed slower (interaction estimate = −.53, P = .002). More autocorrection events were associated with more severe depressive symptoms (CMDI; interaction estimate = −.60, P < .001) and less severe anxiety symptoms (STAI composite; interaction estimate = .40, P < .001) among fast typers. There was no significant moderating effect of median interkey delay on the association between autocorrection and MFIS, but there was a significant main effect such that more autocorrection events were associated with less profound impact from fatigue (estimate = −.64, P = .030).

Moderating effects of median interkey delay on associations between autocorrection events and neuropsychological performance/symptom rating. To visualize moderating effects, interkey delay is split into lower, median, and upper quartiles. Plots represent predicted values from multilevel models, and error bands represent 95% confidence intervals. Better neuropsychological performance was associated with more autocorrection events among fast typers only (Panels A and B). More severe depressive symptoms were associated with more autocorrection events (Panel C), and more severe anxiety symptoms were associated with fewer autocorrection events (Panel D); these relationships were only found among fast typers. SDMT: Symbol Digit Modalities Test; CMDI: Chicago Multiscale Depression Inventory; STAI: State-Trait Anxiety Inventory.
Moderating effects of median interkey delay on associations between autocorrection events per session and baseline neuropsychological test performance and symptom ratings.
Note. Models were generalized multilevel models with keystroke sessions nested within persons and Poisson distributions. Estimates were from interactions between neuropsychological test performance/symptom rating and median interkey delay, predicting autocorrection events per session. Models were offset with the log-transformed number of characters per session and adjusted for age, sex, years of formal education, operating system, and use of one or two hands (see Supplemental Table 3 for more details). All neuropsychological and symptom scores are standardized z scores based on demographics (see Supplemental Table 2 for more information).
SDMT: Symbol Digit Modalities Test; TMT: Trail-Making Test; CWIT: Color-Word Interference; MFIS: Modified Fatigue Impact Scale; CMDI: Chicago Multiscale Depression Inventory; STAI: State-Trait Anxiety Inventory.
*Statistically significant interactions.
There was a significant moderating effect of neuropsychological performance/symptom rating on the association between rate of autocorrection events and use of backspace (see Table 4). Those with better cognitive functioning and less severe symptoms showed a stronger negative association between autocorrection and backspace frequency (see Figure 5). This was evident for a range of neuropsychological tests and symptom inventories including the SDMT (interaction estimate = −.32, P < .001), TMT composite (interaction estimate = −.70, P < .001), CWIT (interaction estimate = −.93, P = .003), verbal fluency composite (interaction estimate = −.64, P < .001), MFIS (interaction estimate = 2.39, P < .001), CMDI (interaction estimate = .33, P < .001), and STAI composite (interaction estimate = .80, P < .001).

Moderating effects of neuropsychological performance/symptom rating on associations between autocorrection events and backspace use. To visualize moderating effects, each neuropsychological test, and symptom inventory is split into the mean and one standard deviation above and below the mean. Plots represent predicted values from multilevel models, and error bands represent 95% confidence intervals. Those with better cognitive functioning (Panels A and B) or less severe symptom ratings (Panels C and D) exhibited a stronger negative correlation between backspace and autocorrection frequencies. SDMT: Symbol Digit Modalities Test; CWI: Color-Word Interference; MFIS: Modified Fatigue Impact Scale; STAI: State-Trait Anxiety Inventory.
Moderating effects of baseline neuropsychological test performance/symptom ratings on associations between autocorrection events and use of backspace per session.
Note. Models were generalized multilevel models with keystroke sessions nested within persons and Poisson distributions. Estimates were from interactions between neuropsychological test performance/symptom rating and rate of autocorrection events per session, predicting backspace events per session. Models were offset with the log-transformed number of characters per session and adjusted for age, sex, years of formal education, operating system, and use of one or two hands (see Supplemental Table 3 for more details). All neuropsychological and symptom scores are standardized z scores based on demographics (see Supplemental Table 2 for more information).
SDMT: Symbol Digit Modalities Test; TMT: Trail-Making Test; CWIT: Color-Word Interference; MFIS: Modified Fatigue Impact Scale; CMDI: Chicago Multiscale Depression Inventory; STAI: State-Trait Anxiety Inventory.
*Statistically significant interactions.
Discussion
The current study examined the associations between smartphone keystroke dynamics and cognitive performance/symptom rating. The MS group typed slower than the HC group. Faster typing speed was associated with better performance on measures of processing speed, attention, and executive functioning (i.e., higher-order cognitive abilities such as cognitive flexibility, behavioral inhibition, and verbal fluency as used in the current study), as well as less severe impact from fatigue and less severe anxiety symptoms (and a statistical trend with less severe depressive symptoms). Our results are consistent with and expanded findings from previous MS studies,14,15 which showed significant associations between typing speed and performance on the SDMT. In terms of typing accuracy, there was a speed/accuracy trade-off, such that individuals with faster typing speed had more typing errors (autocorrection). Thus, because the HC group on average as well as those with better cognitive functioning and less severe symptoms typed faster, they also had more typing errors. Finally, more frequent use of backspaces was related to fewer typing errors, which confirmed our conceptualization that frequency of backspaces can be used as a measure of self-monitoring of typing errors. This relationship was stronger in the HC group on average as well as those with better cognitive functioning and less severe symptoms, suggesting that these individuals were better at monitoring their errors compared to their counterparts.
Episodic memory was the only cognitive construct that was not associated with typing speed. This was not surprising given different brain regions underlying memory compared to other cognitive domains assessed in this study. Memory consolidation is dependent on hippocampal functions, 38 whereas processing speed, attention, and executive functioning are reliant on frontal–subcortical brain circuits.39,40 Therefore, it appears that typing speed may be more indicative of frontal–subcortical brain functions. Interestingly, we found that individuals who typed slower showed a negative relationship between episodic memory performance and autocorrection events. It is possible that for individuals who did not favor speed over accuracy, intact memory function was critical for making fewer typing errors. More research needs to be conducted to fully delineate the relationships among typing speed, typing errors, and cognitive functioning.
Based on our findings, keystroke dynamics have the potential to be used as an unobtrusive, remote monitoring method for real-life cognitive functioning. The high dimensional nature of keystroke data may be more sensitive to subtle changes in cognition and enable longitudinal comparisons to the individual's baseline. 41 Because it is a direct examination of a real-life behavior (i.e., typing), keystroke dynamics may be more ecologically valid than traditional lab-/clinic-based neuropsychological tests.42,43 Remote cognitive monitoring via keystroke dynamics may lead to more timely identification of relapses and thus interventions. The continuous nature of keystroke monitoring may enable real-time assessment of treatment efficacy, allowing clinicians to modify medication regimen without requiring the patient to come in for an in-person visit. This will be especially helpful for physically disabled persons with MS, who often have difficulties arranging transportation and attending in-person medical appointments. Finally, the unobtrusive nature of keystroke monitoring may be less prone to bias and is conducive to long-term monitoring with limited patient burden, which may be used to document the progression of cognitive disability. Cognitive decline over time is a major predictor of reduced/loss of employment and receipt of disability benefits among persons with MS. 44
There are several limitations that require discussion. Firstly, given this was a proof-of-concept study, the sample size was small. Secondly, while we did exclude individuals with self-reported upper extremity or typing difficulties, we did not include objective measures of upper extremity motor skills due to the unavailability of such data in the parent studies. Thirdly, we only examined cross-sectional associations between keystroke dynamics and baseline neuropsychological test scores. We are currently conducting a larger trial with keystroke monitoring combined with real-time assessments of cognitive and motor performances (delivered through a smartphone) as well as real-time symptom ratings using ecological momentary assessment, which will allow us to establish real-time correspondence between keystroke dynamics and cognitive/motor performance and symptom ratings. Fourthly, because we used neuropsychological data from six parent studies with different test batteries, we had to transform the data to make them comparable, which could have decreased their sensitivity. Finally, although the BiAffect app keyboard looked very similar to the native smartphone keyboards, some participants reported that it was more difficult to use (e.g., the autocorrection algorithm was not as robust).
Conclusions
The current study provided proof-of-concept evidence that keystroke dynamics have the potential to be used as an unobtrusive remote monitoring method for real-life cognitive functioning, which may improve the detection of MS relapses, evaluate treatment efficacy, and track disability progression among persons with MS.
Supplemental Material
sj-docx-1-dhj-10.1177_20552076221143234 - Supplemental material for Associations between smartphone keystroke dynamics and cognition in MS
Supplemental material, sj-docx-1-dhj-10.1177_20552076221143234 for Associations between smartphone keystroke dynamics and cognition in MS by Michelle H Chen, Alex Leow, Mindy K Ross, John DeLuca, Nancy Chiaravalloti, Silvana L Costa, Helen M Genova, Erica Weber, Faraz Hussain and Alexander P Demos in Digital Health
Footnotes
Acknowledgements
The authors thank Dr Joshua Sandry (Montclair State University, NJ) for helping with recruitment and providing neuropsychological test data for this study.
Contributorship
MHC designed, obtained funding, analyzed data, and wrote the full draft of the manuscript. AL headed the team that created the study app and managed app data, obtained funding, and provided feedback on data analysis and manuscript draft. MKR and FH managed and processed the app data. DL and APD provided feedback for the analysis and manuscript draft. NC, SLC, HMG, and EW obtained funding, provided neuropsychological test data, and provided feedback for the manuscript draft.
Declaration of Conflicting Interests
The author(s) declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: Alex Leow is a cofounder of KeyWise AI, has served as a consultant for Otsuka US, and is currently on the medical board of Buoy Health. These entities had no involvement in the collection, analysis and interpretation of data; writing of the report; or in the decision to submit the article for publication.
Ethical approval
The study was approved by the Kessler Foundation Institutional Review Board.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This study was funded by Robert Wood Johnson Foundation (a New Venture Fund/MoodChallenge for ResearchKit); National Multiple Sclerosis Society (MB-1606-08779; RG-1802-30111; RG-1901-33304; CA 1069-A-7); National Institutes of Health/Eunice Kennedy Shriver National Institute of Child Health and Human Development (1R01HD095915); National Academy of Neuropsychology; and New Jersey Commission on Spinal Cord Research (CSCR16ERG018). The funding sources had no involvement in the collection, analysis and interpretation of data; writing of the report; or in the decision to submit the article for publication.
Guarantor
MHC.
Supplemental Material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
