Objective wearable measures correlate with self
npj Digital Medicine volume 6, Article number: 146 (2023) Cite this article
1107 Accesses
1 Altmetric
Metrics details
Spinal Cord Stimulation (SCS) is a well-established therapy for treating chronic pain. However, perceived treatment response to SCS therapy may vary among people with chronic pain due to diverse needs and backgrounds. Patient Reported Outcomes (PROs) from standard survey questions do not provide the full picture of what has happened to a patient since their last visit, and digital PROs require patients to visit an app or otherwise regularly engage with software. This study aims to assess the feasibility of using digital biomarkers collected from wearables during SCS treatment to predict pain and PRO outcomes. Twenty participants with chronic pain were recruited and implanted with SCS. During the six months of the study, activity and physiological metrics were collected and data from 15 participants was used to develop a machine learning pipeline to objectively predict pain levels and categories of PRO measures. The model reached an accuracy of 0.768 ± 0.012 in predicting the pain intensity of mild, moderate, and severe. Feature importance analysis showed that digital biomarkers from the smartwatch such as heart rate, heart rate variability, step count, and stand time can contribute to modeling different aspects of pain. The results of the study suggest that wearable biomarkers can be used to predict therapy outcomes in people with chronic pain, enabling continuous, real-time monitoring of patients during the use of implanted therapies.
Chronic pain is a debilitating condition affecting a widespread population in the United States, estimated at over 50 million American adults1. Pain that lasts for more than three to six months is often considered chronic and is influenced by a complex combination of biopsychosocial factors including but not limited to emotional, psychological, physical, and social considerations2,3. Spinal cord stimulation (SCS) is an effective treatment option for chronic pain and often leads to pain reduction and improvement in quality of life4,5,6,7. Response to therapy over time varies from person to person and often requires interactive adjustment of therapy parameters8. Thus, appropriate individual selection and long-term monitoring are crucial in optimizing the outcomes of SCS therapy9,10.
The condition of a person with chronic pain is usually evaluated through several patient-reported outcome (PRO) measures administrated manually in an in-clinic visit. The current gold standard for evaluating pain is using unidimensional PROs such as the Numerical Rating Scale (NRS) or the Visual Analog Scale (VAS)11,12. Efforts to capture multidimensional aspects of chronic pain and treatment effects have historically been done through the addition of other validated PRO questionnaires13,14. To capture the more comprehensive and multidimensional effects of pain, people with chronic pain often answer several other questionnaires, such as the Pain Catastrophizing Scale (PCS)15, which provides insight into the psychological aspects of pain, Oswestry Disability Index (ODI) for disability and function16, and Patient-Reported Outcomes Measurement Information System 29 (PROMIS-29) for global health measure17, Patient Health Questionnaire-9 (PHQ-9) for depression18, and Patient Global Impression of Change (PGIC)19 for the perception of improvement with different therapies. These tools rely on the person’s assessment, which is subject to memory, cognitive, social desirability, and other psychologically influenced response biases. Additionally, there are limitations and the potential for subjective bias on the part of clinical evaluators20. An individual’s perception of pain and its effect on daily activities and overall health are hard to capture in a single data point recorded in a clinical visit20. However, frequent collection of multiple questionnaires at shorter interval visits is burdensome for people living with chronic pain (and their clinicians), as it requires up to 54 questions across instruments.
To date, there are no established and validated objective measures for assessing pain and its impact on the person’s overall well-being, and objectively quantifying the effect of SCS treatment on reducing chronic pain. Prior research has emphasized the need for improved metrics to better characterize an individual’s response and change in chronic pain levels with neurostimulation therapies21,22. Recent advances in the development of wearable technologies enabling objective measurement of movement, physical activity, and function23,24,25,26,27,28,29, gait and posture30,31,32,33,34, neuromuscular and physiological data10,30,31,32,33, sleep35,36, and behavioral assesment37,38 have resulted in the emergence of “digital biomarkers” which could be measured outside the physical confines of the clinics avoiding some of the bias introduced in clinic measurements30,31,32,33,39,40. Many of these biomarkers from wearables have shown a potential to objectively measure different aspects of an individual’s chronic pain and its effect on physical activity, sleep, psychological health, and social participation27,28,31,35,41.
Machine learning (ML) has been extensively used in healthcare to provide insights, enhance decision-making, improve patient outcomes, automate workflows, accelerate medical research, and enhance operational efficiency by analyzing large amounts of data42,43,44. Despite recent research highlighting the importance of using machine learning techniques in pain research, previous works have focused on correlating and monitoring symptoms and side effects of pain with digital biomarkers and not necessarily predicting the subject-reported outcomes45,46,47.
Recent advances in wearable technologies and machine learning algorithms provide a promising opportunity for predicting pain and other subjective measurements of pain48,49. Previous studies have emphasized machine learning-based classification of pain intensity, but far less attention has been paid to predicting pain and its multi-dimensional effects that are usually captured with multiple patient-reported outcome questionnaires49.
Here, we combined objective measures collected from a custom smartwatch application with predictive machine learning algorithms to predict commonly used PROs to measure chronic pain perception. While these objective assessments are not direct measurements of pain, they have the potential to serve as a highly accurate tool to evaluate changes in the quality of life and the level of disability in people with chronic pain and to develop prediction models to measure response to SCS. The goal of this study was to predict pain perception with machine learning models as measured by PROs from objective data measured before and after the SCS implant collected from commercially available smartwatches.
Twenty participants were enrolled as part of this study. Five patients discontinued participation in the study: one participant withdrew consent prior to permanent SCS system implantation, two participants withdrew consent after permanent implantation, one participant’s participation was terminated by the investigator, and one participant was excluded from analysis due to a lack of wearable data. During the study, the median compliance percentage for completing the PROs through the custom application on the iPhone was 88.8% (Interquartile range of 66.6% to 100%). The participants had to input PROs in a separate research phone and were deemed compliant if they completed PROs at least three times during the baseline and once every month after the implant for a duration of 6 months. In addition, a median compliance of 84.7% (Interquartile range of 70.4 to 95.4%) was achieved for using the smartwatch during the study duration. Participants were deemed compliant if they wore the watch for at least 7 days during the baseline and 180 non-consecutive days after the implant. No significant difference was found in compliance between completing PROs on the iPhone and wearing the watch (Wilcoxon rank-sum test). Participants received follow-up phone calls from clinic staff if they missed providing data for more than 3 consecutive days or completing PROs on the custom application. The average age of participants was 52.25 (±9.7) years at baseline. On average, all participants suffered from 12 years of chronic pain. Back pain was the primary pain diagnosis of the majority of the participants (85%) in the presented cohort. Table 1 summarizes the baseline characteristics of the study participants.
During the course of the treatment, participants showed improvements in NRS and all other PROs collected through the wearable application and in-clinic visits (for comparison of the baseline visit to the last pre-op in-clinic visit). A summary of all PROs at different time points and a comparison of in-clinic and application data is presented in Table 2. The average and standard deviation of in-clinic NRS values dropped from 7.2 (±0.88) at baseline to 3.14 (±1.83), and 3.34 (±2.12) at the 3- and 6-month visits, respectively. Moreover, the average daily NRS collected from the custom watch application was reduced across all participants (Fig. 1a). Similar improvements were seen for all PROs included in this study (Table 2). Figure 1b shows improvement seen in PGIC across all participants based on monthly reported values on the custom iPhone application. The raw scores for the PROs collected using the wearable application and in-clinic visits at baseline, 3-months, and 6-months show significant improvement in PROMIS-29’s sleep disturbance, social roles, pain interference, and fatigue compared to baseline (Table 2). The raw scores for the PCS and ODI showed significant improvements compared to baseline as well. The improved trend for the values collected through the digital health custom wearable application is like the improvement trend in the values collected during clinic visits. The sample size for each comparison is listed for each measurement. All participants must complete PROs in the clinic as part of the required case report forms for the study.
a The weekly average of NRS scores across all participants during the study; 0 indicates the permanent implant date; the baseline period values (−1 to 0 month) are the weekly average of NRS pulled from the watch application; statistical testing indicates a comparison to pre-implant pain scores per subject and is calculated using the Wilcoxon signed-rank test on median values; *p ≤ 0.05, **p ≤ 0.001. b Monthly average of PGIC scores for all participants during the study; NRS numerical rating scale, PGIC patient global impression of change, CI confidence intervals.
The physiological and behavioral features collected passively throughout the study were used to construct a machine learning model to predict daily categorical pain levels in the participants. A variety of different machine learning models were attempted for predicting three categorical levels of pain intensity, with the random forest model yielding the best predictive performance. Specifically, the random forest model showed high accuracy in predicting three intensity levels of mild, moderate, and severe pain, corresponding to NRS levels of <4, ≥4 & ≤6, and >6, respectively (F1 Score = 0.768 ± 0.012 and Accuracy = 0.768 ± 0.012, Sensitivity = 0.737 ± 0.016, and Specificity = 0.869 ± 0.007) (Table 3, Supplementary Fig. 1). This model was driven using objective features as an input and could theoretically be used to passively monitor daily pain intensity categories in people with chronic pain.
To holistically predict the well-being of people with chronic pain, we used the objective measures collected from the custom watch app to predict categories of common PROs for chronic pain assessments. We developed 11 machine learning models to predict score categories from PROs collected throughout the study (PGIC, PCS total score, 7 domains of PROMIS-29, ODI total score, PHQ-9 total score) using objective features. The results of PRO modeling and evaluation metrics for each of the PROs are shown in Table 3. All models had high accuracy and F1 score indicating the application of digital biomarkers in predicting categories of these subjective measures for chronic pain. In addition to pain intensity, these models could be used to predict different dimensions of pain including emotional aspects using depression, anxiety, and fatigue (PROMIS-29, PHQ-9) models, physical function aspects using physical activity (PROMIS-29) model and sleep aspects using sleep disturbance (PROMIS-29) model. The PCS model could help with predicting pain catastrophizing in people with chronic pain while the PGIC model could help with predicting individual response to therapy and be used as a tool to find the right candidates for spinal cord stimulation. Most models achieved high sensitivity and specificity, but the datasets for PROMIS-29 domains of physical function and sleep disturbance were highly imbalanced. This was resolved by use of the synthetic minority oversampling technique (SMOTE). However, a remaining limitation was the limited number of data points in minority classes and a lack of sleep data from the smartwatch as an input for predicting sleep disturbance. This limitation likely explains the low specificity for physical function and low sensitivity for the sleep model (Table 3).
The feature importance for modeling three levels of pain intensity and SHAP (SHapley Additive exPlanations) values are plotted in Fig. 2. The figure includes the feature importance for the model built with objective features (a and b). Several wearable biomarkers such as heart rate, step count, and stand time were shown to be important predictors of pain and various aspects of it. The feature importance analysis of the model using objective biomarkers depicts those physiological biomarkers from the Apple® Watch that played an important role in modeling three categorical levels of pain (e.g., heart rate, heart rate variability, step count, and stand time). The number of days pre/post implant was also found to be a prominent feature for predicting pain, likely due to the time it takes for SCS therapy to be adjusted to optimal settings post-implant. The device programming features that were pulled from the patient’s SCS controller app were also among the important features. Participants who reported more variation of pain score had a higher engagement with the patient controller app (i.e., they visited the patient controller application more often to adjust therapy settings).
a The average of feature importance for modeling pain levels using objective data across 10 runtimes. The standard deviation is shown with a line in the bar plot. b A sample of SHAP (SHapley Additive exPlanations) values for one run time. RMSSD: The root mean square of successive inter-beat intervals of heartbeats differences. SDANN: The standard deviation of the average inter-beat intervals without artifacts for each 5-minute interval over a 24-hour recording of heart rate variability. SDNNI: The average of the standard deviations of all inter-beat intervals without artifacts for each 5-minute interval over a 24-hour recording of heart rate variability.
Furthermore, there was a correlation between the median heart rate (as an important feature) and pain intensity demonstrating that on days with lower pain levels, participants experienced lower heart rate values during the day (Fig. 3)50.
The weekly average of heart rate median values across all participants during the study along with the weekly average of NRS scores; NRS numerical rating scale, CI confidence intervals.
The feature analysis for the other predictive models for other aspects of pain showed the same pattern as the predictive model for pain levels, and features such as heart rate, heart rate variability, step count, and stand time were identified as top predictors. The number of months post-implant also appeared as one of the top features in most of the PRO models, suggesting a gradual wash-in of the therapy effect over time. Moreover, these results from pain and PRO modeling indicated that heart rate variability can contribute to the understanding of pain. Further investigations showed that participants with lower pain intensity have a higher heart rate variability which is seen in people with chronic pain41,51. Finally, the total number of programming changes, the target stimulation amplitude (from programming data), and the local weather temperature and precipitation were among the top predictors in most of PRO models. This suggests that the programming setting and weather may influence the daily pain level reported by patients52,53,54.
SCS treatment is an effective therapy that can provide much-needed pain relief, improve physical activity and social participation, and ultimately enhance the quality of life for people dealing with chronic pain. Here we report, the results of this research which further support the efficacy of SCS by showing an improvement in average pain relief post-implant across participants accompanied by average improvement in the quality-of-life metrics including physical activity, social roles anxiety, fatigue, sleep disturbance, and pain interference. Objective measures to monitor each individual’s progress over the period of treatment could help with increasing therapeutic efficacy and better adoption of new technologies. Our approach in this paper is to highlight the feasibility of using objective data from a wearable device to not only create a model for predicting categorical pain intensity but also to predict other outcome quality-of-life outcome measures typically measured in clinics. These results suggest that a machine learning model can use passive data to predict and categorize participants based on NRS, PCS, PROMIS-29 domains, ODI, PHQ-9, and PGIC. This is an important and distinct improvement over just categorizing participants using unidimensional scales.
The smartwatch compliance that passively collected participant’s physiological signals highlights the importance of wearables in new technology adoption. The collection of PROs through the custom application and use of a smartwatch provides frequent data required for validating these predictive models. The objective data from wearables can be used to develop predictive algorithms for long-term passive monitoring of symptoms and reducing the burden of completing PROs in the future. It is noteworthy to mention that patients in this study are selected based on their comfort level with technology and willingness to engage in digital health activities. The study’s device compliance rates may overestimate real-world use because patients are actively monitored, and clinical staff call them if they miss completing PROs or providing data for more than three days.
There are distinguishing differences in the features extracted from the Apple® Watch during the baseline period compared to post-SCS treatment across all participants. These differences indicate physical and physiological changes in people with chronic pain which are measured using a wearable watch. For example, in the physical aspect of pain, the average total stand time increases after implant across all participants which suggests higher activity and social engagement in people with chronic pain treated with SCS. Additionally, the average daily stand time follows the NRS improvement in participants, suggesting that better pain relief leads to a higher average standing time throughout the day. The predictive models accurately predict three levels of daily pain and various aspects of pain captured by commonly used and validated PROs using objective data as inputs. The feature importance analysis reveals activity metrics, heart rate, and heart rate variability as important predictors of pain. The number of days pre/post implant is another crucial feature, indicating that participants take time to experience a lower pain level, and the change in pain intensity is not immediate. This suggests that participants experience varying levels of pain relief depending on the timing of their SCS implants until their pain levels reach a more stable state. Monitoring of patient data using passive means may help in the future by informing optimization of settings with a closed-loop SCS system and choosing the appropriate window to change the SCS configuration based on the pain level.
Chronic pain often affects other dimensions of an individual’s life such as sleep, physical function, psychological health, and quality of life. The study outcomes demonstrate improvement of different aspects of pain in people with chronic pain after spinal cord stimulation therapy and the potential use of wearables to capture these measures objectively since different pain domains such as physical function, social behavior, and sleep can be quantified through wearable sensing55,56,57. The predictive models of PROs developed in this study could be used to monitor an individual’s progress through the SCS continuum and decrease the burden of completing PROs in the app or the clinic. The predictive model for PGIC developed with high accuracy can be used for patient selection and to provide therapy to people with chronic pain for whom SCS is more effective.
The main strength of this work is developing predictive models to predict pain and other aspects of it using objective data. We develop a large set of biomarkers and build accurate and robust models that could be used to characterize pain and well-being in people with chronic pain. One limitation of the current study is the small sample size for developing machine learning models which can affect the generalizability of our predictions given the variability across different patients. To mitigate this, we randomize the training and testing data 10 times and report the average model performance. The short-term application of this work is a population model that can be personalized to each patient using some of their initial data. But in the longer term, with more data and a more diverse dataset, we may be able to generate population models that operate without personalization.
Another limitation of the study is the lack of reliable sleep data as an important predicting factor for pain, due to the inadequate time resolution and the fact that only binary values are provided for sleep data. Additionally, the amount of physical activity information is limited for this version of the watch compared to future generations and clinically validated sensors which could affect our ability to predict pain levels and categories of PRO measures. Moreover, the imbalanced datasets for PROMIS-29 physical function, and sleep disturbance with a few data points in the minority class limit us in building robust models with high specificity and sensitivity in predicting these two domains of PROMIS-29.
There are multiple confounding factors such as post-surgery recovery time, inactivity due to surgery, effects of medication, and other factors that can affect the interpretation of wearable features selected (e.g., heart rate, step count) for this study. Including daily averages and variations of the features as input to the model could decrease such effects. In addition, the pain modeling is performed using data up to 6 months post-implant to have a more robust prediction. The validity of all the features of the model needs to be further studied to better understand the nature of the relationship between the predictor and a patient’s pain level and to rule out potential confounding factors. This study is designed to demonstrate feasibility in a small patient population but future studies including a larger sample size of people with chronic pain are required to overcome these limitations. In addition, improvements in future generations of wearable devices can provide access to additional sensors and data; resulting in helping with the robustness of predictive models aimed at solving complex modeling of individual’s pain.
In summary, with increasing the availability of both consumer and research-grade wearable devices and accessing sophisticated machine learning techniques, the opportunity of developing novel methods to passively monitor the daily changes in people with chronic pain and predict their pain states becomes more achievable. The ability to identify patients passively with waning therapy, or other painful clinical events (such as a fall), would help bring them back in for evaluation and thereby drive improvement in their long-term outcomes. Wearables can objectively measure many features which are influenced by participants’ chronic pain such as activity, sleep, psychological health, and social participation. Adding objective measurements could improve the accuracy of classification models and enable us to move toward a more personalized therapy with a limited burden on both people with a chronic pain and clinicians.
Data were extracted as a sub-study of the prospective, multicenter, international REALITY (Long-Term Real-World Outcomes Study on Patients Implanted with a Neurostimulator) study (NCT03876054; March 15, 2019). Prior to initiating the study, Western Institutional Review Board approval was received for the study sites and all participants were provided with written informed consent. The study inclusion criteria for REALITY were designed with few restrictions on the pain indication as allowed by the regulatory bodies in each geographical region and according to standard clinical practice to replicate the range of complex participants that would be seen in everyday medical practice. The sub-study was designed to compare changes in pain and physical function, and behavioral markers before and 6 months after SCS implantation. The sub-study participants were asked to wear a smartwatch and answer multiple PROs frequently on a custom-designed application in addition to the regular in-clinic visits. Study visits occurred at enrollment, baseline, three- and six-month post-implantation.
Demographics were collected at baseline and included the duration of pain, work status, exercise level, and the number of previous surgeries. PRO measures to assess pain intensity (NRS), physical function and disability (ODI), emotional distress and depression (PCS), and global health (PROMIS-29 and PGIC) were collected at baseline and each follow-up study visit. All sub-study participants were provided with an Apple ® Watch (Series 3) at enrollment. Participants were prompted to enter NRS scores collected on the investigational custom watch application daily from baseline until six months after the implant. The watch application was given access to HealthKit data and passively collected several HealthKit metrics for activity, behavior, and cardiac measures such as heart rate, heart rate variability, step count, stand time, and distance walking/running. The subject needed to access the watch application at least once a day. Once participants selected their current pain level from 0 (no pain) to 10 (the worst pain imaginable) in the custom watch application, the app then sent the NRS data to a secured cloud storage. The watch application is an iOS-based application that pulls Healthkit data from the Apple Watch. As soon as a user rates the pain intensity on the UI, the app is activated, and physical activity and heart rate data are collected in the background. The UI screen shows the elapsed time from the start of the data collection to the subject. The REALITY iPhone custom application is a companion to the watch application and installation of the REALITY wearable application on the iPhone will automatically install the watch application on the paired Apple® Watch. The iPhone application is an iOS-based application to collect behavioral data from subjects. PROs such as PROMIS-29, ODI, PCS, and PHQ-9 were collected on a regular basis through the phone application (3 times before implant and once every month after the implant for a period of 6 months). PGIC was collected monthly for 6 months after the implant.
The normality of PROs data was assessed using the Shapiro-Wilk test. The two-sided Wilcoxon signed-rank test was used to measure the significance of the change in PROs pre- versus post-implant on median values. The two-sided Wilcoxon rank-sum test for the independent non-normal sample was used to measure the significance of compliance with completing PROs on the phone and the watch application. P-values less than 0.05 are considered as the significance level.
Apple® HealthKit provided features for step counts, stand time, walking/running distance, sleep, heart rate and heart rate variability, and the number of flights climbed. Post-processing of the data showed that there are a high number of missing values for sleep and flights climbed acquired through the HealthKit app. We removed these measures for further analysis in this manuscript. We used a threshold that was calculated based on the number of data points recorded each day to discard days with inadequate data points, or sparse data. Specifically, a threshold was determined by using 5% of the median value derived from the daily number of data points within the same pain level. We removed data from days that had fewer data points than the threshold. To balance feature weights and handle missing data for low-resolution features on the Apple® Watch, we used a daily window for data points with the same pain level in our analyses. Statistical features such as maximum, minimum, sum, mean, standard deviation, 25th, 75th, and 90th percentiles were extracted from the daily windowed data. Furthermore, since there were many missing data points in the heart rate variability (HRV) of the Apple® HealthKit data and the inter-beat interval was not accessible, heart rate was used to estimate this time interval in order to calculate HRV in three different methods, 1) the root mean square of successive inter-beat intervals of heartbeats differences (RMSSD), 2) the standard deviation of the average inter-beat intervals without artifacts (NN intervals) for every 5 min over a 24 h-period of HRV recording (SDANN), and 3) mean of the standard deviations of all the NN intervals for every 5 min over a 24 h-period of HRV recording (SDNNI)58.
Different data streams were used and aggregated for comprehensive analyses and to provide a deeper understanding of digital biomarkers contributing to participant’s therapy outcomes. The SCS device programming information was also pulled from the patient’s controller application (Abbott, Plano, TX). However, programming data from the patient controller application on the phone had missing values due to retention policies. Therefore, we imputed programming data using the minimum value for each programming feature. Additionally, we used publicly available datasets for weather information (https://www.ncei.noaa.gov/) based on the location of participants, moon phase (https://www.timeanddate.com/), and the stock market data of three popular stocks, NASDAQ Composite, DIJA, and S&P 500 (https://finance.yahoo.com/). The stock market data was also imputed when the market was closed, and the price on the last day was used to fill the missing prices. The features with high correlation (>0.90) were removed from the dataset and the remaining features were used for developing machine learning models.
The machine learning models were developed on Apple® HealthKit data, programming data, and other features discussed in the data collection and featurization section. Daily NRS values and weekly scores for each PRO were used as the main output variables for the predictive machine learning models.
Several ML techniques such as Logistic Regression59, Support Vector Machine60, K-Nearest Neighbors61, Random Forest62, and Catboost63 were implemented on the dataset for predicting daily pain values as well as other PROs and their performance was compared on the testing data, and random forest, a tree-based model, provided the best performance and interpretability of features (Supplementary Table 1). Additionally, SHAP (SHapley Additive exPlanations)64,65 technique was used to calculate the contribution of features. Of the total data available, 80% of the data was used for training and 20% for the testing phase.
The random forest model was developed using digital biomarkers collected from the Apple® Watch, programming data from the patient controller, and other features such as weather data based on the participant’s primary residence zip code. To increase the robustness of predictions among the training sets, the random forest model was trained 10 times using randomly selected 80% of the input data available. The reported outcomes were then averaged across all 10 different runs. Figure 4 shows the machine learning pipeline for predicting pain and other PRO measurements.
The patient-reported outcomes (PROs) of numerical rating scale (NRS), patient-reported outcomes measurement information system 29 (PROMIS-29), pain catastrophizing scale (PCS), patient health questionnaire-9 (PHQ-9), Oswestry disability index (ODI), patient global impression of change (PGIC) are outputs of machine learning models.
A balanced number of training sets for each class of output variable were considered and NRS and other PROs were grouped into different classes. We categorized NRS values into three groups: mild (NRS < 4), moderate (NRS ≥ 4 and NRS ≤ 6), and severe (NRS > 6) pain; PROMIS-29 into two groups, physical function, and social roles were grouped as responders (T-score ≥ 40) and non-responders (T-score < 40), depression anxiety, fatigue, sleep disturbance, and pain interference into two groups of responders (T-score ≤ 60), and non-responders (T-score > 60);14,66 PCS into two groups of catastrophizing (total score ≥ 30), and non-catastrophizing (total score < 30); PHQ-9 into two groups of responders (total score ≤ 9) and non-responders (total score > 9)18, ODI into three classes of high responders (≤ 20), low responders (> 20 & ≤ 40), and non-responders (> 40);67 PGIC, into two groups of responders (Moderately better, Better, and A great deal better) and non-responders (No change, Almost the same, A little better, and Somewhat better). The synthetic minority oversampling technique (SMOTE)68 was used to address imbalanced datasets for PROMIS-29 physical function and sleep disturbance for building their predictive models.
Further information on research design is available in the Nature Research Reporting Summary linked to this article.
Data used in this study may be made available to qualified individuals for collaboration if a written request is made to and granted in writing by Abbott at Abbott’s sole discretion. The requester should include their name, title, contact information, and the institution they work for as well as the specifics regarding the use and necessity of the requested dataset to the corresponding author. Abbott retains full discretion over its data and is under no obligation, legal or otherwise, to release or provide it to third parties regardless of the request being made.
The codes for data processing, modeling, and visualization were developed using open-source Python libraries including pandas version 1.5.0, scikit-learn version 1.0.2, scipy version 1.9.1, numpy version 1.21.5, and matplotlib version 3.5.3. The code may be made available to qualified individuals for collaboration if a written request is made to and granted in writing by Abbott at Abbott’s sole discretion. The requestor should include their name, title, contact information, and the institution they work for as well as the specifics regarding the use and necessity of requested codes to the corresponding author. Abbott retains full discretion over its code and is under no obligation, legal or otherwise, to release or provide it to third parties regardless of the request being made.
Steingrímsdóttir, Ó. A., Landmark, T., Macfarlane, G. J. & Nielsen, C. S. Defining chronic pain in epidemiological studies: a systematic review and meta-analysis. Pain 158, 2092–2107 (2017).
Article PubMed Google Scholar
Russo, C. M. & Brose, W. G. Chronic pain. Annu. Rev. Med. 49, 123 (1998).
Article CAS PubMed Google Scholar
Gatchel, R. J., Peng, Y. B., Peters, M. L., Fuchs, P. N. & Turk, D. C. The biopsychosocial approach to chronic pain: scientific advances and future directions. Psychol. Bull. 133, 581 (2007).
Article PubMed Google Scholar
Taylor, R. S., Van Buyten, J.-P. & Buchser, E. Spinal cord stimulation for complex regional pain syndrome: a systematic review of the clinical and cost-effectiveness literature and assessment of prognostic factors. Eur. J. Pain. 10, 91–101 (2006).
Article PubMed Google Scholar
Deer, T. et al. Ultra-Low Energy Cycled Burst Spinal Cord Stimulation Yields Robust Outcomes in Pain, Function, and Affective Domains: A Subanalysis From Two Prospective, Multicenter, International Clinical Trials. Neuromodulation Technol. Neural Interface 25, 137–144 (2021).
Article Google Scholar
Deer, T. R. et al. Dorsal root ganglion stimulation yielded higher treatment success rate for complex regional pain syndrome and causalgia at 3 and 12 months: a randomized comparative trial. Pain 158, 669–681 (2017).
Article PubMed Google Scholar
Kapural, L. et al. Treatment of nonsurgical refractory back pain with high-frequency spinal cord stimulation at 10 kHz: 12-month results of a pragmatic, multicenter, randomized controlled trial. J. Neurosurg. Spine 11, 1–12 (2022).
Google Scholar
Deer, T. R. et al. Dorsal root ganglion stimulation yielded higher treatment success rate for complex regional pain syndrome and causalgia at 3 and 12 months. PAIN 158, 669–681 (2017).
Article PubMed Google Scholar
Goudman, L. et al. Patient Selection for Spinal Cord Stimulation in Treatment of Pain: Sequential Decision-Making Model—A Narrative Review. J. pain. Res. 15, 1163 (2022).
Article PubMed PubMed Central Google Scholar
Goudman, L., Brouns, R., Linderoth, B. & Moens, M. Effects of spinal cord stimulation on heart rate variability in patients with failed back surgery syndrome: comparison between a 2-lead ECG and a wearable device. Neuromodulation Technol. Neural Interface 24, 512–519 (2021).
Article Google Scholar
Thong, I. S. K., Jensen, M. P., Miro, J. & Tan, G. The validity of pain intensity measures: what do the NRS, VAS, VRS, and FPS-R measure? Scand. J. Pain. 18, 99–107 (2018).
Article PubMed Google Scholar
Farrar, J. T., Young, J. P., LaMoreaux, L., Werth, J. L. & Poole, R. M. Clinical importance of changes in chronic pain intensity measured on an 11-point numerical pain rating scale. PAIN 94, 149–158 (2001).
Article PubMed Google Scholar
Dworkin, R. H. et al. Interpreting the clinical importance of treatment outcomes in chronic pain clinical trials: IMMPACT recommendations. J. Pain. Off. J. Am. Pain. Soc. 9, 105–121 (2008).
Article Google Scholar
Hays, R. D., Spritzer, K. L., Schalet, B. D. & Cella, D. PROMIS®-29 v2. 0 profile physical and mental health summary scores. Qual. Life Res. 27, 1885–1891 (2018).
Article PubMed PubMed Central Google Scholar
Sullivan, M. J. L., Lynch, M. E. & Clark, A. J. Dimensions of catastrophic thinking associated with pain experience and disability in patients with neuropathic pain conditions. Pain 113, 310–315 (2005).
Article PubMed Google Scholar
Fairbank, J. C., Couper, J., Davies, J. B. & O’Brien, J. P. The Oswestry low back pain disability questionnaire. Physiotherapy 66, 271–273 (1980).
CAS PubMed Google Scholar
Cella, D. et al. The Patient-Reported Outcomes Measurement Information System (PROMIS): progress of an NIH Roadmap cooperative group during its first two years. Med. Care 45, S3 (2007).
Article PubMed PubMed Central Google Scholar
Cannon, D. S. et al. The PHQ-9 as a brief assessment of lifetime major depression. Psychol. Assess. 19, 247 (2007).
Article PubMed Google Scholar
Geisser, M. E. et al. Contributions of change in clinical status parameters to Patient Global Impression of Change (PGIC) scores among persons with fibromyalgia treated with milnacipran. PAIN® 149, 373–378 (2010).
Article PubMed Google Scholar
Leroux, A., Rzasa-Lynn, R., Crainiceanu, C. & Sharma, T. Wearable devices: current status and opportunities in pain assessment and management. Digit. Biomark. 5, 89–102 (2021).
Article PubMed PubMed Central Google Scholar
Hagedorn, J. M. et al. Differences in calculated percentage improvement versus patient-reported percentage improvement in pain scores: a review of spinal cord stimulation trials. Reg. Anesth. Pain. Med. 46, 293–297 (2021).
Article PubMed Google Scholar
Pilitsis, J. G., Fahey, M., Custozzo, A., Chakravarthy, K. & Capobianco, R. Composite Score is a Better Reflection of Patient Response to Chronic Pain Therapy Compared With Pain Intensity Alone. Neuromodulation Technol. Neural Interface 24, 68–75 (2020).
Article Google Scholar
Caramia, C. et al. IMU-based classification of Parkinson’s disease from gait: A sensitivity analysis on sensor location and feature selection. IEEE J. Biomed. health Inform. 22, 1765–1774 (2018).
Article PubMed Google Scholar
Maceira-Elvira, P., Popa, T., Schmid, A.-C. & Hummel, F. C. Wearable technology in stroke rehabilitation: towards improved diagnosis and treatment of upper-limb motor impairment. J. Neuroeng. Rehabil. 16, 1–18 (2019).
Article Google Scholar
Yan, X., Li, H., Li, A. R. & Zhang, H. Wearable IMU-based real-time motion warning system for construction workers’ musculoskeletal disorders prevention. Autom. Constr. 74, 2–11 (2017).
Article Google Scholar
Motl, R. W., McAuley, E., Snook, E. M. & Gliottoni, R. C. Physical activity and quality of life in multiple sclerosis: intermediary roles of disability, fatigue, mood, pain, self-efficacy and social support. Psychol. Health Med. 14, 111–124 (2009).
Article PubMed PubMed Central Google Scholar
Smuck, M., Tomkins-Lane, C., Ith, M. A., Jarosz, R. & Kao, M. J. Physical performance analysis: A new approach to assessing free-living physical activity in musculoskeletal pain and mobility-limited populations. PLoS One 12, e0172804 (2017).
Article PubMed PubMed Central Google Scholar
Tomkins-Lane, C. et al. Objective features of sedentary time and light activity differentiate people with low back pain from healthy controls: a pilot study. Spine J. 22, 629–634 (2022).
Article PubMed Google Scholar
Smuck, M. et al. Objective measurement of function following lumbar spinal stenosis decompression reveals improved functional capacity with stagnant real-life physical activity. Spine J. 18, 15–21 (2018).
Article PubMed Google Scholar
Rodríguez-Fernández, A., Lobo-Prat, J. & Font-Llagunes, J. M. Systematic review on wearable lower-limb exoskeletons for gait training in neuromuscular impairments. J. Neuroeng. Rehabil. 18, 1–21 (2021).
Article Google Scholar
Avila, F. R. et al. Wearable electronic devices for chronic pain intensity assessment: A systematic review. Pain. Pract. 21, 955–965 (2021).
Article PubMed Google Scholar
Xia, S., Song, S., Jia, F. & Gao, G. A flexible, adhesive and self-healable hydrogel-based wearable strain sensor for human motion and physiological signal monitoring. J. Mater. Chem. B 7, 4638–4648 (2019).
Article CAS PubMed Google Scholar
Pathak, Y. J. et al. Digital health integration with neuromodulation therapies: The future of patient-centric innovation in neuromodulation. Front. Digit. Health 3, 618959 (2021).
Article PubMed PubMed Central Google Scholar
Kushioka, J. et al. Gait Variability to Phenotype Common Orthopedic Gait Impairments Using Wearable Sensors. Sens. (Basel) 22, 9301 (2022).
Article Google Scholar
Costa, N. et al. Are objective measures of sleep and sedentary behaviours related to low back pain flares? Pain 163, 1829–1837 (2022).
Article PubMed Google Scholar
Perraudin, C. G. M. et al. Observational Study of a Wearable Sensor and Smartphone Application Supporting Unsupervised Exercises to Assess Pain and Stiffness. Digit Biomark. 2, 106–125 (2018).
Article PubMed PubMed Central Google Scholar
Chen, J., Abbod, M. & Shieh, J.-S. Pain and stress detection using wearable sensors and devices—A review. Sensors 21, 1030 (2021).
Article PubMed PubMed Central Google Scholar
Naeini, E. K., et al. An edge-assisted and smart system for real-time pain monitoring. in 2019 IEEE/ACM International Conference on Connected Health: Applications, Systems and Engineering Technologies (CHASE) 47-52 (IEEE, 2019).
Sett, N., et al. Are you in pain? Predicting pain and stiffness from wearable sensor activity data. in International Conference on Innovative Techniques and Applications of Artificial Intelligence 183-197 (Springer, 2019).
Chhikara, A., Rice, A., McGregor, A. H. & Bello, F. Wearable device for monitoring disability associated with low back pain. World 10, 13 (2008).
Google Scholar
Koenig, J., Loerbroks, A., Jarczok, M. N., Fischer, J. E. & Thayer, J. F. Chronic pain and heart rate variability in a cross-sectional occupational sample. Clin. J. pain. 32, 218–225 (2016).
Article PubMed Google Scholar
Miotto, R., Wang, F., Wang, S., Jiang, X. & Dudley, J. T. Deep learning for healthcare: review, opportunities and challenges. Brief. Bioinforma. 19, 1236–1246 (2018).
Article Google Scholar
Qayyum, A., Qadir, J., Bilal, M. & Al-Fuqaha, A. Secure and robust machine learning for healthcare: A survey. IEEE Rev. Biomed. Eng. 14, 156–180 (2020).
Article Google Scholar
Waring, J., Lindvall, C. & Umeton, R. Automated machine learning: Review of the state-of-the-art and opportunities for healthcare. Artif. Intell. Med. 104, 101822 (2020).
Article PubMed Google Scholar
Falla, D., Devecchi, V., Jimenez-Grande, D., Rugamer, D. & Liew, B. X. W. Machine learning approaches applied in spinal pain research. J. Electromyogr. Kinesiol 61, 102599 (2021).
Article PubMed Google Scholar
Lotsch, J. & Ultsch, A. Machine learning in pain research. Pain 159, 623–630 (2018).
Article PubMed Google Scholar
Lotsch, J., Ultsch, A., Mayer, B. & Kringel, D. Artificial intelligence and machine learning in pain research: a data scientometric analysis. Pain. Rep. 7, e1044 (2022).
Article PubMed PubMed Central Google Scholar
Miettinen, T. et al. Machine learning suggests sleep as a core factor in chronic pain. Pain 162, 109–123 (2021).
Article PubMed Google Scholar
Jenssen, M. D. K. et al. Machine learning in chronic pain research: a scoping review. Appl. Sci. 11, 3205 (2021).
Article CAS Google Scholar
Dudarev, V., Zhang, C., Barral, O., Davis, G. & Enns, J. T. Night-time cardiac metrics from a wearable sensor predict intensity of next-day chronic pain. Procedia Comput. Sci. 206, 34–44 (2022).
Article Google Scholar
Evans, S. et al. Heart rate variability as a biomarker for autonomic nervous system response differences between children with chronic pain and healthy control children. J. Pain. Res. 6, 449 (2013).
PubMed PubMed Central Google Scholar
Hamm-Faber, T. E., Gültuna, I., van Gorp, E.-J. & Aukes, H. High-dose spinal cord stimulation for treatment of chronic low back pain and leg pain in patients with FBSS, 12-month results: a prospective pilot study. Neuromodulation Technol. Neural Interface 23, 118–125 (2020).
Article Google Scholar
Kirketeig, T., Schultheis, C., Zuidema, X., Hunter, C. W. & Deer, T. Burst spinal cord stimulation: a clinical review. Pain. Med. 20, S31–S40 (2019).
Article PubMed PubMed Central Google Scholar
Beukenhorst, A. L., Schultz, D. M., McBeth, J., Sergeant, J. C. & Dixon, W. G. Are weather conditions associated with chronic musculoskeletal pain? Review of results and methodologies. Pain 161, 668–683 (2020).
Article PubMed Google Scholar
Barkley, J. E. et al. Increased Physical Activity and Reduced Pain with Spinal Cord Stimulation: A 12-Month Study. Int. J. Exerc. Sci. 13, 1583 (2020).
PubMed PubMed Central Google Scholar
Dueñas, M., Ojeda, B., Salazar, A., Mico, J. A. & Failde, I. A review of chronic pain impact on patients, their social environment and the health care system. J. Pain. Res. 9, 457 (2016).
Article PubMed PubMed Central Google Scholar
Ramineni, T. et al. The impact of spinal cord stimulation on sleep patterns. Neuromodulation Technol. Neural Interface 19, 477–481 (2016).
Article Google Scholar
Shaffer, F. & Ginsberg, J. P. An overview of heart rate variability metrics and norms. Front. Public Health 5, 258 (2017).
Article PubMed PubMed Central Google Scholar
Sperandei, S. Understanding logistic regression analysis. Biochem. Med. (Zagreb) 24, 12–18 (2014).
Article PubMed Google Scholar
Boser, B. E., Guyon, I. M. & Vapnik, V. N. A training algorithm for optimal margin classifiers. in Proceedings of the fifth annual workshop on Computational learning theory 144–152 (1992).
Cover, T. & Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. theory 13, 21–27 (1967).
Article Google Scholar
Breiman, L. Random Forests. Mach. Learn. 45, 5–32 (2001).
Article Google Scholar
Dorogush, A. V., Ershov, V. & Gulin, A. CatBoost: gradient boosting with categorical features support. arXiv preprint arXiv:1810.11363 (2018).
Lundberg, S. M. & Lee, S.-I. A unified approach to interpreting model predictions. Adv. Neural Inform Process. Syst. 30, 4768–4777 (2017).
Google Scholar
Lundberg, S. M., Erion, G. G. & Lee, S.-I. Consistent individualized feature attribution for tree ensembles. arXiv preprint arXiv:1802.03888 (2018).
Cella, D. et al. PROMIS® adult health profiles: efficient short-form measures of seven health domains. Value Health 22, 537–544 (2019).
Article PubMed PubMed Central Google Scholar
Fairbank, J. C. & Pynsent, P. B. The Oswestry disability index. Spine 25, 2940–2953 (2000).
Article CAS PubMed Google Scholar
Chawla, N. V., Bowyer, K. W., Hall, L. O. & Kegelmeyer, W. P. SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357 (2002).
Article Google Scholar
Download references
The REALITY sub-study is funded by Abbott Laboratories. We thank Dr. Misagh Mansouri, Dr. David Page, and Parisa Sarikhani for their help, time and dedication in reviewing the data analysis and revising the text. We would like to thank Prof. Joydeep Ghosh and Dr. Aryan Mokhtari for their advice on the data analysis and machine learning execution of the project.
Nevada Advanced Pain Specialists, Reno, NV, USA
Denis G. Patterson
Goodman Campbell Brain & Spine, Carmel, IN, USA
Derron Wilson
Center for Interventional Pain and Spine, Lancaster, PA, USA
Michael A. Fishman
Pacific Sports and Spine, Eugene, OR, USA
Gregory Moore
Expert Pain, Houston, TX, USA
Ioannis Skaribas
Spinal Diagnostics, Tualatin, OR, USA
Robert Heros
Abbott Neuromodulation, Plano, TX, USA
Soroush Dehghan, Erika Ross & Anahita Kyani
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
D.P., M.F., G.M., I.S., R.H., and D.W. were investigators in the study, and they were involved in enrolling subjects, data collection, and revising the manuscript. S.D. performed analyses of the objective and subjective data collected in the study, led statistical analysis, and was involved in writing the manuscript. E.R. was involved in the study design and revising of the manuscript. A.K. was the scientist of the study and was involved in the study design and execution, writing the protocol and consent form, designing the REALITY watch and iPhone custom application, testing the application, developing the backend for data collection and processing, creating dashboards for data monitoring and writing the manuscript.
Correspondence to Denis G. Patterson.
D.P. is a consultant, investigator, advisory board member, and proctor for Vertiflex; consultant, investigator, proctor, and speaker for Abbott. D.W. is a consultant for Abbott, Biotronik, and Boston Scientific and he is on the advisory board for Abbott and Biotronik. R.H. is a consultant for Abbott, Biotronik, and Boston Scientific, performs research for Abbott, Ethos, Mainstay, and Saluda, and speaker’s bureau for Abbott and Boston Scientific. M.F. is a consultant for Biotronik, Biowave, Brixton Biosciences, Medtronic, and Saluda and receives grants from Abbott, Biotronik, Boston Scientific, Medtronic, Nalu Medical, SGX Medical, and Thermaquil. I.S. is a consultant for Abbott. G.M. is a consultant for Abbott, Boston Scientific, Relievant, and Vertos and does research for Abbott, Boston Scientific, Nalu, and Mainstay Medical. S.D., E.R., and A.K. are Abbott employees.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
Reprints and Permissions
Patterson, D.G., Wilson, D., Fishman, M.A. et al. Objective wearable measures correlate with self-reported chronic pain levels in people with spinal cord stimulation systems. npj Digit. Med. 6, 146 (2023). https://doi.org/10.1038/s41746-023-00892-x
Download citation
Received: 27 January 2023
Accepted: 03 August 2023
Published: 15 August 2023
DOI: https://doi.org/10.1038/s41746-023-00892-x
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative