Skip to main content

Reliability of a new measure to assess modern screen time in adults



Screen time among adults represents a continuing and growing problem in relation to health behaviors and health outcomes. However, no instrument currently exists in the literature that quantifies the use of modern screen-based devices. The primary purpose of this study was to develop and assess the reliability of a new screen time questionnaire, an instrument designed to quantify use of multiple popular screen-based devices among the US population.


An 18-item screen-time questionnaire was created to quantify use of commonly used screen devices (e.g. television, smartphone, tablet) across different time points during the week (e.g. weekday, weeknight, weekend). Test-retest reliability was assessed through intra-class correlation coefficients (ICCs) and standard error of measurement (SEM). The questionnaire was delivered online using Qualtrics and administered through Amazon Mechanical Turk (MTurk).


Eighty MTurk workers completed full study participation and were included in the final analyses. All items in the screen time questionnaire showed fair to excellent relative reliability (ICCs = 0.50–0.90; all < 0.000), except for the item inquiring about the use of smartphone during an average weekend day (ICC = 0.16, p = 0.069). The SEM values were large for all screen types across the different periods under study.


Results from this study suggest this self-administered questionnaire may be used to successfully classify individuals into different categories of screen time use (e.g. high vs. low); however, it is likely that objective measures are needed to increase precision of screen time assessment.

Peer Review reports


Screen time among adults represents a continuing and growing problem in relation to health behaviors and health outcomes. Extended periods of screen use – interaction with electronic devices that primarily deliver content via screen-based displays – have been associated with multiple adverse health outcomes including obesity, type 2 diabetes, cardiovascular disease, and early mortality in adults [1,2,3,4,5]. Associations remain even when controlling for age, health history, and health-related behaviors such as smoking, alcohol consumption, physical activity and dietary intake [1,2,3,4,5].

The impact of screen use on health is a complex phenomenon that may go above and beyond the sedentary behavior that may result from prolonged sitting. Television watching has been previously associated with poor dietary choices in part due to heavy commercial advertisement [6]. The frequent use of smartphones has been associated to sleep disturbances [7, 8] potentially resulting from exposure to radio frequency electromagnetic fields that can affect brain physiology [9]. The use of e-readers in the evening has also been shown to affect sleep via suppression of melatonin secretion and alterations to the circadian rhythm [10]. In addition, long periods of screen time have been associated with poor mental health including depression [11, 12], which in turn may disturb the hypothalamic adrenal axis adversely affecting immune function and metabolism [13].

The problem is unlikely to resolve itself given the near ubiquity of screens in modern life. The average American family, for example, owns three television sets [14] and among television-owning households, close to 60% own at least one internet-enabled device such as a Smart TV or a video game console [15]. In addition, approximately 77% of Americans own a smartphone and 78% have at least one desktop or laptop at home [16, 17].

Moreover, with the rapid evolution of screen-based technologies, the landscape of media use has changed at an equally dramatic pace. For example, even though television remains the most widely used screen-based device among US adults, its popularity has steadily decreased in recent years [18]. Simultaneously, the use of portable screen-based devices increased quickly. Recent Nielsen data showed that use of ‘apps’ and internet services on a smartphone more than doubled from 2015 to 2017, and use of apps and web services on tablets increased in the same period by 70% [15].

Moreover, television-connected devices (e.g. gaming consoles, smart TVs) have given rise to the phenomenon of “binge watching” – viewing multiple episodes of a television show in a single sitting – which is observed across multiple generations [19]. Hence, these newer devices represent critical considerations for capturing an updated view of how adults interact and spend time with screens.

Past research on screen time has focused almost exclusively on quantifying “screen use” as the number of hours of television viewing in a given week [20] and only a small number of studies have aggregated television/videos/games and computer use [21, 22]. As such, the instruments available within the health literature designed to capture screen time behavior generally only measure television viewing time and sometimes non-occupational computer use alongside other sedentary behaviors such as reading, driving, and socializing. To the authors’ knowledge, no instrument currently exists that quantifies the use of other specific screen-based devices alongside use of televisions and computers. This represents a critical gap in screen-time assessment as it might be important to distinguish how and when different modern devices are used along with traditional screen-based devices, as well as the variety of ways that different types of screens might be used (e.g. committed use vs. use only as background noise), which may possibly be associated with different patterns of sedentary behavior activity and health outcomes.

Although a variety of screen time tracking applications (‘apps’) and devices are available for smartphones, tablets, PCs, and televisions, to the authors’ knowledge no single application or tool exists that can track all devices at the same time for overall estimation of total screen time. Asking participants to track usage per device via various apps could be burdensome and possibly reveal more information than a participant might want, such as minutes on social media or on particular apps. Also, current television screen time trackers can be significantly expensive and could potentially require research personnel to perform a home visit for installation.

A questionnaire with strong psychometric properties can be a useful research tool that estimates global screen time in a simple, fast, no-cost, and completely anonymous manner. This questionnaire then may be used with online anonymous samples and/or large clinical studies that examine the association between modern screen time and health outcomes. Therefore, the purpose of this study was to develop and assess the reliability of a new screen time questionnaire designed to quantify use of multiple popular screen devices among the US adult population, including television, television-connected devices, laptops, smartphones, and tablets.


Screen time questionnaire

An 18-item screen time online questionnaire was created to quantify use of commonly used screen-based devices (please see Additional file 1: Screen Time Questionnaire). Five different categories of devices were created – TV, TV-connected devices (e.g. streaming devices, video game consoles), laptop/computer, smartphone, and tablet – based on the classification scheme used by publicly available reports on screen time usage among the American population [15, 23,24,25]. These categories were believed to appropriately reflect the purpose of the questionnaire, which is to quantify different forms of screen time use among American adults. For instance, it was important to differentiate between TV and TV-connected devices because of the growing trend of adults watching subscription-based video and on-demand content instead of regularly programmed television [26, 27].

Moreover, it was considered appropriate to include a variety of internet-enabled devices, such as game consoles and multimedia devices, under the same category (TV-connected devices) since the use of these devices would reflect the same sedentary behavior of sitting for prolonged periods of time that is initiated at any time during the day, instead of being a behavior that is driven by a predetermined schedule like TV programming.

In the present online questionnaire, participants were instructed to estimate total time spent in hours and minutes using each device. Total time for each screen-based device was quantified in minutes (e.g. 1 h and 30 min = 90 min). Because screen time use shows variation throughout the day and week [24], the questionnaire further inquired about screen use during an average weekday, an average weeknight, and an average weekend day (Saturday or Sunday) separately.

In addition, because the use of a screen while performing other activities that require body movement would not represent sedentary behavior that impacts physical health and thus health outcomes, the questionnaire was divided into sections exploring screen use as primary activity and screen use in the background. Taking into account the time spent only in a primary activity has been used previously by time-use surveys [28]. The “primary activity” was defined in the survey as “the main activity you are engaged in rather than using a television or other screen in the background while performing another activity, such as cooking or exercising.” Background use was defined as “the use of a television or another screen near you while performing other activities such as exercising, cooking, and interacting with family/friends.”

Lastly, basic demographic information was also collected (e.g., age, sex) along with self-reported height, weight, and physical activity levels. Body mass index (BMI) was calculated by dividing weight in kilograms over height in meters squared. Physical activity levels were estimated with the Stanford Leisure-Time Activity Categorical Item (L-Cat), which has previously shown adequate psychometric properties [29].

Data collection

The questionnaire was delivered online using Qualtrics and administered through Amazon Mechanical Turk (MTurk). U.S. workers on MTurk are more similar to the U.S. population compared to subjects recruited from traditional university subject pools and provide greater diversity in terms of age, ethnicity, and socio-economic status [30,31,32]. In addition, previous studies have shown that MTurk samples can provide valid and reliable data for health and social science research [30,31,32]. For instance, Casler, Bickel, & Hackett adapted a behavioral, face-to-face task to an online test and found that test results were almost identical between a standard sample of college students attending an in-person lab session and participants recruited online through MTurk [33].

Data collection was conducted between March and October 2018. To achieve a high level of data quality, multiple strategies were employed such as attention checks and a ‘captcha’ verification to exclude spam and automated responses. Inclusion criteria included adults 18 years of age and older who watched television or a television-connected device for at least 2 h daily and owned at least one other screen-based device other than a television, English speakers, and current residents of the US. The study was approved by the Institutional Review Board of a university in the Southwest of the United States.

Eligible MTurk workers signed an electronic informed consent form and were redirected to the Qualtrics survey where they were asked about their demographics, height, weight, current physical activity levels, and screen time (time 1). At the end of the survey, a completion code and a new Qualtrics link was provided. Participants were instructed to access the link 3 days later to complete the screen-time questionnaire a second time (time 2). Afterwards, they were provided a final code to enter into the MTurk website in order to receive a $5 payment. The research team included only participants whose data entries were separated by at least 3 days.

Data analyses

Data were analyzed using the Statistical Package for the Social Sciences (SPSS) version 21.0. Because all variables under investigation were found positively skewed violating normality assumptions, even after transformation attempts, screen time variables are presented as medians and interquartile ranges. Test-retest reliability of the screen-time questionnaire was assessed via two different approaches. Relative reliability was assessed through intra-class correlation coefficients (ICCs) using a two-way mixed effects, single measurement, absolute agreement model. Values represent the following: less than 0.40 - poor reliability, between 0.40 and 0.59 – fair reliability, between 0.60 and 0.74 - good reliability, and greater than 0.75 - excellent reliability [34]. Relative reliability refers to consistency of the position of individuals in the group relative to others, and hence allows for the determination of how well participants can be distinguished from each other regardless of measurement errors [34, 35].

Absolute reliability, or agreement, was assessed through the standard error of measurement (SEM) using the following formula:

$$ \mathrm{SEM}={{\sqrt{\sigma}}_{\mathrm{e}}}^2 $$

where σ e2 is the error variance in a repeated measures analysis of variance (ANOVA). In addition, two-sided 95% confidence intervals were estimated from the following formula:

$$ \surd \left[\mathrm{SSE}/{x^2}_{\mathrm{a},\mathrm{dfe}};\mathrm{SSE}/{x^2}_{1-\mathrm{a},\mathrm{dfe}}\right] $$

where SSE is the sum of squares error from the repeated measures ANOVA, x2a, dfe is the chi-square value for the probability level alpha and dfe is the degrees of freedom associated with SSE. Absolute reliability or agreement refers to the consistency of scores of individuals, and therefore indicates how similar the scores for repeated measures are when measurement error (systematic and random) is present [34, 35].

Sample size calculation was performed using the Donner & Eliasziw approach [36] using an alpha level of 0.05 and a power of 0.80. For an ICC between 0.4 and 06, corresponding to acceptable reliability, a sample size of at least 86 is required. On the other hand, for an ICC between 0.6 and 0.8, corresponding to good to excellent reliability, a sample size of at least 39 is required [36]. In addition, it was determined that two observations were adequate for the present analysis. Previous authors have suggested that when examining the reproducibility of a questionnaire, two observations per subject are more adequate in order to avoid memory effect; also, the width of the 95 confidence interval does not change between two and three observations [37]. Due to follow-up issues with previous data collection using MTurk by our research team, it was decided to request the participation of 200 workers in order to approximate a sample size of 86.


A total of 170 MTurk workers enrolled in the study; 83 participants did not take the second survey and seven participants were excluded for varied reasons (e.g. failing the check question, incomplete data, etc.). Hence, a total of 80 participants completed full study participation and were included in the final analyses. The majority of participants were male (62.5%), 35 years of age or younger (60%), non-Hispanic White (91.3%), single (58.8%), employed full-time (78.8%), had a Bachelor’s or higher degree (58.8%), had a household income of less than $60,000 per year (68.8%), and had a BMI of 26.7 or less (58.8%). Additionally, approximately half (53.8%) reported engaging in current physical activity recommendations of at least 30 min or more of moderate-intensity activity 5 days per week [38].

Minutes spent on each screen device as a primary form of activity are presented in Table 1. The most commonly used screen during a weekday was a laptop/computer, followed by television and television-connected devices, smartphone, and tablet. The pattern was similar for screen use during a weeknight and a weekend day. Lastly, participants reported the most background screen use during the weekend with an aggregate of approximately 3 h across devices, whereas background screen use during an average weekday and average weeknight was approximately 2 h (see Table 2).

Table 1 Minutes spent on different screen devices as a primary activity during an average weekday, weeknight, and weekend day by MTurk participants (N = 80)
Table 2 Minutes exposed to background screen use during an average weekday, weeknight, and weekend by MTurk participants (N = 80)

Relative reliability results for the screen-time questionnaire are presented in Table 3. Items inquiring about television, laptop/computer, smartphone, and tablet use during a weekday and the three questions associated with screen use in the background showed good to excellent reliability (ICCs = 0.61–0.90). Items inquiring about screen use during a weeknight showed fair to excellent reliability (ICCs = 0.50–0.82). Items inquiring about screen use during a weekend day showed excellent reliability (ICCs = 0.84–0.87), except for smartphone use (ICC = 0.16).

Table 3 Intra-class correlation results for the different sections of the screen-time questionnaire

Absolute reliability results for the screen-time questionnaire is presented in Table 4. Overall, SEM values were large for all types of screens across the different periods under study. However, measurement error was smaller among the items inquiring about television, laptop/computer, smartphone, and tablet use during an average weeknight, as were confidence intervals. Among the different types of screen, television-connected devices and laptop/computer use during a weekday and weeknight showed the most precision, whereas smartphone use during a weekend day showed the greatest measurement error.

Table 4 Standard error of measurement results for the different sections of the screen-time questionnaire


The present study assessed the reliability, both relative and absolute, of a newly developed screen time questionnaire. The questionnaire quantifies the use of a variety of screen-based devices that better reflects the more varied screen-time behaviors of US adults, and it makes the distinction between different contexts in which screens may be used (actual use vs. background use as well as day vs. evening vs. weekend use). To our knowledge, this is the first questionnaire that inquiries about the use of commonly used screens beyond television and computers among the US adult population.

Current questionnaires predominantly focus on overall sedentary behavior and estimate “screen time” by summing hours spent per day or week watching television/television or computer/television or playing games [39], but they do not quantify other forms of screen use such as tablets and smartphones. The questionnaire by Marshall et al. [40] additionally included an item asking respondents to report hours spent sitting during leisure time engaged in other activities not including television, such as visiting friends, watching movies away from the home, or dining. The inclusion of this catch-all “other” section might potentially include the use of additional screen devices; however, the questionnaire does not include a way to explicitly make this distinction.

The present questionnaire is also the first one to include a separate assessment for weeknights and background use. It has been previously reported that a large percentage of Americans engage in screen time during the evenings compared to the rest of the day [24], and hence screen time specifically during weeknights may serve as a valuable target of change for interventions that aim to reduce overall sedentary behavior. Similarly, prior research has found that background television use has detrimental effects on cognitive processing [41, 42] and thus background screen time may be an important target of change for behavioral interventions that seek to improve overall well-being including intellectual development and learning.

Results from this study compare favorably to previous test-retest reliability studies in a number of ways. Previous questionnaires have shown ICCs ranging from 0.62 to 0.69 for computer use [39], whereas the present questionnaire demonstrated higher ICCs ranging from 0.68 to 0.89 across the different periods under study, indicating good to excellent relative reliability. Prior questionnaires have also found ICCs ranging from 0.32 to 0.82 for television viewing [39, 43], whereas this questionnaire found slightly higher ICCs ranging from 0.50 to 0.87. In addition, items inquiring about additional types of screen such as television-connected devices and tablets showed ICCs ranging from 0.61 to 0.90, indicating good to excellent relative reliability, except for smartphone during an average weekend day. As a result, the present questionnaire was able to successfully distinguish between low- and high- users of a variety of different screen-based devices in addition to television and computer.

Furthermore, to our knowledge this is the first study that examines the absolute reliability of a self-administered questionnaire inquiring about screen time in adults. Our analyses indicated that even though our questionnaire would be able to adequately classify individuals into different categories (e.g. heavy screen time users vs. moderate screen time users) as evidenced by high ICCs, this self-report measure may not provide consistent results over repeated assessments as indicated by measurement error. The results indicated that the most precise estimation of screen time was for television-connected devices and laptop/computer use during a weekday and weeknight, whereas the least precise was smartphone use during the weekend. This may the result of distinct periods of time when people watch favorite TV shows and/or engage in specific computer tasks on a daily basis allowing for easy recall and thus more precise estimation of screen time. On the other hand, during weekends when individuals have more free time at their disposal, it could be difficult to precisely estimate how much screen time is spent with a particular device, particularly smartphones that provide easy access to browsing at any time during the day and at almost any location.

These results highlight the importance of using objective measures in addition to self-administered questionnaires when assessing screen time. Objective measures such as television timers and smartphones apps may provide a more accurate quantification of time spent on a variety of screen devices, which may be particularly helpful for studies that seek to demonstrate a reduction in screen time after a behavioral intervention that is due to real change and not simply due to measurement error.

Lastly, these results emphasize the need to quantify different types of screen use in order to provide a more accurate representation of overall screen time for adults. Participants in this study reported spending more hours combined using television-connected devices and smartphones compared to television alone across all periods under study; as such, including television only as a measure of “screen time” may substantially underestimate true screen time and potentially sedentary behavior that may contribute to obesity and other chronic conditions.

Future directions

Some limitations to this study exist, including a brief test-retest reliability period. Future studies should examine the stability of reliability coefficients using a longer test-retest period. Also, the questionnaire needs to be examined for additional psychometric properties such as convergent and discriminant validity. Future studies could employ measures to objectively quantify hours of screen use (e.g. television monitors, smartphone apps) in addition to activity monitors that estimate sedentary behavior (e.g. activPAL™) and compare these against self-reported screen time as measured by the present screen-time questionnaire.

Lastly, it would be interesting to develop another questionnaire that quantifies the concurrent use of different screen-based devices. A recent meta-analysis found that media multitasking has a detrimental impact on cognitive outcomes such as attention, comprehension, and recall; nevertheless, little is known about the consequences of media multitasking on long-term mental health [44].


To the authors’ knowledge, this is the first study to test reliability (both relative and absolute) of a screen-time questionnaire tool that incorporates a greater variety of modern screen-based devices. Relative reliability results suggest this tool could be used to appropriately classify individuals into different screen time categories across multiple devices (e.g. heavy users of television vs. light users of television; heavy users of laptop/computer vs. moderate users of laptop/computer). Our results also emphasize the importance of assessing diverse types of screen to obtain an accurate representation of overall screen time.

Availability of data and materials

The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.



Analysis of variance


Body Mass Index


Intra-class correlation coefficients


Stanford Leisure-Time Activity Categorical Item


Amazon Mechanical Turk


Standard error of measurement


Statistical Package for the Social Sciences


  1. 1.

    Dunstan DW, Barr EL, Healy GN, Salmon J, Shaw JE, Balkau B, et al. Television viewing time and mortality: the Australian diabetes, obesity and lifestyle study (AusDiab). Circulation. 2010;121:384–91.

    CAS  PubMed  Article  Google Scholar 

  2. 2.

    Hu FB, Leitzmann MF, Stampfer MJ, Colditz GA, Willett WC, Rimm EB. Physical activity and television watching in relation to risk for type 2 diabetes mellitus in men. Arch Intern Med. 2001;161:1542–8.

    CAS  PubMed  Article  Google Scholar 

  3. 3.

    Hu FB, Li TY, Colditz GA, Willett WC, Manson JE. Television watching and other sedentary behaviors in relation to risk of obesity and type 2 diabetes mellitus in women. JAMA. 2003;289:1785–91.

    PubMed  Article  Google Scholar 

  4. 4.

    Salmon J, Bauman A, Crawford D, Timperio A, Owen N. The association between television viewing and overweight among Australian adults participating in varying levels of leisure-time physical activity. Int J Obes. 2000;24:600–6.

    CAS  Article  Google Scholar 

  5. 5.

    Wijndaele K, Brage S, Besson H, Khaw K, Sharp SJ, Luben R, et al. Television viewing and incident cardiovascular disease: prospective associations and mediation analysis in the EPIC Norfolk study. PLoS One. 2011;6:e20058.

    CAS  PubMed  PubMed Central  Article  Google Scholar 

  6. 6.

    Coon KA, Tucker K. Television and children’s consumption patterns. Minerva Pediatr. 2002;54:423–36.

    CAS  PubMed  Google Scholar 

  7. 7.

    Yogesh S, Abha S, Priyanka S. Mobile usage and sleep patterns among medical students. Indian J Physiol Pharmacol. 2014;58:100–3.

    PubMed  Google Scholar 

  8. 8.

    Demirci K, Argonul M, Akpinar A. Relationship of smartphone use severity with sleep quality, depression, and anxiety in university students. J Behav Addict. 2015;4:85–92.

    PubMed  PubMed Central  Article  Google Scholar 

  9. 9.

    Regel SJ, Tinguely G, Schuderer J, Adam M, Kuster N, Landolt H, et al. Pulsed radio-frequency electromagnetic fields: dose-dependent effects on sleep, the sleep EEG and cognitive performance. J Sleep Res. 2007;16:253–8.

    PubMed  Article  Google Scholar 

  10. 10.

    Chang A, Aeschbach D, Duffy JF, Czeisler CA. Evening use of light-emitting eReaders negatively affects sleep, circadian timing, and next-morning alertness. PNAS. 2015;112:1232–7.

    CAS  PubMed  Article  Google Scholar 

  11. 11.

    Hamer M, Stamatakis E, Mishra GD. Television- and screen-based activity and mental well-being in adults. Am J Prev Med. 2010;38:375–80.

    PubMed  Article  Google Scholar 

  12. 12.

    Hamer M, Poole L, Messerli-Burgy N. Television viewing, c-reactive protein, and depressive symptoms in older adults. Brain Behav Immun. 2013;33:29–32.

    CAS  PubMed  Article  Google Scholar 

  13. 13.

    Tsigos C, Chrousos GP. Hypothalamic-pituitary-adrenal axis, neuroendocrine factors and stress. J Psychosom Res. 2002;53:865–71.

    PubMed  Article  Google Scholar 

  14. 14.

    Nielsen. Television audience report 2010 & 2011. US: The Nielsen Company, LLC; 2011. Available from: Cited 2011 Nov 30

    Google Scholar 

  15. 15.

    Nielsen. The Nielsen total audience report: Q2 2017. US: The Nielsen Company, LLC; 2017. Available from: Cited 2017 Nov 16

    Google Scholar 

  16. 16.

    Pew Research Center. Mobile fact sheet. Washington, DC: Pew Research Center; 2018. Available from: Cited 2018 Feb 5

    Google Scholar 

  17. 17.

    Ryan C, Lewis JM. Computer and internet use in the United States: 2015. United States Census Bureau: US; 2017. Available from: Cited 2017 Sept

    Google Scholar 

  18. 18.

    Mann G, Venturini F, Murdoch R, Mishra B, Moorby G, Carlier B. Digital video and the connected consumer. Accenture: US; 2015. Available from: Cited 2015

    Google Scholar 

  19. 19.

    Jenner M. Binge-watching: video-on-demand, quality TV and mainstreaming fandom. Int J Cult Stud. 2017;20:304–20.

    Article  Google Scholar 

  20. 20.

    Wilmot EG, Edwardson CL, Achana FA, Davies MJ, Gorely T, Gray LJ, et al. Sedentary time in adults and the association with diabetes, cardiovascular disease and death: systematic review and meta-analysis. Diabetologia. 2012;55:2895–905.

    CAS  PubMed  Article  Google Scholar 

  21. 21.

    Stamatakis E, Hamer M, Dunstan DW. Screen-based entertainment time, all-cause mortality, and cardiovascular events: population-based study with ongoing mortality and hospital events follow-up. J Am Coll Cardiol. 2011;57:292–9.

    PubMed  Article  Google Scholar 

  22. 22.

    Wijndaele K, Brage S, Besson H, Khaw KT, Sharp SJ, Luben R, et al. Television viewing time independently predicts all-cause and cardiovascular mortality: the EPIC Norfolk study. Int J Epidemiol. 2011;40:150–9.

    PubMed  Article  Google Scholar 

  23. 23.

    Nielsen. The total audience report: December 2014. US: The Nielsen Company, LLC; 2014. Available from: Cited 2014 Dec 3

    Google Scholar 

  24. 24.

    Nielsen. The total audience report: Q2 2015. US: The Nielsen Company, LLC; 2015. Available from: Cited 2015 Sept 21

    Google Scholar 

  25. 25.

    Nielsen. The Nielsen comparable metrics report: Q4 2016. US: The Nielsen Company, LLC; 2017. Available from: Cited 2017 May 25

    Google Scholar 

  26. 26.

    Nielsen. More of what we want: The cross-platform report June 2014. US: The Nielsen Company, LLC; 2014. Available from: Cited 2014 June 30

    Google Scholar 

  27. 27.

    Nielsen. Video 360: 2018 Report Highlights. US: The Nielsen Company, LLC; 2018. Available from: Cited 2018 Aug 27

    Google Scholar 

  28. 28.

    Hamermesh DS, Frazis H, Stewart J. Data watch: the American time use survey. J Econ Perspect. 2005;19:221–32.

    Article  Google Scholar 

  29. 29.

    Kiernan M, Schoffman DE, Lee K, Brown SD, Fair JM, Perri MG, et al. The Stanford leisure-time activity categorical item (L-cat): a single item sensitive to physical activity changes in overweight/obese women. Int J Obes. 2013;37:1597–602.

    CAS  Article  Google Scholar 

  30. 30.

    Paolacci G, Chandler J, Ipeirotis PG. Running experiments on Amazon mechanical Turk. Judgm Decis Mak. 2010;5:411–9.

    Google Scholar 

  31. 31.

    Paolacci G, Chandler J. Inside the Turk: understanding mechanical Turk as a participant tool. Curr Dir Psychol Sci. 2014;23:184–8.

    Article  Google Scholar 

  32. 32.

    Mason W, Suri S. Conducting behavioral research on Amazon’s mechanical Turk. Behav Res Methods. 2012;44:1–23.

    PubMed  Article  Google Scholar 

  33. 33.

    Casler K, Bickel L, Hackett E. Separate but equal? A comparison of participants and data gathered via Amazon’s Mturk, social media, and face-to-face behavioral testing. Comput Hum Behav. 2013;29:2156–60.

    Article  Google Scholar 

  34. 34.

    Hernaez R. Reliability and agreement studies: a guide for clinical investigators. Gut. 2015;64:1018–27.

    PubMed  Article  Google Scholar 

  35. 35.

    Weir JP. Quantifying test-retest reliability using the intraclass correlation coefficient and the SEM. J Strength Cond Res. 2005;19:231–40.

    PubMed  Google Scholar 

  36. 36.

    Donner A, Eliasziw M. Sample size requirements for reliability studies. Stat Med. 1987;6:441–8.

    CAS  PubMed  Article  Google Scholar 

  37. 37.

    Giraudeau B, Mary JY. Planning a reproducibility study: how many subjects and how many replicates per subject for an expected width of the 95 percent confidence interval of the intraclass correlation coefficient. Stat Med. 2001;20:3205–14.

    CAS  PubMed  Article  Google Scholar 

  38. 38.

    Garber CE, Blissmer B, Deschenes MR, Franklin BA, Lamonte MJ, Lee IM, et al. American College of Sports Medicine position stand. Quantity and quality of exercise for developing and maintaining cardiorespiratory, musculoskeletal, and neuromotor fitness in apparently healthy adults: guidance for prescribing exercise. Med Sci Sports Exerc. 2011;43:1334–59.

    PubMed  Article  Google Scholar 

  39. 39.

    Clark BK, Sugiyama T, Healy GN, Salmon J, Dunstan DW, Owen N. Validity and reliability of measures of television viewing time and other non-occupational sedentary behaviour of adults: a review. Obes Rev. 2009;10:7–16.

    PubMed  Article  Google Scholar 

  40. 40.

    Marshall AL, Miller YD, Burton NW, Brown WJ. Measuring total and domain-specific sitting: a study of reliability and validity. Med Sci Sports Exerc. 2010;42:1094–102.

    PubMed  Google Scholar 

  41. 41.

    Armstrong GB, Greenberg BS. Background television as an inhibitor of cognitive processing. Hum Commun Res. 1990;16:355–86.

    Article  Google Scholar 

  42. 42.

    Armstrong GB, Chung L. Background television and reading memory in context: assessing TV interference and facilitative context effects on encoding versus retrieval processes. Commun Res. 2000;27:327–52.

    Article  Google Scholar 

  43. 43.

    Pettee KK, Ham SA, Macera CA, Ainsworth BE. The reliability of a survey question on television viewing and associations with health risk factors in US adults. Behav Psychol. 2009;17:487–93.

    Google Scholar 

  44. 44.

    Jeong S, Hwang Y. Media multitasking effects on cognitive vs. attitudinal outcomes: a meta-analysis. Hum Commun Res. 2016;42:599–618.

    Article  Google Scholar 

Download references


This research was presented as a poster at the Society for Behavioral Medicine conference, Washington, D.C., 2019. The abstract can be found here:

Vizcaino M, Buman M, DesRoches CT, Wharton CM. Reliability of a new measure to assess modern screen time habits in adults. Ann Behav Med. 2019; 53 (Suppl 1):S1–S842.


This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Author information




MV, CW, and CTD conceived of the study. MV and CW designed the survey and the scale. MV conducted data collection, analysis, and initial interpretation; and drafted the manuscript. MB substantially contributed in data analysis. MV, CW, CTD, and MB contributed to interpret the data, substantively revised the manuscript and approved the submitted version.

Corresponding author

Correspondence to Christopher Wharton.

Ethics declarations

Ethics approval and consent to participate

All procedures performed in this study were in accordance with the ethical standards of the Institutional Review Board from Arizona State University. Informed consent was obtained from all individual participants included in the study. As this was an online study, participants signed an electronic informed consent form, which then directed them to the survey for data collection. This method of informed consent was approved by Arizona State University’s Institutional Review Board.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary information

Additional file 1.

Screen-time Questionnaire. The questionnaire includes all items used to quantify the use of a variety of modern screen-based devices.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Vizcaino, M., Buman, M., DesRoches, C. et al. Reliability of a new measure to assess modern screen time in adults. BMC Public Health 19, 1386 (2019).

Download citation


  • Screen-time
  • Television
  • Smartphone
  • ICC
  • SEM
  • Reliability