scholarly journals Pilot study of a new concussion assessment tool using computerized cognitive testing and instrumented balance: Test-retest reliability

2017 ◽  
Vol 2 (6) ◽  
Author(s):  
Neville C ◽  
Rieger B ◽  
Viterise T ◽  
Brindle A
Author(s):  
Matthew L. Hall ◽  
Stephanie De Anda

Purpose The purposes of this study were (a) to introduce “language access profiles” as a viable alternative construct to “communication mode” for describing experience with language input during early childhood for deaf and hard-of-hearing (DHH) children; (b) to describe the development of a new tool for measuring DHH children's language access profiles during infancy and toddlerhood; and (c) to evaluate the novelty, reliability, and validity of this tool. Method We adapted an existing retrospective parent report measure of early language experience (the Language Exposure Assessment Tool) to make it suitable for use with DHH populations. We administered the adapted instrument (DHH Language Exposure Assessment Tool [D-LEAT]) to the caregivers of 105 DHH children aged 12 years and younger. To measure convergent validity, we also administered another novel instrument: the Language Access Profile Tool. To measure test–retest reliability, half of the participants were interviewed again after 1 month. We identified groups of children with similar language access profiles by using hierarchical cluster analysis. Results The D-LEAT revealed DHH children's diverse experiences with access to language during infancy and toddlerhood. Cluster analysis groupings were markedly different from those derived from more traditional grouping rules (e.g., communication modes). Test–retest reliability was good, especially for the same-interviewer condition. Content, convergent, and face validity were strong. Conclusions To optimize DHH children's developmental potential, stakeholders who work at the individual and population levels would benefit from replacing communication mode with language access profiles. The D-LEAT is the first tool that aims to measure this novel construct. Despite limitations that future work aims to address, the present results demonstrate that the D-LEAT represents progress over the status quo.


2013 ◽  
Vol 02 (04) ◽  
pp. 11-16 ◽  
Author(s):  
Yanhong Dong ◽  
Claire L. Thompson ◽  
Shi Huey Joanne Tan ◽  
Leon Ben Swie Lim ◽  
Wanshin Pang ◽  
...  

2017 ◽  
Vol 49 (5S) ◽  
pp. 48
Author(s):  
Jacob Shepherd ◽  
Autumn Oostindie ◽  
Jared Locke ◽  
Srikant Vallabhajosula ◽  
Deborah Stetts ◽  
...  

2020 ◽  
Vol 14 (2) ◽  
pp. 309-317
Author(s):  
Michelle L. Manning ◽  
Harsimran Singh ◽  
Keaton Stoner ◽  
Steph Habif

Background: With the rapid development of new insulin delivery technology, measuring patient experience has become especially pertinent. The current study reports on item development, psychometric validation, and intended use of the newly developed Diabetes Impact and Device Satisfaction (DIDS) Scale. Method: The DIDS Scale was informed by a comprehensive literature review, and field tested as part of two focus groups. The finalized measure was used at baseline and 6 months post-assessment with a large US cohort. Exploratory factor analyses (EFAs) were conducted to determine and confirm factor structure and item selection. Internal reliability, test–retest reliability, and convergent/divergent validity of the emerged factors were tested with demographics, diabetes-specific information, and diabetes behavioral and satisfaction measures. Results: In all, 778 participants with type 1 diabetes (66% female, mean age 47.13 ± 17.76 years, 74% insulin pump users) completed surveys at both baseline and post-assessment. EFA highlighted two factors—Device Satisfaction (seven items, Cronbach’s α = 0.85-0.90) and Diabetes Impact (four items, Cronbach’s α = 0.71-0.75). DIDS Scale demonstrated good concurrent validity and test–retest reliability. Conclusion: The DIDS Scale is a novel and a brief assessment tool with robust psychometric properties. It is recommended for use across all insulin delivery devices and is considered appropriate for use in longitudinal studies. Future studies are recommended to evaluate the performance of DIDS Scale in diverse populations with diabetes.


2020 ◽  
Vol 15 (6) ◽  
pp. 795-800 ◽  
Author(s):  
Nicola Taylor ◽  
David Giles ◽  
Micha Panáčková ◽  
James Mitchell ◽  
Joel Chidley ◽  
...  

Purpose: To assess the validity and reliability of a novel movement-performance assessment tool for climbing/sport climbing. Methods: First, salient climbing movement-performance factors were identified through an iterative consultation process with 10 expert climbing coaches; the resulting Climber’s Movement Performance Assessment Tool (CM-PAT) contained 14 items in 5 categories. Second, 61 intermediate to advanced climbers ascended a single route, which was video recorded. Subsequently, 4 experienced (>10 y coaching) coaches used the CM-PAT to observe and score the climbers’ performance. Interrater reliability and comparisons with existing measures of climbing performance (6-mo self-reported ability, success and failure, climbing pace [m·min−1], and geometric entropy) were made. Results: Intraclass correlation coefficient (2,k) for the 4 raters demonstrated excellent reliability (>.81) between observers and good to excellent test–retest reliability (.71–.91). Pearson correlations between self-reported ability and CM-PAT scores explained 61% of the variance in self-reported climbing performance compared with 16% for geometric entropy and 52% for climbing pace. Considering differences in successful and unsuccessful climbers, the CM-PAT (P < .0005; d = 2.14), geometric entropy (P = .014; d = 0.67), and pace (P < .0005; d = 1.88) were able to differentiate between groups. Conclusions: The CM-PAT is the first sport climbing performance observational instrument to be developed through a thorough iterative process with expert coaches. Excellent interrater and test–retest reliability and excellent agreement with self-reported ability and with existing quantitative measures of performance support its recommendation for use in coaching and research contexts. Notably, a key advantage over existing measures is the identification of coachable elements of performance.


2019 ◽  
Vol 23 (4) ◽  
pp. 388-390 ◽  
Author(s):  
Aditi Senthilnathan ◽  
Sree S. Kolli ◽  
Leah A. Cardwell ◽  
Irma Richardson ◽  
Steven R. Feldman ◽  
...  

Background: Hidradenitis suppurativa (HS) is a debilitating dermatologic condition presenting with recurrent abscesses. While there are multiple scales to determine HS severity, none are designed for self-administration. A validated severity self-assessment tool may facilitate survey research and improve communication by allowing patients to objectively report their HS severity between clinic visits. Objectives: The purpose of this study was to assess a self-administered HS measure. Methods: An HS self-assessment tool (HSSA) with 10 photographs of different Hurley stages was developed. The tool was administered to patients diagnosed with HS who visited the Wake Forest Baptist Health dermatology clinic over a span of 2 months. Physician-administered Hurley stage was recorded to determine criterion validity. To assess test-retest reliability of the measure, patients completed the HSSA again at least 30 minutes after the first completion. Results: Twenty-four patients completed the measure, and 20 of these patients completed it twice. Agreement between physician-determined Hurley stage and self-determined Hurley stage was 66.7% with a weighted kappa of 0.57 (95% confidence interval [CI]: 0.30-0.84). The weighted kappa for agreement between patients’ initial and second completion of the HSSA was 0.81 (95% CI: 0.64-0.99). Conclusions: The self-administered measure provides moderate agreement with physician-determined Hurley stage and good test-retest reliability.


BMJ Open ◽  
2020 ◽  
Vol 10 (7) ◽  
pp. e031580
Author(s):  
Cai Le ◽  
Ke Ma ◽  
Pingfen Tang ◽  
David Edvardsson ◽  
Lina Behm ◽  
...  

ObjectiveThis study aims to test a Chinese cross-cultural adaptation of the English version of the Person-Centred Care Assessment Tool (P-CAT) and evaluate its psychometric properties.DesignP-CAT was translated/back-translated using established procedures before the psychometric evaluation of the Chinese version was made.SettingTwo hospitals covering urban and suburban areas of Kunming in the Yunnan province of China.Participants152 female hospital staff completed the survey.Main outcome measure(s)Construct validity and reliability, including internal consistency and test–retest reliability, were assessed among a sample of hospital staff.ResultsThe factor analysis resulted in a two-component solution that consisted of two subscales. The corrected item-total correlations for all of the items ranged from 0.14 to 0.44, with six items not meeting the cut-off level for item-total correlation (>0.3). The Chinese P-CAT demonstrated strong reliability, with a Cronbach’s alpha of 0.91–0.94 for the scales and a test–retest reliability coefficient of 0.88 for the overall scale scores. The intraclass correlation was 0.92 (95% CI 0.90 to 0.95).ConclusionP-CAT appears to be a promising measure for evaluating staff perceptions of person-centredness in Chinese hospital environments. The results show that P-CAT can be a useful tool for improving the quality of healthcare in terms of person-centred care in the Chinese context.


2013 ◽  
Vol 10 (8) ◽  
pp. 1079-1090 ◽  
Author(s):  
Adewale L. Oyeyemi ◽  
James F. Sallis ◽  
Adetoyeje Y. Oyeyemi ◽  
Mariam M. Amin ◽  
Ilse De Bourdeaudhuij ◽  
...  

Background:This study adapted the Physical Activity Neighborhood Environment Scale (PANES) to the Nigerian context and assessed the test-retest reliability and construct validity of the Nigerian version (PANES-N).Methods:A multidisciplinary panel of experts adapted the original PANES to reflect the built and social environment of Nigeria. The adapted PANES was subjected to cognitive testing and test retest reliability in a diverse sample of Nigerian adults (N = 132) from different neighborhood types. Intraclass Correlation Coefficients (ICC) was used to assess test-retest reliability, and construct validity was investigated with Analysis of Covariance for differences in environmental attributes between neighborhoods.Results:Four of the 17 items on the original PANES were significantly modified, 3 were removed and 2 new items were incorporated into the final version of adapted PANES-N. Test-retest reliability was substantial to almost perfect (ICC = 0.62–1.00) for all items on the PANES-N, and residents of neighborhoods in the inner city reported higher residential density, land use mix and safety, but lower pedestrian facilities and aesthetics than did residents of government reserved area/new layout neighborhoods.Conclusion:The PANES-N appears promising for assessing environmental perceptions related to physical activity in Nigeria, but further testing is required to assess its applicability across Africa.


Author(s):  
Louis Jenkins ◽  
Bob Mash ◽  
Anselme Derese

Background: Competency-based education and the validity and reliability of workplace-based assessment of postgraduate trainees have received increasing attention worldwide.Family medicine was recognised as a speciality in South Africa six years ago and a satisfactory portfolio of learning is a prerequisite to sit the national exit exam. A massive scaling up of the number of family physicians is needed in order to meet the health needs of the country.Aim: The aim of this study was to develop a reliable, robust and feasible portfolio assessment tool (PAT) for South Africa.Methods: Six raters each rated nine portfolios from the Stellenbosch University programme, using the PAT, to test for inter-rater reliability. This rating was repeated three months later to determine test–retest reliability. Following initial analysis and feedback the PAT was modified and the inter-rater reliability again assessed on nine new portfolios. An acceptable intra-classcorrelation was considered to be > 0.80.Results: The total score was found to be reliable, with a coefficient of 0.92. For test–retest reliability, the difference in mean total score was 1.7%, which was not statistically significant. Amongst the subsections, only assessment of the educational meetings and the logbook showed reliability coefficients > 0.80.Conclusion: This was the first attempt to develop a reliable, robust and feasible national portfolio assessment tool to assess postgraduate family medicine training in the South African context. The tool was reliable for the total score, but the low reliability of several sections in the PAT helped us to develop 12 recommendations regarding the use of the portfolio, the design of the PAT and the training of raters.


Sign in / Sign up

Export Citation Format

Share Document