Emergency medicine residents have much access to computer technology and possess some computer skills; however, many are unable to produce a usable product or conduct a literature search. Emergency medicine residents have not had sufficient computer training prior to residency. The computer skills of EMRs should be assessed through skills testing rather than self-assessment, and computer training during residency should be improved.
Fifty patients who underwent 60 triple innominate osteotomies were reviewed radiographically and clinically using a modified Harris hip score (HHS). Average patient age was 26 years (range: 13-48 years). At average 9-year follow-up (range: 5-14 years), 12 (20%) hips had been converted to total hip arthroplasty (THA) and 4 (7%) hips had incapacitating pain. Sixteen (27%) hips were considered failures. Average modified HHS at final follow-up was 67 (range: 28-91). Forty-nine (98%) of 50 patients reported they would recommend the procedure to others in the same situation. Radiographically, there was significant improvement in the center-to-edge angle of Wiberg and the acetabular angle of Sharp. There also was a statistically significant relationship between failure of the osteotomy and severity of preexisting hip arthrosis as measured by the Tonnis criteria. The results demonstrate triple innominate osteotomy was effective in eliminating pain, but the fact that 27% of hips required or will require THA indicates results may deteriorate with time.
Abstract. Objective: To describe emergency medicine residents' (EMRs') personal computer (PC) use and educational needs and to compare their perceived and actual PC skills. Methods: This was a prospective, cross-sectional study. Subjects were all EMRs at seven midwestern Accreditation Council for Graduate Medical Education (ACGME) residency programs. The EMRs completed a questionnaire about their PC use and ability to perform 23 tasks derived from two national retail-training programs. The tasks covered word processing, slide making, and Internet use. The EMRs then took a three-part test performing the skills in the questionnaire. Two independent raters scored the tests. Frequencies with 95% confidence intervals (95% CIs) were calculated for categorical data. Positive and negative predictive values were used to report information comparing residents' performance with their self-assessment of skills. Cohen's kappa was used to test agreement between raters. Results: One hundred twenty-four of 158 (79%) eligible EMRs participated. Since not all participants engaged in all parts of the study, the sample size varies between 121 and 124. One hundred one of 122 (83%; 95% CI = 75 to 89) owned a PC. The EMRs use home PCs a mean of 3.8 hours/week for physician duties and use residency PCs 1.9 hours/week (range 0-20). Ninety-six of 122 (79%; 95% CI = 70 to 86) EMRs reported no formal PC training during residency. Thirty-five percent (43/122; 95% CI = 27 to 44) passed the word-processing test and 50% (62/123; 95% CI = 41 to 60) passed the slide-making test. Reasons for failure were because of errors and not having a presentable product. Thirty-eight of 122 (31%; 95% CI = 23 to 40) failed the literature search, including 33 who said they could perform it. One hundred fifteen of 123 (94%; 95% CI = 88 to 98) EMRs were able to find an Internet address, including ten who stated they could not. Twenty-one percent of the residents who attempted any test (26/124; 95% CI = 14 to 29) passed all three tests. There was no association between year of training and success on the tests (p = 0.374). Thirty-seven of 115 (32%; 95% CI = 24 to 42) EMRs said they had insufficient PC training to meet their physician needs. Conclusions: Emergency medicine residents have much access to computer technology and possess some computer skills; however, many are unable to produce a usable product or conduct a literature search. Emergency medicine residents have not had sufficient computer training prior to residency. The computer skills of EMRs should be assessed through skills testing rather than self-assessment, and computer training during residency should be improved.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.