Society for Academic Emergency Medicine Annual Meeting
Transcription
Society for Academic Emergency Medicine Annual Meeting
SOCIETY FOR ACADEMIC EMERGENCY MEDICINE ANNUAL MEETING ABSTRACTS - 2012 The editors of Academic Emergency Medicine (AEM) are honored to present these abstracts accepted for presentation at the 2012 annual meeting of the Society for Academic Emergency Medicine (SAEM), May 9 to 12 in Chicago, Illinois. These abstracts represent countless hours of labor, exciting intellectual discovery, and unending dedication by our specialty’s academicians. We are grateful for their consistent enthusiasm, and are privileged to publish these brief summaries of their research. This year, SAEM received 1172 abstracts for consideration, and accepted 746. Each abstract was independently reviewed by up to six dedicated topic experts blinded to the identity of the authors. Final determinations for scientific presentation were made by the SAEM Program Scientific Subcommittee co-chaired by Ali S. Raja, MD, MBA, MPH and Steven B. Bird, MD, and the SAEM Program Committee, chaired by Michael L. Hochberg, MD. Their decisions were based on the final review scores and the time and space available at the annual meeting for oral and poster presentations. There were also 125 Innovation in Emergency Medicine Education (IEME) abstracts submitted, of which 37 were accepted. The IEME Subcommittee was co-chaired by JoAnna Leuck, MD and Laurie Thibodeau, MD. We present these abstracts as they were received, with minimal proofreading and copy editing. Any questions related to the content of the abstracts should be directed to the authors. Presentation numbers precede the abstract titles; these match the listings for the various oral and poster sessions at the annual meeting in Chicago, as well as the abstract numbers (not page numbers) shown in the key word and author indexes at the end of this supplement. All authors attested to institutional review board or animal care and use committee approval at the time of abstract submission, when relevant. Abstracts marked as ‘‘late-breakers’’ are prospective research projects that were still in the process of data collection at the time of the December abstract deadline, but were deemed by the Scientific Subcommittee to be of exceptional interest. These projects will be completed by the time of the annual meeting; data shown here may be preliminary or interim. On behalf of the editors of AEM, the membership of SAEM, and the leadership of our specialty, we sincerely thank our research colleagues for these contributions, and their continuing efforts to expand our knowledge base and allow us to better treat our patients. David C. Cone, MD Editor-in-Chief Academic Emergency Medicine 1 Policy-driven Improvements In Crowding: System-level Changes Introduced By A Provincial Health Authority And Its Impact On Emergency Department Operations In 15 Centers Grant Innes1, Andrew McRae1, Brian Holroyd2, Brian Rowe2, Christian Schmid3, MingFu Liu3, Lester Mercuur1, Nancy Guebert3, Dongmei Wang3, Jason Scarlett3, Eddy S. Lang1 1 University of Calgary, Calgary, AB, Canada; 2 University of Alberta, Edmonton, AB, Canada; 3 Alberta Health Services, Calgary, AB, Canada Background: System-level changes that target both ED throughput and output show the most promise in alleviating crowding. In December 2010, Alberta Health Services (AHS) implemented a province-wide hospital overcapacity protocol (OCP) structured upon the Viccellio model. S4 ISSN 1069-6563 PII ISSN 1069-6563583 Objectives: We sought to determine if the OCP policy resulted in a meaningful and sustained improvement in ED throughput and output metrics. Methods: A prospective pre-post experimental study was conducted using administrative data from 15 community and tertiary centers across the province. The study phases consisted of the 8 months from February to September 2010 compared against the same months in 2011. Operational data for all centres were collected through the EDIS tracking systems used in the province. The OCP included 3 main triggers: ED bed occupancy >110%, at least 35% of ED stretchers blocked by patients awaiting inpatient bed or disposition decision, and no stretcher available for high acuity patients. When all criteria were met, selected boarded patients were moved to an inpatient unit (non-traditional care space if no bed available). The primary outcome was ED length of stay (LOS) for admitted patients. The ED load of boarded patients from 10–11 am was reported ª 2012 by the Society for Academic Emergency Medicine doi: 10.1111/j.1553-2712.2012.01332.x ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 in patient-hours (pt-hrs). Throughput is reported as time from ED arrival to physician assessment and percent left without being seen (LWBS). Continuous variables were compared with the Student’s t-test. Results: The volume of ED patients across all sites increased by 6.3% from the pre phase to the post phase (579071 vs. 615787; p < 0.001) while admission rates remained constant (12.9% vs. 13.1%; p = NS). ED LOS for admitted patients decreased from 17.2 hours to 11.6 hours (p < 0.001); the load of admitted patients at 10 am declined from 11.3 pt-hrs to 6.1 (p < 0.001). Average time from ED arrival to physician assessment decreased in the post phase (113.2 vs. 99.3 minutes; p < 0.001) as did % LWBS (4.0% vs 3.8%; p < 0.001). All OCP effects remained constant over time; however, there were regional disparities in its impact. Conclusion: Policy-driven changes in ED and hospital operations were associated with significant improvements in both throughput and output metrics, despite increased input. Which components of the OCP program had the greatest impact are uncertain, as are explanations for the differential regional impact. 2 Prevalence of Non-convulsive Seizure and Other Electroencephalographic Abnormalities In Emergency Department Patients With Altered Mental Status Shahriar Zehtabchi1, Arthur C. Grant2, Samah G. Abdel Baki3, Omurtag Ahmet3, Richard Sinert1, Geetha Chari2, Shweta Malhotra1, Jeremy Weedon4, Andre A. Fenton5 1 Department of Emergency Medicine, State University of New York, Downstate Medical Center, Brooklyn, NY; 2Department of Neurology, State University of New York, Downstate Medical Center, Brooklyn, NY; 3Biosignal Group Inc., Brooklyn, NY; 4Scientific Computing Center, State University of New York, Downstate Medical Center, Brooklyn, NY; 5Center for Neural Science, New York University, New York, NY Background: Two to ten percent of patients evaluated in the emergency departments (ED) present with altered mental status (AMS). The prevalence of non-convulsive seizure (NCS) and other electroencephalographic (EEG) abnormalities in this population is not known. This information is needed to make recommendations regarding the routine use of emergent EEG in AMS patients. Objectives: To identify the prevalence of NCS and other EEG abnormalities in ED patients with AMS. Methods: An ongoing prospective study at two academic urban ED. Inclusion: Patients ‡ 13 years old with AMS. Exclusion: An easily correctable cause of AMS (e.g. hypoglycemia, opioid overdose). A 30-minute EEG with the standard 19 electrodes was performed on each subject as soon as possible after presentation (usually within 1 hour). Outcome: The rate of EEG abnormalities based on blinded review of all EEGs by two boardcertified epileptologists. Descriptive statistics are used to report EEG findings. Frequencies are reported as percentages with 95% confidence intervals (CI), and inter-rater variability is reported with kappa. • www.aemj.org S5 Results: The interim analysis was performed on 130 consecutive patients (target sample size: 260) enrolled from May to October 2011 (median age: 61, range 13–100, 40% male). EEGs for 20 patients were reported uninterpretable by at least one rater (6 by both raters). Of the remaining 110, only 30 (27%, 95%CI 20–36%) were normal according to either rater (n = 15 by both). The most common abnormality was background slowing (n = 75, 68%, 95%CI 59–76%) by either rater (n = 47 by both), indicating underlying encephalopathy. NCS was diagnosed in 8 patients (7%, 95%CI, 4–14%) by at least one rater (n = 4 by both), including 6 (5%, 95%CI 2–12%) patients in non-convulsive status epilepticus (NCSE). 29 patients (26%,95%CI 19–35%) had interictal epileptiform discharges read by at least one rater (n = 12 by both) indicating cortical irritability and an increased risk of spontaneous seizure. Inter-rater reliability for EEG interpretations was modest (kappa: 0.53, 95%CI 0.39–0.67). Conclusion: ED patients with AMS have a high prevalence of EEG abnormalities, including encephalopathy and NCS. ED physicians should have a high index of suspicion for such pathologies in AMS patients. EEG is necessary to make the diagnosis of NCS/NCSE, for which early treatment could significantly reduce morbidity. (Originally submitted as a ‘‘late-breaker.’’) 3 RNA Transcriptional Profiling for Diagnosis of Serious Bacterial Infections (SBIs) in Young Febrile Infants P. Mahajan1, N. Kuppermann2, A. Mejias3, D. Chaussabel4, T. Casper5, B. Dimo6, H. Gramse5, O. Ramilo6 1 Children’s Hospital of Michigan, Detroit, MI; 2 University of California, Davis School of Medicine, Davis, CA; 3Nationwide Childrens Hospital, Columbus, OH; 4Benaroya Research Institute, Seattle, WA; 5University of Utah, Salt Lake City, UT; 6Nationwide Children’s Hospital, Columbus, OH Background: Genomic technologies allow us to determine the etiology of infection by evaluating specific host responses (RNA biosignatures) to different pathogens and have the potential to replace cultures of relevant body fluids as the reference standard. Objectives: To define diagnostic SBI and non-bacterial (non-SBI) biosignatures using RNA microarrays in febrile infants presenting to emergency departments (EDs). Methods: We prospectively collected blood for RNA microarray analysis in addition to routine screening tests including white blood cell (WBC) counts, urinalyses, cultures of blood, urine, and cerebrospinal fluid, and viral studies in febrile infants 60 days of age in 22 EDs (2008–09). We defined SBI as bacteremia, urinary tract infection (UTI), or bacterial meningitis. We used class comparisons (Mann-Whitney p < 0.01, Benjamini for MTC and 1.25 fold change filter), modular gene analysis, and K-NN algorithms to define and validate SBI and non-SBI biosignatures in a subset of samples. Results: 81% (939/1162) of febrile infants were evaluated for SBI. 6.8% (64/939) had SBI (14 (1.5%) bac- S6 2012 SAEM ANNUAL MEETING ABSTRACTS teremia, 56 (6.0%) UTIs, and 4 (0.4%) bacterial meningitis). Infants with SBIs had higher mean temperatures, and higher WBC, neutrophil, and band counts. We analyzed RNA biosignatures on 141 febrile infants: 35 SBIs (2 meningitis, 5 bacteremia, 28 UTI), 106 non-SBIs (49 influenza, 29 enterovirus, 28 undefined viral infections), and 11 healthy controls. Class comparisons identified 1,288 differentially expressed genes between SBIs and non-SBIs. Modular analysis revealed overexpression of interferon related genes in non-SBIs and inflammation related genes in SBIs. 232 genes were differently expressed (p < 0.01) in each of the three non-SBI groups vs SBI group. Unsupervised cluster analysis of these 232 genes correctly clustered 91% (128/141) of non-SBIs and SBIs. K-NN algorithm identified 33 discriminatory genes in training set (30 non-SBIs vs 17 SBIs) which classified an independent test (76 non-SBIs vs 18 SBIs) with 87% accuracy. Four misclassified SBIs had over-expression of interferon-related genes, suggesting viral-bacterial co-infections, which was confirmed in one patient. Conclusion: Analysis on this initial sample set confirms the potential of RNA biosignatures for distinguishing young febrile infants with SBIs vs those without bacterial illness. Funded by HRSA/MCHB grant H34MC16870. 4 Saving Maternal, Newborn, and Child Lives in Developing Countries: Evaluation of a Novel Training Package among Frontline Health Workers in South Sudan Brett D. Nelson, Roy Ahn, Maya Fehling, Melody J. Eckardt, Kathryn L. Conn, Alaa El-Bashir, Margaret Tiernan, Thomas F. Burke Massachusetts General Hospital, Boston, MA Background: Improving maternal, newborn, and child health (MNCH) is a leading priority worldwide. However, limited frontline health care capacity is a major barrier to improving MNCH in developing countries. Objectives: We sought to develop, implement, and evaluate an evidence-based Maternal, Newborn, and Child Survival (MNCS) package for frontline health workers (FHWs). We hypothesized that FHWs could be trained and equipped to manage and refer the leading MNCH emergencies. Methods: SETTING - South Sudan, which suffers from some of the world’s worst MNCH indices. ASSESSMENT/INTERVENTION - A multi-modal needs assessment was conducted to develop a best-evidence package comprised of targeted trainings, pictorial checklists, and reusable equipment and commodities (Figure 1). Program implementation utilized a trainingof-trainers model. EVALUTION - 1) Pre/post knowledge assessments, 2) pre/post objective structured clinical examinations (OSCEs), 3) focus group discussions, and 4) closed-response questionnaires. Results: Between Nov 2010 to Oct 2011, 72 local trainers and 708 FHWs were trained in 7 of the 10 states in South Sudan. Knowledge assessments among trainers (n = 57) improved significantly from 62.7% (SD 20.1) to 92.0% (SD 11.8) (p < 0.001). Mean scores a maternal OSCE and a newborn OSCE pre-training, immediately post-training, and upon 2–3 month follow-up are shown in the table. Closed-response questionnaires with 54 FHWs revealed high levels of satisfaction, use, and confidence with MNCS materials. Participants reported an average of 3.0 referrals (range 0–20) to a higher level of care in the 2–3 months since training. Furthermore, 78.3% of FHWs were more likely to refer patients as a result of the training program. During seven focus group discussions with trained FHWs, respondents (n = 41) reported high satisfaction with MNCS trainings, commodities, and checklists, with few barriers to implementation or use. Conclusion: These findings suggest MNCS has led to improvements in South Sudanese FHWs’ knowledge, skills, and referral practices with respect to appropriate management of MNCH emergencies. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Table - Abstract 4: FHW skills tests before training, after training, and at 2–3 months follow-up Maternal OSCE (n = 55) Newborn OSCE (n = 54) 5 Pre-training mean % (SD) Post-training mean % (SD) Follow-up mean % (SD) 21.1 (13.8) 83.4 (21.5) 61.5 (25.8) 41.6 (16.5) 89.8 (14.0) 45.7 (23.1) Whole Blood Lactate Kinetics in Patients Undergoing Quantitative Resuscitation for Septic Shock Alan E. Jones1, Michael Puskarich1, Stephen Trzeciak2, Nathan Shapiro3, Jeffrey Kline4 1 University of Mississippi Medical Center, Jackson, MS; 2Cooper University Hospital, Camden, NJ; 3BIDMC, Boston, MA; 4Carolinas Medical Center, Charlotte, NC Background: Previous studies have suggested lactate clearance as an endpoint of early sepsis resuscitation. No study has compared various lactate measurements to determine the optimal parameter to target. Objectives: To compare the association of blood lactate kinetics with survival in patients with septic shock undergoing early quantitative resuscitation. Methods: Preplanned analysis of a multicenter EDbased RCT of early sepsis resuscitation targeting three physiological variables: CVP, MAP, and either central venous oxygen saturation or lactate clearance. Inclusion criteria: suspected infection, two or more SIRS criteria, and either SBP <90 mmHg after a fluid bolus or lactate >4 mmol/L. All patients had an initial lactate measured with repeat at two hours. Normalization of lactate was defined a lactate decline to <2.0 mmol/L in a patient with an intial lactate ‡2.0. Absolute lactate clearance (initial - delayed value), and relative ((absolute clearance)/(initial value)*100) were calculated if the initial lactate was ‡2.0. The outcome was in-hospital survival. Receiver operating characteristic curves were constructed and areas under the curve (AUC) were calculated. Difference in proportions of survival between the two groups at different lactate cutoffs were analyzed using 95% CI and Fisher exact tests. Results: Of 272 included patients, the median initial lactate was 3.1 mmol/L (IQR 1.7, 5.8), and the median absolute and relative lactate clearance were 1 mmol/L (IQR 0.3, 2.5) and 37% (IQR 14, 57). An initial lactate >2.0 mmol/L was seen in 187/272 (69%), and 68/187 (36%) patients normalized their lactate. Overall mortality was 19.7%. AUCs for initial lactate, relative lactate clearance, and absolute lactate clearance were 0.70, 0.69, and 0.58. Lactate normalization best predicted survival (OR 6.1, 95% CI 2.2–21), followed by lactate clearance of 50% (OR 4.3, 95% CI 1.8–10.3), initial lactate of <2 mmol/L (OR 3.4, 95% CI 1.5–7.8), and initial lactate <4 mmol/L (OR 2.3, 95% CI 1.3–4.3), with lactate clearance of 10% not reaching significance (OR 2.3, 95% CI 0.96–5.6). Conclusion: In ED sepsis patients undergoing early resuscitation, normalization of lactate during resuscitation was more strongly associated with survival than • www.aemj.org S7 any absolute value or absolute/relative change in lactate. Further studies should address if targeting lactate normalization leads to improved outcomes. 6 A Comparison of Cosmetic Outcomes of Lacerations of the Trunk and Extremity Repaired Using Absorbable Versus Nonabsorbable Sutures Cena Tejani1, Adam Sivitz1, Michael Rosen1, Albert Nakanishi2, Robert Flood2, Matthew Clott1, Paul Saccone1, Raemma Luck3 1 Newark Beth Israel Hospital, Newark, NJ; 2 Cardinal Glennon Children’s Medical Center, St. Louis, MO; 3Saint Christopher’s Hospital for Children, Philadelphia, PA Background: Although prior studies have compared the use of absorbable versus nonabsorbable sutures for traumatic lacerations, most of these studies have focused on facial lacerations. A review of the literature indicates that there are no randomized prospective studies to date that have looked exclusively at the cosmesis of absorbable sutures on trunk and extremity lacerations that present in the ED. The use of absorbable sutures in the ED setting confers several advantages: patients do not need to return for suture removal which results in a reduction in ED crowding, ED wait times, missed work or school days, and stressful procedures (suture removal) for children. Objectives: The primary objective of this study is to compare the cosmetic outcome of trunk and extremity lacerations repaired using absorbable versus nonabsorbable sutures in children and adults. A secondary objective is to compare complication rates between the two groups. Methods: Eligible patients with lacerations were randomly allocated to have their wounds repaired with Vicryl Rapide (absorbable) or Prolene (nonabsorbable) sutures. At a 10 day follow-up visit the wounds were evaluated for infection and dehiscence. After 3 months, patients were asked to return to have a photograph of the wound taken. Two blinded plastic surgeons using a previously validated 100 mm visual analogue scale (VAS) rated the cosmetic outcome of each wound. A VAS score of 15 mm or greater was considered to be a clinically significant difference. Results: Of the 100 patients enrolled, 45 have currently completed the study including 19 in the Vicryl Rapide group and 26 in the Prolene group. There were no significant differences in the age, race, sex, length of wound, number of sutures, or layers of repair in the two groups. The observer’s mean VAS for the Vicryl Rapide group was 55.76 mm (95%CI 41.95–69.57) and that for the Prolene group was 55.9 mm (95%CI 44.77– 67.03), resulting in a mean difference of 0.14 mm (95%CI–16.95 to 17.23, p = .98). There were no significant differences in the rates of infection, dehiscence, or keloid formation between the two groups. Conclusion: The use of Vicryl Rapide instead of nonabsorbable sutures for the repair of lacerations on the trunk and extremities should be considered by emergency physicians as it is an alternative that provides a similar cosmetic outcome. S8 7 2012 SAEM ANNUAL MEETING ABSTRACTS Minimally Invasive Burn Care: A Report Of Six Clinical Studies Of Rapid And Selective Debridement Using A Bromelain-based Debriding Gel Dressing Lior Rosenberg1, Yuval Krieger1, Eldad Silberstein1, Alex Bogdanov-Berezovsky1, Ofer Arnon1, Yaron Shoam1, Nir Rosenberg1, Amiram Sagi1, Keren David2, Guy Rubin3, Adam J. Singer4 1 Ben-Gurion University of the Negev, Beer-Sheva, Israel; 2MediWound Ltd, Beer-Sheva, Israel; 3 Haemek Hospital, Afula, Israel; 4Stony Brook University, Stony Brook, NY Background: Burns are characterized by an eschar that delays diagnosis and increases the risk of infection and scarring. As a result, surgical excision of the eschar is a cornerstone of care requiring specialized personnel and facilities. Objectives: A novel debriding agent that could be used by emergency physicians (EPs) was developed to overcome the weaknesses of surgical and conventional enzymatic debridement. We hypothesized that the novel agent would reduce the need for surgical excision and skin grafting compared with standard care (SC). Methods: The safety and efficacy of a novel Debriding Gel Dressing (DGD) was determined in six studies; five were RCTs. Treatments (DGD, control vehicle, or SC) were randomly allocated to deep partial and full thickness burns covering less than 30% TBSA by a computer-generated randomization scheme. Primary endpoints were percentage eschar debridement, rate of surgical burn excision and total area excised. Efficacy analyses were intention to treat. Results: 518 patients were enrolled. Percentage eschar debridement was greater than 90% in all studies for DGD, which was comparable to SC, and significantly greater than control vehicle, which was negligible. In the third study, the total area surgically excised was significantly less in DGD-treated patients compared with patients treated with the control vehicle (22.9% vs. 73.2%, P < 0.001) or SC (50.5%, P = 0.006). In the sixth, phase III RCT the rate of surgical excision was significantly lower in DGD-treated patients than in control patients treated with SC (40/163 [24.5%] vs. 119/170 [70.0%], P < 0.001). The total area surgically excised was also significantly less in DGD-treated patients compared with patients treated with SC (13.1% vs. 56.7%, P < 0.001). Local and systemic adverse events were similar for DGD and SC. Conclusion: DGD is a safe and effective method of burn debridement that offers an alternative, minimally invasive burn care modality to standard surgical excision that could be used by EPs. 8 The Golden Period of Laceration Repair Has Disappeared James Quinn1, Michael Kohn2, Steven Polevoi2 1 Stanford University, Stanford, CA; 2University of California, San Francisco, San Francisco, CA Background: Much has been written about the ‘‘golden period’’ for lacerations and that ‘‘older’’ wounds are at increased risk of infection. Objectives: To determine the relationship between infection and time from injury to closure, and the characteristics of lacerations closed before and after 12 hours of injury. Methods: Over an 18 month period, a prospective multi-center cohort study was conducted at a teaching hospital, trauma center and community hospital. Emergency physicians completed a structured data form when treating patients with lacerations. Patients were followed to determine whether they had suffered a wound infection requiring treatment and to determine a cosmetic outcome rating. We compared infection rates and clinical characteristics of lacerations with chisquare and t-tests as appropriate. Results: There were 2663 patients with lacerations; 2342 had documented times from injury to closure. The mean times from injury to repair for infected and noninfected wounds were 2.4 vs. 3.0 hrs (p = 0.39) with 78% of lacerations treated within 3 hours and 4% (85) treated 12 hours after injury. There were no differences in the infection rates for lacerations closed before (2.9%, 95%CI 2.2–3.7) or after (2.1%, 95%CI 0.4–6.0) 6 hours and before (3.0%, 95% CI 2.3%–3.8%) or after (1.2%, 95% CI 0.03%–6.4%) 12 hours. The patients treated 12 hours after injury tended to be older (41 vs. 34 yrs p = 0.02) and fewer were treated with primary closure (85% vs. 96% P < 0.0001). Comparing wounds 12 or more hours after injury with more recent wounds, there was no effect of location on decision to close. Wounds closed after 12 hours did not differ from wounds closed before 12 hours with respect to use of prophylactic antibiotics, type of repair, length of laceration, or cosmetic outcome. Conclusion: Closing older lacerations, even those greater than 12 hours after injury, does not appear to be associated with any increased risk of infection or adverse outcomes. Excellent irrigation and decontamination over the last 30 years may have led to this change in outcome. 9 The Effects of a Novel TGF-beta Antagonist on Scarring in a Vertical Progression Porcine Burn Model Adam J. Singer1, Steve A. McClain1, Ji-Hun Kang1, Shuan S. Huang2, Jung S. Huang3 1 Stony Brook University, Stony Brook, NY; 2 Auxagen, Inc., St Louis, MO; 3St Louis University, St Louis, MO Background: Deep burns may result in significant scarring leading to aesthetic disfigurement and functional disability. TGF-b is a growth factor that plays a significant role in wound healing and scar formation. Objectives: The current study was designed to test the hypothesis that a novel TGF-b antagonist would reduce scar contracture compared with its vehicle in a porcine partial thickness burn model. Methods: Ninety-six mid-dermal contact burns were created on the backs and flanks of four anesthetized young swine using a 150 gm aluminum bar preheated to 80 Celsius for 20 seconds. The burns were randomized to treatment with topical TGF-b antagonist at one of three concentrations (0, 187, and 375 lL) in replicates of ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 8 in each pig. Dressing changes and reapplication of the topical therapy were performed every 2 days for 2 weeks then twice weekly for an additional 2 weeks. Burns were photographed and full thickness biopsies were obtained at 5, 7, 9, 14, and 28 days to determine reepithelialization and scar formation grossly and microscopically. A sample of 32 burns in each group had 80% power to detect a 10% difference in percentage scar contracture. Results: A total of 32 burns were created in each of the three study groups. Burns treated with the high dose TGF-b antagonist healed with less scar contracture than those treated with the low dose and control (52 ± 20%, 63 ± 15%, and 62 ± 14%; ANOVA P = 0.02). Additionally, burns treated with the higher, but not the lower dose of TGF-b antagonist healed with significantly fewer full thickness scars than controls (62.5% vs. 100% vs. 93.8% respectively; P < 0.001). There were no infections and no differences in the percentage wound reepithelialization among all study groups at any of the time points. Conclusion: Treatment of mid-dermal porcine contact burns with the higher dose TGF-b antagonist reduced scar contracture and rate of deep scars compared with the low dose and controls. A Double-Blinded Comparison of Insulin Regimens in Diabetic Ketoacidosis: Does Bolus Insulin Make a Difference? Joel Kravitz1, Patricia Giraldo2, Rika N. O’Malley3, Elizabeth Aguilera3, Claudia Lares3, Sorin Cadar3 1 Community Medical Center, Toms River, Southampton, NJ; 2Albert Einstein Medical Center, Philadelphia, PA, NJ; 3Albert Einstein Medical Center, Philadelphia, PA Background: Diabetic ketoacidosis (DKA) is a common and lethal complication of diabetes. The American Diabetes Association recommends treating adult patients with a bolus dose of regular insulin followed by a continuous insulin infusion. The ADA also suggests a glucose correction rate of 75–100 mg/dl/hr to minimize complications. Objectives: Compare the effect of bolus dose insulin therapy with insulin infusion to insulin infusion alone on serum glucose, bicarbonate, and pH in the initial treatment of DKA. Methods: Consecutive DKA patients were screened in the ED between March ’06 and June ’10. Inclusion criteria were: age >18 years, glucose >350 mg/dL, serum bicarbonate 15 or ketonemia or ketonuria. Exclusion criteria were: congestive heart failure, current hemodialysis, pregnancy, or inability to consent. No patient was enrolled more than once. Patients were randomized to receive either regular insulin 0.1 units/kg or the same volume of normal saline. Patients, medical and research staff were blinded. Baseline glucose, electrolytes, and venous blood gases were collected on arrival. Bolus insulin or placebo was then administered and all enrolled patients received regular insulin at rate of 0.1 unit/kg/hr, as well as fluid and potassium repletion per the research protocol. Glucose, electrolytes, and www.aemj.org S9 venous blood gases were drawn hourly for 4 hours. Data between two groups were compared using unpaired t-test. Results: 99 patients were enrolled, with 30 being excluded. 35 patients received bolus insulin; 34 received placebo. No significant differences were noted in initial glucose, pH, bicarbonate, age, or weight between the two groups. After the first hour, glucose levels in the insulin group decreased by 151 mg/dL compared to 94 mg/dL in the placebo group (p = 0.0391, 95% CI 2.7 to 102.0). Changes in mean glucose levels, pH, bicarbonate level, and AG were not statistically different between the two groups for the remainder of the 4 hour study period. There was no difference in the incidence of hypoglycemia in the two groups. Conclusion: Administering a bolus dose of regular insulin decreased mean glucose levels more than placebo, although only for the first hour. There was no difference in the change in pH, serum bicarbonate or anion gap at any interval. This suggests that bolus dose insulin may not add significant benefit in the emergency management of DKA. 11 10 • Calibration Of APACHE II Score To Predict Mortality In Out-of-hospital And In-hospital Cardiac Arrest Justin D. Salciccioli, Cristal Cristia, Andre Dejam, Brandon Giberson, David Toomey, Michael N. Cocchi, Michael W. Donnino BIDMC Center for Resuscitation Science, Boston, MA Background: Severity of illness scores can predict outcomes in critically ill patients. However, the calibration of previously established scores in post-cardiac arrest is poorly established. Objectives: To assess the calibration of the Acute Physiology and Chronic Health Evaluation (APACHE II) score in out-of-hospital (OHCA) and in-hospital cardiac arrest (IHCA). Methods: We performed a prospective observational study of adult cardiac arrest at an urban tertiary care hospital during the period from 12/2007 to 12/2010. Inclusion criteria: 1. Adult (>18 years); 2. OHCA or S10 2012 SAEM ANNUAL MEETING ABSTRACTS IHCA; 3. Return of spontaneous circulation (RSOC). Traumatic cardiac arrests were excluded. We recorded baseline demographics, arrest event characteristics, follow-up vitals and laboratory data, and in-hospital mortality. APACHE II scores were calculated at the time of ROSC, and at 24 hrs, 48 hrs, and 72 hrs. We used simple descriptive statistics to describe the study population. Univariate logistic regression was used to predict mortality with APACHE II as a continuous predictor variable. Discrimination of APACHE II scores was assessed using the area under the curve (AUC) of the receiver operator characteristic (ROC) curve. Results: A total of 229 patients were analyzed. The median age was 70 years (IQR: 56–79) and 32% were female. APACHE II score was a significant predictor of mortality for both OHCA and IHCA at baseline and at all follow-up time points (all p < 0.01). Discrimination of the score increased over time and achieved very good discrimination after 24 hrs (Table, Figure). Conclusion: The ability of APACHE II score to predict mortality improves over time in the 72 hours following cardiac arrest. These data suggest that after 24 hours, APACHE II scoring is a useful severity of illness score in all post-cardiac arrest patients. Table - Abstract 11: Comparison of APACHE II with AUC to Predict Mortality hyperlactatemia (lactate ‡ 4.0 mmol/L), a logistic regression model was created; outcome- hyperlactatemia; primary variable of interest- hyperglycemia. A second model was created to determine if concurrent hyperlactatemia affects hyperglycemia’s association with mortality; outcome- 28-day mortality; primary risk variablehyperglycemia with an interaction term for concurrent hyperlactatemia. Both models were adjusted for demographics, comorbidities, presenting infectious syndrome, and objective evidence of renal, respiratory, hematologic, or cardiovascular dysfunction. Results: 1236 ED patients were included; mean age 76 ± 19 years. 133 (9%) subjects were hyperglycemic, 182 (13%) hyperlactatemic, and 225 (16%) died within 28 days of the initial ED visit. After adjustment, hyperglycemia was significantly associated with simultaneous hyperlactatemia (OR 3.9, 95%CI 2.48, 5.98). Hyperglycemia with concurrent hyperlactatemia was associated with increased mortality risk (OR 4.4, 95%CI 2.27, 8.59), but hyperglycemia in the absence of simultaneous hyperlactatemia was not (OR 0.86, 95%CI 0.45, 1.65). Conclusion: In this cohort of septic adult non-diabetic patients, mortality risk did not increase with hyperglycemia unless associated with simultaneous hyperlactatemia. The previously reported association of hyperglycemia with mortality in this population may be due to the association of hyperglycemia with hyperlactatemia. APACHE II Score (AUC) All-CA OHCA IHCA 12 0 hr 24 hr 48 hr 72 hr Mortality n (%) 28 (0.67) 26 (0.66) 28 (0.71) 29 (0.71) 29 (0.72) 29 (0.69) 26 (0.81) 26 (0.82) 26 (0.81) 22 (0.84) 23 (0.89) 22 (0.79) 124 (54) 69 (56) 55 (44) Hyperlactatemia Affects the Association of Hyperglycemia with Mortality in Non-Diabetic Septic Adults Jeffrey P. Green1, Tony Berger1, Nidhi Garg2, Alison Suarez2, Michael S. Radeos2, Sanjay Gupta2, Edward A. Panacek1 1 UC Davis Medical Center, Davis, CA; 2New York Hospital Queens, Flushing, NY Background: Admission hyperglycemia has been described as a mortality risk factor for septic non-diabetics, but the known association of hyperglycemia with hyperlactatemia (a validated mortality risk factor in sepsis) has not previously been accounted for. Objectives: To determine whether the association of hyperglycemia with mortality remains significant when adjusted for concurrent hyperlactatemia. Methods: This was a post-hoc, nested analysis of a single-center cohort study. Providers identified study subjects during their ED encounters; all data were collected from the electronic medical record. Patients: Nondiabetic adult ED patients with a provider-suspected infection, two or more Systemic Inflammatory Response Syndrome criteria, and concurrent lactate and glucose testing in the ED. Setting: The ED of an urban teaching hospital; 2007 to 2009. Analysis: To evaluate the association of hyperglycemia (glucose >200 mg/dL) with 13 The Effect of Near Infrared Spectroscopy Monitoring on Patients Undergoing Resuscitation for Shock James Miner, Johanna Bischoff, Nathaniel Scott, Roma Patel, Rebecca Nelson, Stephen W. Smith Hennepin County Medical Center, Minneapolis, MN Background: Near infrared spectroscopy (StO2) represents a measure of perfusion that provides the treating physician with an assessment of a patient’s shock state and response to therapy. It has been shown to correlate with lactate and acid/base status. It is not known if using information from this monitor to guide resuscitation will result in improved patient outcomes. Objectives: To compare the resuscitation of patients in shock when the StO2 monitor is or is not being used to guide resuscitation. Methods: This was a prospective study of patients undergoing resuscitation in the ED for shock from any cause. During alternating 30 day periods, physicians were blinded to the data from the monitor followed by 30 days in which physicians were able to see the information from the StO2 monitor and were instructed to resuscitate patients to a target StO2 value of 75. Adult patients (age>17) with a shock index (SI) of >0.9 (SI = heart rate/systolic blood pressure) or a blood pressure <80 mmHg systolic who underwent resuscitation were enrolled. Patients had a StO2 monitor placed on the thenar eminence of their least-injured hand. Data from the StO2 monitor were recorded continuously and noted every minute along with blood pressure, heart rate, and ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 oxygen saturation. All treatments were recorded. Patients’ charts were reviewed to determine the diagnosis, ICU-free days in the 28 days after enrollment, inpatient LOS, and 28-day mortality. Data were compared using Wilcoxon rank sum and chi-square tests. Results: 107 patients were enrolled, 51 during blinded periods and 56 during unblinded periods. The median presenting shock index was 1.24 (range 0.5 to 4.0) for the blinded group and 1.10 (0.5–3.3) for the unblinded group (p = 0.13). The median time in department was 70 minutes (range 22–407) for the blinded and 76 minutes (range 11–275) for the unblinded groups (p = 0.99). The median hospital LOS was 1 day (range 0–30) for the blinded group, and 2 days (range 0–23) in the unblinded group (p = 0.63). The mean ICU-free days was 22 ± 9 for the blinded group and 19 ± 11 for the unblinded group (p = 0.26). Among patients where the physician indicated using the StO2 monitor data to guide patient care, the ICU-free days were 21.4 ± 9 for the blinded group and 16.3 ± 12 for the blinded group (p = 0.06). Conclusion: StO2 monitoring of patients in undifferentiated shock did not demonstrate a difference in hospital LOS or ICU-free days in this preliminary study. (Originally submitted as a ‘‘late-breaker.’’) 14 A Laboratory Study Assessing The Influence Of Flow Rate And Insulation Upon Intravenous Fluid Infusion Temperature Jonathan Studnek1, John Watts1, Steven Vandeventer2, David Pearson1 1 Carolians Medical Center, Charlotte, NC; 2 Mecklenburg EMS Agency, Charlotte, NC Background: Inducing therapeutic hypothermia (TH) using 4C IV fluids in resuscitated cardiac arrest patients has been shown to be feasible and effective. Limited research exists assessing the efficiency of this cooling method. Objectives: The objective was to determine an efficient infusion method for keeping fluid close to 4C upon exiting an IV. It was hypothesized that colder temperatures would be associated with both higher flow rate and insulation of the fluid bag. • www.aemj.org S11 Methods: Efficiency was studied by assessing change in fluid temperature (0C) during the infusion, under three laboratory conditions. Each condition was performed four times using 1 liter bags of normal saline. Fluid was infused into a 1000 mL beaker through 10 gtts tubing. Flow rate was controlled using a tubing clamp and in-line transducer with a flowmeter, while temperature was continuously monitored in a side port at the terminal end of the IV tubing using a digital thermometer. The three conditions included infusing chilled fluid at a rate of 40 mL/min, which is equivalent to 30 mL/kg/hr for an 80 kg patient, 105 mL/min, and 105 mL/min using a chilled and insulated pressure bag. Descriptive statistics and analysis of variance was performed to assess changes in fluid temperature. Results: The average fluid temperatures at time 0 were 3.40 (95% CI 3.12–3.69) (40 mL/min), 3.35 (95% CI 3.25–3.45) (105 mL/min), and 2.92 (95% CI 2.40–3.45) (105 mL/min + insulation). There was no significant difference in starting temperature between groups (p = 0.16). The average fluid temperatures after 100 mL had been infused were 10.02 (95% CI 9.30–10.74) (40 mL/min), 7.35 (95% CI 6.91–7.79) (105 mL/min), and 6.95 (95% CI 6.47–7.43) (105 mL/min + insulation). The higher flow rate groups had significantly lower temperature than the lower flow rate after 100 mL of fluid had been infused (p < 0.001). The average fluid temperatures after 1000 mL had been infused were 16.77 (95% CI 15.96–17.58) (40 mL/min), 11.40 (95% CI 11.18–11.61) (105 mL/min), and 7.75 (95% CI 7.55–7.99) (105 mL/min + insulation). There was a significant difference in temperature between all three groups after 1000 mL of fluid had been infused (p < 0.001). Conclusion: In a laboratory setting, the most efficient method of infusing cold fluid appears to be a method that both keeps the bag of fluid insulated and is infused at a faster rate. 15 Outcomes of Patients with Vasoplegic versus Tissue Dysoxic Septic Shock Sarah Sterling1, Michael Puskarich1, Stephen Trzeciak2, Nathan Shapiro3, Jeffrey Kline4, Alan Jones1 1 University of Mississippi Medical Center, Jackson, MS; 2Cooper University Hospital, Camden, NJ; 3BIDMC, Boston, MA; 4Carolinas Medical Center, Charlotte, NC Background: The current consensus definition of septic shock requires hypotension after adequate fluid challenge or vasopressor requirement. Some patients with septic shock present with hypotension and hyperlactemia >2 mM/L (tissue dysoxic shock), while others have hypotension alone with normal lactate (vasoplegic shock). Objectives: To determine differences in outcomes of patients with tissue dysoxic versus vasoplegic septic shock. Methods: Pre-planned secondary analysis of a large, multi-center RCT. Inclusion criteria: suspected infection, two or more systemic inflammatory response criteria, and systolic blood pressure <90 mmHg after a S12 2012 SAEM ANNUAL MEETING ABSTRACTS fluid bolus. Patients were categorized by presence of vasoplegic or tissue dysoxic shock. Demographics and sequential organ failure assessment (SOFA) scores were evaluated between the groups. The primary outcome was in-hospital mortality. Data were analyzed using t-tests, chi-squared test, and proportion differences with 95% confidence intervals as appropriate. Results: A total of 242 patients were included: 89 patients with vasoplegic shock and 153 with tissue dysoxic shock. There were no significant differences in age (61 vs. 58 years), Caucasian race (53% vs. 58%), or male sex (57% vs. 52%) between the dysoxic shock and vasoplegic shock groups, respectively. The group with vasoplegic shock had a lower initial SOFA score than did the group with tissue dysoxic shock (5.7 vs. 7.3 points, p = 0.0002). The primary outcome of in-hospital mortality occurred in 8/89 (9%) of patients with vasoplegic shock compared to 40/153 (26%) in the group with tissue dysoxic shock (proportion difference 17%, 95% CI 7–26%, p < 0.0001). Conclusion: In this analysis of patients with septic shock, we found a significant difference in in-hospital mortality between patients with vasoplegic versus tissue dysoxic septic shock. These findings suggest a need to consider these differences when designing future studies of septic shock therapies. 16 Assessment of Clinical Deterioration and Progressive Organ Failure in Moderate Severity Emergency Department Sepsis Patients Lindsey J. Glaspey, Steven M. Hollenberg, Samantha A. Ni, Stephen Trzeciak, Ryan C. Arnold Cooper University Hospital, Camden, NJ Background: The PRE-SHOCK population, ED sepsis patients with tissue hypoperfusion (lactate of 2.0– 3.9 mM), commonly deteriorates after admission and requires transfer to critical care. Objectives: To determine the physiologic parameters and disease severity indices in the ED PRE-SHOCK sepsis population that predict clinical deterioration. We hypothesized that neither initial physiologic parameters nor organ function scores will be predictive. Methods: Design: Retrospective analysis of a prospectively maintained registry of sepsis patients with lactate measurements. Setting: An urban, academic medical center. Participants: The PRE-SHOCK population, defined as adult ED sepsis patients with either elevated lactate (2.0–3.9 mM) or transient hypotension (any sBP <90 mmHg) receiving IV antibiotics and admitted to a medical floor. Consecutive patients meeting PRESHOCK criteria were enrolled over a 1-year period. Patients with overt shock in the ED, pregnancy, or acute trauma were excluded. Outcome: Primary patientcentered outcome of increased organ failure (sequential organ failure assessment [SOFA] score increase >1 point, mechanical ventilation, or vasopressor utilization) within 72 hours of admission or in-hospital mortality. Results: We identified 248 PRE-SHOCK patients from 2649 screened. The primary outcome was met in 54% of the cohort and 44% were transferred to the ICU from a medical floor. Patients meeting the outcome of increased organ failure had a greater Shock Index (1.02 vs 0.93, p = 0.042) and heart rate (115 vs 105, p < 0.001) with no difference in initial lactate, age, MAP, or exposure to hypotension (sBP <100 mmHg). There was no difference in the Predisposition, Infection, Response, and Organ dysfunction (PIRO) score between groups (6.4 vs 5.7, p = 0.052). Outcome patients had similar initial levels of organ dysfunction but had higher SOFA scores at 24, 48, and 72 hours, a higher ICU transfer rate (60 vs 24%, p < 0.001), and increased ICU and hospital lengths of stay. Conclusion: The PRE-SHOCK sepsis population has a high incidence of clinical deterioration, progressive organ failure, and ICU transfer. Physiologic data in the ED were unable to differentiate the PRE-SHOCK sepsis patients who developed increased organ failure. This study supports the need for an objective organ failure assessment in the emergency department to supplement clinical decision-making. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Table - Abstract 16: Physiologic Parameters and Resource Utilization Mean arterial pressure (MAP) [mean (SD)] Any Systolic blood pressure < 100 mmHg [n(%)] Shock Index (HR/sBP) [mean (SD)] Lactate: Initial [mean (SD)] SOFA score: Initial [mean (SD)] SOFA score: 24 hours [mean (SD)] SOFA score: 48 hours [mean (SD)] ICU LOS [mean (SD)] Hospital LOS [mean (SD)] 17 Outcome (n = 135) No Outcome (n = 113) p value 76 (18) 73 (16) 0.171 63 (47) 63 (56) 0.199 1.02 (0.38) 0.93 (0.30) 0.042 2.45 (0.74) 2.29 (0.78) 0.099 2.1 (1.9) 1.8 (1.7) 0.195 2.9 (2.3) 1.3 (1.4) <0.001 2.5 (2.0) 1.0 (1.3) <0.001 3 (5) 11 (12) 1 (2) 6 (7) <0.001 <0.001 Lipopolysaccharide Detection in Patients with Septic Shock in the ED Daren M. Beam1, Michael A. Puskarich1, Mary Beth Fulkerson1, Jeffrey A. Kline1, Alan E. Jones2 1 Carolinas Medical Center, Charlotte, NC; 2 University of Mississippi Medical Center, Jackson, MS Background: Lipopolysaccharide (LPS) has long been recognized to initiate the host inflammatory response to infection with gram negative bacteria (GNB). Large clinical trials of potentially very expensive therapies continue to have the objective of reducing circulating LPS. Previous studies have found varying prevalence of LPS in blood of patients with severe sepsis. Compared with sepsis trials conducted 20 years ago, the frequency of GNB in culture specimens from emergency department (ED) patients enrolled in clinical trials of severe sepsis has decreased. Objectives: Test the hypothesis that prior to antibiotic administration, circulating LPS can be detected in the plasma of fewer than 10% of ED patients with severe sepsis. Methods: Secondary analysis of a prospective EDbased RCT of early quantitative resuscitation for severe sepsis. Blood specimens were drawn at the time severe sepsis was recognized, defined as two or more systemic inflammatory criteria and a serum lactate >4 mM or SPB<90 mmHg after fluid challenge. Blood was drawn in EDTA prior to antibiotic administration or within the first several hours, immediately centrifuged, and plasma frozen at )80C. Plasma LPS was quantified using the limulus amebocyte lysate assay (LAL) by a technician blinded to all clinical data. Results: 180 patients were enrolled with 140 plasma samples available for testing. Median age was 59 ± 17 years, 50% female, with overall mortality of 18%. Forty of 140 patients (29%) had any culture specimen positive for GNB including 21 (15%) with blood cultures positive. Only five specimens had detectable LPS, including two with a GNB-positive culture specimen, and three were LPS-positive without GNB in any culture. Prevalence of detectable LPS was 3.5% (CI: 1.5%–8.1%). • www.aemj.org S13 Conclusion: The frequency of detectable LPS in antibiotic-naive plasma is too low to serve as a useful diagnostic test or therapeutic target in ED patients with severe sepsis. The data raise the question of whether post-antibiotic plasma may have a higher frequency of detectable LPS. 18 Evaluation of the Efficacy of an Early Goal Directed Therapy (EGDT) Protocol When Using MEDS Score for Risk Stratification Ameer F. Ibrahim, Michael Mann, Jeff Scull, Michael Spino Jr., Lisa Voigt, Kimberly Zammit, Robert McCormack The State University of New York at Buffalo, Buffalo General Hospital, Buffalo, NY Background: EGDT is known to reduce mortality in septic patients. There is no evidence to date that delineates the role of using a risk stratification tool, such as the Mortality in Emergency Department Sepsis (MEDS) score, to determine which subgroups of patients may have a greater benefit with EGDT. Objectives: Our objective was to determine if our EGDT protocol differentially affects mortality based on the severity of illness using MEDS score. Methods: This study is a retrospective chart review of 243 patients, conducted at an urban tertiary care center, after implementing an EGDT protocol on July 1, 2008 (Figure). This study compares in-hospital mortality, length of stay (LOS) in ICU, and LOS in ED between the control group (126 patients from 1/1/07–12/31/07) and the postimplementation group (117 patients from 7/1/08–6/ 30/09), using MEDS score as a risk stratification tool. Inclusion criteria: patients who presented to our ED with a suspected infection, and two or more SIRS criteria, a MAP<65 mmHg, a SBP< 90 mmol/L. Exclusion criteria: age<18, death on arrival to ED, DNR or DNI, emergent surgical intervention, or those with an acute myocardial infarction or CHF exacerbation. A two-sample t-test was used to show that the mean age and number of comorbidities was similar between the control and study groups (p = 0.27 and 0.87 respectively). Mortality was compared and adjusted for MEDS score using logistic regression. The odds ratios and predicted probabilities of death are generated using the fitted logistic regression model. ED and ICU LOS were compared using Mood’s median test. Results: When controlling for illness severity using MEDS score, the relative risk (RR) of death with EGDT is about half that of the control group (RR = 0.52, 95% CI [0.278-0.973], p=0.04). Also, by applying MEDS score to risk stratify patients into various groups of illness severity, we found no specific groups where EGDT is more efficacious at reducing the Predicted Probability of death (Table 1). Without controlling for MEDS score, there is a trend in reduction of absolute mortality by 9.7% when EGDT is used (control = 30.2%, study = 20.5%, p = 0.086). EGDT leads to a 40.3% reduction in the median LOS in ICU (control = 124 hours, study = 74 hours, p = 0.03), without increasing LOS in ED (control = 6 hours, study = 7 hours, p = 0.50). Conclusion: EGDT is beneficial in patients with severe sepsis or septic shock, regardless of their MEDS score. S14 2012 SAEM ANNUAL MEETING ABSTRACTS Table - Abstract 18: RR of Death within Each MEDS Score Illness Severity Subgroup MEDS Score Very High (>15) High (13–15) Moderate (8–12) Low (5–7) Very Low (0–4) 19 Control Post implementation Difference P value 0.621 0.496 0.358 0.181 0.127 0.457 0.336 0.222 0.102 0.069 0.164 0.160 0.136 0.079 0.058 0.801 0.856 0.437 0.484 0.625 Validation of Using Fingerstick Blood Sample with i-Stat POC Device for Cardiac Troponin I Assay Devin Loewenstein, Christine Stake, Mark Cichon Loyola University Health System, Maywood, IL Background: In patients experiencing acute coronary syndrome (ACS), prompt diagnosis is critical in achieving the best health outcome. While ECG analysis is usually sufficient to diagnose ACS in cases of ST elevation, ACS without ST elevation is reliably diagnosed through serial testing of cardiac troponin I (cTnI). Pointof-care testing (POCT) for cTnI by venipuncture has been proven a more rapid means to diagnosis than central laboratory testing. Implementing fingerstick testing for cTnI in place of standard venipuncture methods would allow for faster and easier procurement of patients’ cTnI levels, as well as increase the likelihood of starting a rapid test for cTnI in the prehospital setting, which could allow for even earlier diagnosis of ACS. Objectives: To determine if fingerstick blood samples yield accurate and reliable troponin measurements compared to conventional venous blood draws using the i-STAT POC device. Methods: This experimental study was performed in the ED of a quaternary care suburban medical center between June-August 2011. Fingerstick blood samples were obtained from adult ED patients for whom standard (venipuncture) POC troponin testing was ordered. The time between fingerstick and standard draws was ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Table - Abstract 19: Categorical comparison of standard ED POCT and fingerstick results kept as narrow as possible. cTnI assays were performed at the bedside using the i-STAT 1 (Abbott Point of Care). Results: 94 samples from 87 patients were analyzed by both fingerstick and standard ED POCT methods (see Table). Four resulted in cartridge error. Compared to ‘‘gold standard’’ ED POCT, fingerstick testing has a positive predictive value of 100%, negative predictive value of 96%, sensitivity of 79%, and specificity of 100%. No significant difference in cTnI level was found between the two methods, with a nonparametric intraclass correlation coefficient of 0.994 (95% CI 0.992–0.996, p-value < 0.001). Conclusion: Whole blood fingerstick cTnI testing using the i-STAT device is suitable for rapid evaluation of cTnI level in prehospital and ED settings. However, results must be interpreted with caution if they are within a narrow territory of the cutoff for normal vs. elevated levels. Additional testing on a larger sample would be beneficial. The practicality and clinical benefit of using fingerstick cTnI testing in the EMS setting must still be assessed. 20 Central Versus Local Adjudication of Myocardial Infarction in a Biomarker Trial Stephen W. Smith1, John T. Nagurney2, Deborah B. Diercks3, Rick C. San George4, Fred S. Apple1, Peter A. McCullough5, Christian T. Ruff6, Arturo Sesma7, W. Frank Peacock8 1 Hennepin County Medical Center, Minneapolis, MN; 2Massachusetts General Hospital, Boston, MA; 3University of California, Davis, CA; 4Alere, Biosite, San Diego, CA; 5Providence Park Heart Institute, Novi, MI; 6Brigham and Women’s Hospital, Boston, MA; 7St. Catherine University, St. Paul, MN; 8Cleveland Clinic Foundation, Cleveland, OH Background: Adjudication of diagnosis of acute myocardial infarction (AMI) in clinical studies typically occurs at each site of subject enrollment (local) or by experts at an independent site (central). From 2000– 2007, the troponin (cTn) element of the diagnosis was predicated on the local laboratories, using a mix of the 99th percentile reference cTn and ROC-determined cutpoints. In 2007, the universal definition of AMI (UDAMI) defined it by the 99th percentile reference alone. Objectives: To compare the diagnosis rates of AMI as determined by local adjudication vs. central adjudication using UDAMI criteria. Methods: Retrospective analysis of data from the Myeloperoxidase in the Diagnosis of Acute Coronary • www.aemj.org S15 Syndromes (ACS) Study (MIDAS), an 18-center prospective study with enrollment from 12/19/06 to 9/20/07 of patients with suspected ACS presenting to the ED < 8 hours after symptom onset and in whom serial cTn and objective cardiac perfusion testing was planned. Adjudication of ACS was done by single local principal investigators using clinical data and local cTn cutpoints from 13 different cTn assays, and applying the 2000 definition. Central adjudication was done after completion of the MIDAS primary analysis using the same data and local cTn assay, but by experts at three different institutions, using the UDAMI and the manufacturer’s 99th percentile cTn cutpoint, and not blinded to local adjudications. Discrepant dignoses were resolved by consensus. Local vs. central cTn cutpoints differed for six assays, with central cutpoints lower in all. Statistics were by chi-square and kappa. Results: Excluding 11 cases deemed indeterminate by central adjudication, 1096 cases were successfully adjudicated. Local adjudication resulted in 104 AMI (9.5% of total) and 992 non-AMI; central adjudication resulted in 134 (12.2%) AMI and 962 non-AMI. Overall, 44 local diagnoses (4%) were either changed from non-AMI to AMI or AMI to non-AMI (p < 0.001). Interrater reliability across both methods was found to be kappa = 0.79 (p < 0.001). For ACS diagnosis, local adjudication identified 252 ACS cases (23%) and 854 non-ACS, while central adjudication identified 275 ACS (25%) and 831 non-ACS. Overall, 61 local diagnoses (6%) were either changed from non-ACS to ACS or ACS to non-ACS (p < 0 .001). Interrater reliability found kappa = 0.85 (p < 0.001). Conclusion: Central and local adjudication resulted in significantly different rates of AMI and ACS diagnosis. However, overall agreement of the two methods across these two diagnoses was acceptable. Table - Abstract 20: Differences in AMI Diagnosis Between Central and Local Adjudication Local Central AMI Non-AMI AMI Non-AMI –ACS, not AMI –Non-cardiac chest pain –Chest pain not otherwise specified Total changes 21 AMI Non-AMI Non-AMI AMI AMI AMI AMI Number (%) 97 (8.9%) 955 (87.1%) 7 (0.64%) Total 37 (3.4%) )17 (1.6%) )9 (0.8%) )11 (1.0%) 44 (4.0%, 31% of all MI Dx’s) Myeloperoxidase And C-Reactive Protein In Patients With Cocaine-associated Chest Pain Katie O’Conor1, Anna Marie Chang1, Alan Wu2, Judd E. Hollander1 1 Hospital of the University of Pennsylvania, Philadelphia, PA; 2University of California, San Francisco, San Francisco, CA Background: In ED patients presenting with chest pain, acute coronary syndrome (ACS) was found to S16 2012 SAEM ANNUAL MEETING ABSTRACTS occur four times more often in cocaine users. Biomarkers myeloperoxidase (MPO) and C-reactive protein (CRP) have potential in the diagnosis of ACS. Objectives: To evaluate the utility of MPO and CRP in the diagnosis of ACS in patients presenting to the ED with cocaine-associated chest pain and compare the predictive value to nonusers. We hypothesized that these markers may be more sensitive for ACS in nonusers given the underlying pathophysiology of enhanced plaque inflammation. Methods: A secondary analysis of a cohort study of enrolled ED patients who received evaluation for ACS at an urban, tertiary care hospital. Structured data collection at presentation included demographics, chest pain history, lab, and ECG data. Subjects included those with self-reported or lab-confirmed cocaine use and chest pain. They were matched to controls based on age, sex, and race. Our main outcome was diagnosis of ACS at index visit. We determined median MPO and CRP values, calculated maximal AUC for ROC curves, and found cut-points to maximize sensitivity and specificity. Data are presented with 95% CI. Results: Overall, 95 patients in the cocaine positivegroup and 86 patients in the nonusers group had MPO and CRP levels measured. Patients had a median age of 47 (IQR, 40–52), 90% black or African American, and 62% male (p > 0.05 between groups). Fifteen patients were diagnosed with ACS: 8 patients in the cocaine group and 7 in the nonusers group. Comparing cocaine users to nonusers, there was no difference in MPO (median 162 [IQR, 101–253] v 136 [111–235] ng/mL; p = 0.78) or CRP (3 [1–9] v 5 [1–15] mg/L; p = 0.08). The AUC for MPO was 0.65 (95% CI 0.39–0.90) v 0.54 (95% CI 0.19–0.73). The optimal cut-point to maximize sensitivity and specificity was 242 ng/mL which gave a sensitivity of 0.42 and specificity of 0.75. Using this cutpoint, 57% v 29% of ACS in cocaine users vs the nonusers would be identified. The AUC for CRP was 0.63 (95% CI 0.39–0.88) in cocaine users vs 0.73 (95% CI 0.52–0.95) in nonusers. The optimal cut point was 11.9 mg/L with a sensitivity of 0.67 and specificity of 0.79. Using this cutpoint, 43% v 88% of ACS in cocaine users and nonusers would have been identified. Conclusion: The diagnostic accuracy of MPO and CRP is not different in cocaine users than nonusers and does not appear to have sufficient discriminatory ability in either cohort. 22 A Soluble Guanylate Cyclase Stimulator, Bay 41-8543, Preserves Right Ventricular Function In Experimental Pulmonary Embolism John A. Watts, Michael A. Gellar, Mary-Beth K. Fulkerson, Jeffrey A. Kline Carolinas Medical Center, Charlotte, NC Background: Pulmonary embolism (PE) increases pulmonary vascular hypertension and can cause right ventricular (RV) injury by shear forces, stretch, work, and inflammatory responses. Our recent studies show that the vasodilation observed following in vivo treatment with BAY 41-8543 reduces pulmonary vascular resistance in a 5 hr model of acute experimental PE. Objectives: To test if treatment of rats with BAY 418543 protects against RV dysfunction in two models of acute experimental PE. Methods: Experimental PE was induced in anesthetized, male Sprague-Dawley rats by infusing 25 um polystyrene microspheres in the right jugular vein to produce RV injury from severe PE (2.6 million/100 gm body wt, 5 hrs) or moderate PE (2.0 million/100 gm body wt, 18 hrs). Rats with PE were treated with BAY 41-8543 (50 ug/kg, IV) or its solvent at the time of PE induction. Control rats received vehicle for microspheres. Hearts were perfused by Langendorff technique to assess RV function. Values are mean ± se, n = 7 / group. Comparisons between control values, PE, and PE + BAY 41-8543 were made by ANOVA followed by Neuman-Keuls test (significance = p < 0.05). Results: 18 hrs of moderate PE caused a significant decrease in RV heart function in rats treated with the solvent for BAY 41-8543: peak systolic pressure (PSP) decreased from 39 ± 1.5 mmHg, control to 16 ± 1.5, PE, +dP/dt decreased from 1192 ± 93 mmHg/sec to 463 ± 77, -dP/dt decreased from )576 ± 60 mmHg/sec to )251 ± 9. Treatment of rats with BAY 41-8543 significantly improved all three indices of RV heart function (PSP 29 ± 2.6, +dP/dt 1109 ± 116, -dP/dt )426 ± 69). 5 hrs of severe PE also caused significant RV dysfunction (PSP 25 ± 2, -dP/dt )356 ± 28) and treatment with BAY 41-8543 produced protection of RV heart function (PSP 34 ± 2, -dP/dt )535 ± 41) similar to the 18 hr moderate PE model. Conclusion: Experimental PE produced significant RV dysfunction, which was ameliorated by treatment of the animals with the soluble guanylate cyclase stimulator, BAY 41-8543. 23 Prospective Evaluation of a Simplified Risk Stratification Tool for Chest Pain Patients in an Emergency Department Observation Unit Matthew J. Fuller, Jessica Holly, Thomas Rayner, Camille Broadwater-Hollifield, Michael Mallin, Virgil Davis, Erik Barton, Troy Madsen University of Utah, Salt Lake City, UT Background: The Thrombolysis in Myocardial Infarction (TIMI) score has been validated as a risk stratification tool in the emergency department (ED) setting, but certain aspects of the scoring system may not be applicable when applied to chest pain patients selected for ED observation unit (EDOU) stay. Objectives: We evaluated a simplified, three-point risk stratification tool for EDOU patients which we termed the CARdiac score: Coronary disease [previous myocardial infarction (MI), stent, or coronary artery bypass graft (CABG)], Age (65 years or older), and Risk factors (at least 3 of 5 cardiac risk factors). Methods: We performed a prospective, observational study with 30-day phone follow-up for all chest pain patients admitted to our EDOU over a one-year period. Baseline data, outcomes related to EDOU stay, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 inpatient admission, and 30-day outcomes were recorded. CARdiac scores were calculated based on patient history and were used to evaluate the risk of the composite outcome of MI, stent/CABG, or death during the EDOU stay, inpatient admission, or 30-day follow-up period. The CARdiac score was not used during the EDOU stay and was calculated with blinding to patient outcomes. Results: 552 patients were evaluated. Average age was 54.1 years (19–80 yrs) and 46% were male. Eighteen patients experienced composite outcomes: stent (12), CABG (3), MI and stent (2), and MI and CABG (1). Risk of the composite outcome generally increased by CARdiac score: 0 (1.6%), 1 (3.6%), 2 (9%), and 3 (8.3%). Patients with a CARdiac score of 2 or 3 (moderate risk) were significantly more likely to experience MI, stent, or CABG than those with a score of 0 or 1 (low risk): 7/ 79 moderate-risk patients (8.9%) had the composite outcome vs. 11/473 low-risk patients (2.3%, p = 0.008, relative risk = 3.9). Those moderate-risk by the CARdiac score were also more likely to require inpatient admission from the EDOU (16.5% vs. 9.1%, p = 0.045). Conclusion: The CARdiac score may prove to be a simple tool for risk stratification of chest pain patients in an EDOU. Patients considered moderate-risk by CARdiac score may be appropriate for more intensive evaluation in the EDOU or consideration for inpatient admission rather than EDOU placement. 24 Disease Progression in Patients Without Clinically Significant Stenosis on Coronary CT Angiography Performed for Evaluation of Potential Acute Coronary Syndrome Anna Marie Chang1, Catherine T. Ginty2, Harold I. Litt1, Judd E. Hollander1 1 Hospital of the University of Pennsylvania, Philadelphia, PA; 2Cooper University Hospital, Camden, NJ Background: Patients who present to the ED with symptoms of potential acute coronary syndrome (ACS) can be safely discharged home after a negative coronary computerized tomographic angiography (CTA). However, the duration of time for which a negative coronary CTA can be used to inform decision making when patients have recurrent symptoms is unknown. Objectives: We examined patients who received more than one coronary CTA for evaluation of ACS to determine whether they had disease progression, as defined by crossing the threshold from noncritical (<50% maximal stenosis) to potentially critical disease. Methods: We performed a structured comprehensive record search of all coronary CTAs performed from 2005 to 2010 at a tertiary care health system. Low-tointermediate risk ED patients who received two or more coronary CTAs, at least one from an ED evaluation for potential ACS, were identified. Patients who were revascularized between scans were excluded. We collected demographic data, clinical course, time between scans, and number of ED visits between scans. Record review was structured and done by trained • www.aemj.org S17 abstractors. Our main outcome was progression of coronary stenosis between scans, specifically crossing the threshold from noncritical to potentially critical disease. Results: Overall, 32 patients met study criteria (median age 45, interquartile range [IQR] (37.5–48); 56% female; 88% black). The median time between studies was 27.3 months (IQR, 18.2–33.2). 22 patients did not have stenosis in any vessel on either coronary CTA, two studies showed increasing stenosis of <20%, and the rest showed ‘‘improvement,’’ most due to better imaging quality. No patient initially below the 50% threshold subsequently exceeded it (0%; 95% CI, 0–11.0%). Patients also had varying numbers of ED visits (median number of visits 5, range 0–23), and numbers of ED visits for potentially cardiac complaints (median 1, range 0–6); 10 were re-admitted for potentially cardiac complaints (for example, chest pain or shortness of breath), and 9 received further provocative cardiac testing, all of which had negative results. Conclusion: We did not find clinically significant disease progression within a 2 year time frame in patients who had a negative coronary CTA, despite a high number of repeat visits. This suggests that prior negative coronary CTA may be able to be used to inform decision making within this time period. 25 Triple Rule Out CT Scan for Patients Presenting to the Emergency Department with Chest Pain: A Systematic Review and Meta-Analysis David Ayaram1, Fernanda Bellolio1, Torrey Laack1, Hassan M. Murad1, Victor M. Montori1, Judd Hollander2, Ian G. Stiell3, Erik P. Hess1 1 Mayo Clinic, Rochester, MN; 2University of Pennsylvania, Philadelphia, PA; 3University of Ottawa, Ottawa, ON, Canada Background: ‘‘Triple rule out’’ (TRO) CT has emerged as a new technology to evaluate for coronary artery disease (CAD), pulmonary embolism (PE), and aortic dissection (AD) in a single imaging modality. Objectives: We compared the diagnostic accuracy, image quality, radiation exposure, contrast volume, length of stay, and admission rate of TRO CT to other diagnostic modalities (dedicated coronary, PE, and AD CT; coronary angiography; and nuclear stress testing) for the evaluation of chest pain. Methods: With the assistance of an expert librarian we searched four electronic databases and reference lists of included studies, and contacted content experts to identify articles for review. Included articles met all the following criteria: enrolled patients with non-traumatic chest pain, shortness of breath, suspected ACS, PE, or AD; used 64-slice CT technology; and compared TRO CT to another diagnostic modality. Statistical comparisons were conducted using RevMan, and Meta-DiSc was used to pool diagnostic accuracy estimates. Results: Nine ED studies enrolling 1371 patients (483 TRO, 888 non-TRO) were included (1 RCT and 8 observational). Patients undergoing TRO CT were exposed to more radiation (mean difference [MD] 5.03 mSv, 95% CI 4.16–5.91) and more contrast (MD 45.6 mL, 95% CI S18 2012 SAEM ANNUAL MEETING ABSTRACTS 42.7–48.6) compared to non TRO CT patients. There was no significant difference in image quality between TRO CT images and those of dedicated CT scans in any studies performing this comparison. Similarly, there was no significant difference between TRO CT and other diagnostic modalities in regards to length of stay or admission rate. When compared to conventional coronary angiography as the gold standard for evaluation of CAD, TRO CT had the following pooled diagnostic accuracy estimates: sensitivity 0.94 [95%CI 0.89–0.98], specificity 0.98 [0.97–0.99], LR+ 35.7 [95%CI 11.1–115.0], and LR- 0.07 [95%CI 0.02–0.35]. Conclusion: TRO chest CT is comparable to dedicated PE, coronary, or AD CT in regard to image quality, length of stay, and admission rate and is highly accurate for detecting CAD. The utility of TRO CT depends on the relative pre-test probabilities of the conditions being assessed and its role is yet to be clearly defined. TRO CT, however, involves increased radiation exposure and contrast volume and for this reason clinicians should be selective in its use. 26 Impact of Coronary Computer Tomographic Angiography Findings on the Medical Treatment of CAD Anna Marie Chang, Stephen Kimmel, Jeffrey Le, Judd E. Hollander Hospital of the University of Pennsylvania, Philadelphia, PA Background: Coronary computed tomographic angiography (CCTA) has high sensitivity, specificity, accuracy, and prognostic value for coronary artery disease (CAD) and ACS. However, how a CCTA informs subsequent use of prescription medication is unclear. Objectives: To determine if detection of critical or noncritical CAD on CCTA is associated with initiation of aspirin and statins for patients who presented to the ED with chest pain. We hypothesized that aspirin and statins would be more likely to be prescribed to patients with noncritical disease relative to those without any CAD. Methods: Prospective cohort study of patients who received CCTA as part of evaluation of chest pain in the ED or observation unit. Patients were contacted and medical records were reviewed to obtain clinical follow-up for up to the year after CCTA. The main outcome was new prescription of aspirin or statin. CAD severity on CCTA was graded as absent, mild (1% to 49%), moderate (50% to 69%), or severe (‡70%) stenosis. Logistic regression was used to assess the association of stenosis severity to new medication prescription; covariates were determined a priori. Results: 859 patients who had CCTA performed consented to participate in this study or met waiver of consent for record review only (median age, 48.5 IQR 42.7–53.4, 59% female, 71% black). Median follow-up time was 333 days, IQR 70–725 days. At baseline, 13% of the total cohort was already prescribed aspirin and 8% on statin medication. Two hundred seventy nine (32%) patients were found to have stenosis in at least one vessel. In patients with absent, mild, moderate, and severe CAD on CCTA, aspirin was initiated in 11%, 34%, 52%, and 55%; statins were initiated in 7%, 22%, 32%, and 53% of patients. After adjustment for age, race, sex, hypertension, diabetes, cholesterol, tobacco use, and admission to the hospital after CCTA, higher grades of CAD severity were independently associated with greater post-CCTA use of aspirin (OR 1.9 per grade, 95% CI 1.4–2.2, p < 0.001) and statins (OR 1.9, 95% CI 1.5–2.4, p < 0.001). Conclusion: Greater CAD severity on CCTA is associated with increased medication prescription for CAD. Patients with noncritical disease are more likely than patients without any disease to receive aspirin and statins. Future studies should examine whether these changes lead to decreased hospitalizations and improved cardiovascular health. Table - Abstract 26: New Prescription of Aspirin or Statin by Degree of CAD Degree of CAD None Mild (1–49%) Moderate (50–69%) Severe (>70%) Aspirin Statin 68/580 38/580 41/160 35/160 40/76 25/76 24/43 23/43 27 Validation of a Clinical Decision Rule for ED Patients with Potential Acute Coronary Syndromes (ACS) Anna Marie Chang1, Julie Pitts1, Frances Shofer1, Jeffrey Le1, Emily Barrows1, Shannon Marcoon1, Erik Hess2, Judd Hollander1 1 University of Pennsylvania, Philadelphia, PA; 2 Mayo Clinic, Rochester, MN Background: Hess et al. developed a clinical decision rule for patients with acute chest pain consisting of the absence of five predictors: ischemic ECG changes not known to be old, elevated initial or 6-hour troponin level, known coronary disease, ‘‘typical’’ pain, and age over 50. Patients less than 40 required only a single troponin evaluation. Objectives: To test the hypothesis that patients less than 40 years old without these criteria are at <1% risk for major adverse cardiovascular events (MACE) including death, AMI, PCI, and CABG. Methods: We performed a secondary analysis of several combined prospective cohort studies that enrolled ED patients who received an evaluation for ACS in an urban ED from 1999 to 2009. Cocaine users and STEMI patients were excluded. Structured data collection at presentation included demographics, pain description, history, lab, and ECG data for all studies. Hospital course was followed daily. Thirty-day follow up was done by telephone. Our main outcome was 30-day MACE using objective criteria. The secondary outcome was potential change in ED disposition due to application of the rule. Descriptive statistics and 95% CIs were used. Results: Of 9289 visits for potential ACS, patients had a mean age of 52.4 ± 14.7 yrs; 68% were black and 59% female. There were 638 patients (6.9%) with 30-day CV events (93 dead, 384 AMI, 298 PCI). Sequential removal of patients in order to meet the final rule for patients less than 40 excluded patients based upon: ischemic ECG changes not old (n = 434, 30% MACE rate), ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 elevated initial troponin level (n = 237, 60% MACE), known coronary disease (n = 1622, 11% MACE), ‘‘typical’’ pain (n = 3179, 3% MACE), and age over 40 (n = 2690, 3.4% MACE) leaving 1127 patients less than 40 with 0.8% MACE [95% CI, 0.4–1.5%]. Of this cohort, 70% were discharged home from the ED by the treating physician without application of this rule. Adding a second negative troponin in patients 40–50 years old identified a group of 1139 patients with a 2.0% rate of MACE [1.3–3.0] and a 48% discharge rate. Conclusion: The Hess rule appears to identify a cohort of patients at approximately 1% risk of 30-day MACE, and may enhance discharge of young patients. However, even without application of this rule, the 70% of young patients at low risk are already being discharged home based upon clinical judgment. 28 Utilization of an Electronic Clinical Decision Rule Does Not Change Emergency Physicians’ Pattern of Practice in Evaluating Patients with Possible Pulmonary Embolism Salam Lehrfeld1, Corey Weiner2, Brian Gillett2, Marsia Vermeulen2, Antonios Likourezos2, Eitan Dickman2 1 UTSouthwestern Medical Center, Dallas, TX; 2 Maimonides Medical Center, Brooklyn, NY Background: A Clinical Decision Support System (CDSS) incorporates evidence-based medicine into clinical practice, but this technology is underutilized in the ED. A CDSS can be integrated directly into an electronic medical record (EMR) to improve physician efficiency and ease of use. The Christopher study investigators validated a clinical decision rule for • www.aemj.org S19 patients with suspected pulmonary embolism (PE). The rule stratifies patients using Wells’ criteria to undergo either D-dimer testing or a CT angiogram (CT). The effect of this decision rule, integrated as a CDSS into the EMR, on ordering CTs has not been studied. Objectives: To assess the effect of a mandatory CDSS on the ordering of D-dimers and CTs for patients with suspected PE. Methods: We assessed the number of CTs ordered for patients with suspected PE before and after integrating a mandatory CDSS in an urban community ED. Physicians were educated regarding CDSS use prior to implementation. The CDSS advised physicians as to whether a negative D-dimer alone excluded PE or if a CT was required based on Wells’ criteria. The EMR required physicians to complete the CDSS prior to ordering the CT. However, physicians maintained the ability to order a CT regardless of the CDSS recommendation. Patients ‡18 years of age presenting to the ED with a chief complaint of chest pain, dyspnea, syncope, or palpitations were included in the data analysis. We compared the proportion of D-dimers and CTs ordered during the 8-month periods immediately before and after implementing the CDSS. All 27 physicians who worked in the ED during both time periods were included in the analysis. Patients with an allergy to intravenous contrast agents, renal insufficiency, or pregnancy were excluded. Results were analyzed using a chi-square test. Results: A total of 11,931 patients were included in the data analysis (6054 pre- and 5877 post-implementation). CTs were ordered for 215 patients (3.6%) in the pre-implementation group and 226 patients (3.8%) in the post-implementation group; p = 0.396. A D-dimer was ordered for 392 patients (6.5%) in the pre-implementation group and 382 patients (6.5%) in the post-implementation group; p = 0.958. S20 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: In this single-center study, EMR integration of a mandatory CDSS for evaluation of PE did not significantly alter ordering patterns of CTs and D-Dimers. 29 Identification of Patients with Low-Risk Pulmonary Emboli Suitable for Discharge from the Emergency Department Mike Zimmer, Keith E. Kocher University of Michigan, Ann Arbor, MI Background: Recent data, including a large, multicenter randomized controlled trial, suggest that a low-risk cohort of patients diagnosed with pulmonary embolism (PE) exists who can be safely discharged from the ED for outpatient treatment. Objectives: To determine if there is a similar cohort at our institution who have a low rate of complications from PE suitable for outpatient treatment. Methods: This was a retrospective chart review at a single academic tertiary referral center with an annual ED volume of 80,000 patients. All adult ED patients who were diagnosed with PE during a 24-month period from 11/1/09 through 10/31/11 were identified. The Pulmonary Embolism Severity Index (PESI) score, a previously validated clinical decision rule to risk stratify patients with PE, was calculated. Patients with high PESI (>85) were excluded. Additional exclusion criteria included patients who were at high risk of complications from initiation of therapeutic anticoagulation and those patients with other clear indications for admission to the hospital. The remaining cohort of patients with low risk PE (PESI £ 85) was included in the final analysis. Outcomes were measured at 14 and 90 days after PE diagnosis and included death, major bleeding, and objectively confirmed recurrent venous thromboembolism (VTE). Results: During the study period, 298 total patients were diagnosed with PE. There were 172 (58%) patients categorized as ‘‘low risk’’ (PESI £ 85), with 42 removed because of various pre-defined exclusion criteria. Of the remaining 130 (44%) patients suitable for outpatient treatment, 5 patients (3.8%; 95% CI, 0.5% - 7.2%) had one or more negative outcomes by 90 days. This included 2 (1.5%; 95% CI, 0% - 3.7%) major bleeding events, 2 (1.5%; 95% CI, 0% - 3.7%) recurrent VTE, and 2 (1.5%; 95% CI, 0% - 3.7%) deaths. None of the deaths were attributable to PE or anticoagulation. One patient suffered both a recurrent VTE and died within 90 days. Both patients who died within 90 days were transitioned to hospice care because of worsening metastatic burden. At 14 days, there was 1 bleeding event (0.8%; 95% CI, 0% - 2.3%), no recurrent VTE, and no deaths. The average hospital length of stay for these patients was 2.8 days (SD ±1.6). Conclusion: Over 40% of our patients diagnosed with PE in the ED may have been suitable for outpatient treatment, with 4% suffering a negative outcome within 90 days and 0.8% suffering a negative outcome within 14 days. In addition, the average hospital length of stay for these patients was 2.8 days, which may represent a potential cost savings if these patients had been managed as outpatients. Our experience supports previous studies that suggest the safety of outpatient treatment of patients diagnosed with PE in the ED. Given the potential savings related to a decreased need for hospitalization, these results have health policy implications and support the feasibility of creating protocols to facilitate this clinical practice change. 30 Validation of a Clinical Prediction Rule for Chest Radiography in Emergency Department Patients with Chest Pain and Possible Acute Coronary Syndrome Joshua Guttman1, Eli Segal2, Mark Levental2, Xiaoqing Xue3, Marc Afilalo2 1 McGill University, Montreal, QC, Canada; 2 Jewish General Hospital, McGill University, Montreal, QC, Canada; 3Jewish General Hospital, Montreal, QC, Canada Background: Chest x-rays (CXRs) are commonly obtained on ED chest pain patients presenting with suspected acute coronary syndrome (ACS). A recently derived clinical decision rule (CDR) determined that patients who have no history of congestive heart failure, have never smoked, and have a normal lung examination do not require a CXR in the ED. Objectives: To validate the diagnostic accuracy of the Hess CXR CDR for ED chest pain patients with suspected ACS. Methods: This was a prospective observational study of a convenience sample of chest pain patients over 24 years old with suspected ACS who presented to a single urban academic ED. The primary outcome was the ability of the CDR to identify patients with abnormalities on CXR requiring acute ED intervention. Data were collected by research associates using the chart and physician interviews. Abnormalities on CXR and specific interventions were predetermined, with a positive CXR defined as one with abnormality requiring ED intervention, and a negative CXR defined as either normal or abnormal but not requiring ED intervention. The final radiologist report was used as a reference standard for CXR interpretation. A second radiologist, blinded to the initial radiologist’s report, reviewed the CXRs of patients meeting the CDR criteria to calculate inter-observer agreement. Patients were followed up by chart review and telephone interview 30 days after presentation. Results: Between January and August 2011, 178 patients were enrolled, of whom 38 (21%) were excluded and 10 (5.6%) did not receive CXRs in the ED. Of the 130 remaining patients, 74 (57%) met the CDR. The CDR identified all patients with a positive CXR (sensitivity = 100%, 95%CI 40–100%). The CDR identified 73 of the 126 patients with a negative CXR (specificity = 58%, 95%CI 49–67%). The positive likelihood ratio was 2.4 (95%CI 1.9–2.9). Inter-observer agreement between radiologists was substantial (kappa = 0.63, 95%CI 0.41–0.85). Telephone contact was made with 78% of patients and all patient charts were reviewed at 30 days. None had any adverse events related to a ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 diagnosis made on CXR. If the CDR had been applied, CXR usage would have dropped from 93% to 41%, an absolute reduction of 52%. Conclusion: The Hess CXR CDR was validated in our setting. Implementation of this CDR has the potential to reduce CXR usage in this population. 31 The Detection Rate of Pulmonary Embolisms by Emergency Physicians Has Increased Scott M. Alter, Barnet Eskin, John R. Allegra Morristown Medical Center, Morristown, NJ Background: Currently most pulmonary embolisms (PEs) are diagnosed using chest computed tomography (CT) imaging. A recent study showed that there has been a six-fold increase in the rate of chest CT imaging in the emergency department (ED) from 2001 to 2007. Objectives: We hypothesize that this has led to an increase in the rate of detecting PEs in the ED. Our goal was to analyze the recent trends in the annual rates of detection of PEs in the ED. Methods: Design: Retrospective cohort. Setting: 33 suburban, urban, and rural New York and New Jersey EDs with annual visits between 8,000 and 75,000. Participants: Consecutive patients seen by ED physicians from January 1, 1996 through December 31, 2010. Observations: We identified PEs using ICD-9 codes and calculated annual rates by dividing the number of PEs by the total ED visits for each year. We determined statistical significance using Student’s t-test, calculated 95% confidence intervals (CI), and performed a regression analysis. Alpha was set at 0.05. Results: Of 9,533,827 ED visits, 5,595 (1 in 1,704 visits, 0.059%) were for PEs. The mean patient age was 58 ± 19 years and 57% were female. From 1996 to 2010, the rate increased 3.8-fold (95% CI, 3.2–4.6; p < 0.0001) from 1 in 4,603 visits (0.022%) in 1996 to 1 in 1,213 (0.082%) visits in 2010. The correlation coefficient for the rate increase was R2 = 0.93 (p < 0.0001). Conclusion: The rate of detection of pulmonary embolisms by ED physicians has increased 3.8-fold over the 15 years of our study. Although it may be due to increased incidence, we believe this trend is most likely due to the increased utilization of chest CT imaging. 32 • www.aemj.org S21 D-dimer Threshold Increase With Pretest Probability Unlikely For Pulmonary Embolism To Decrease Unnecessary Computerized Tomographic Pulmonary Angiography Jeffrey Kline1, Melanie M. Hogg1, Daniel M. Courtney2, Chadwick D. Miller3, Alan E. Jones4, Howard A. Smithline5 1 Carolinas Medical Center, Charlotte, NC; 2 Northwestern University, Chicago, IL; 3Wake Forest School of Medicine, Winston Salem, NC; 4 University of Mississippi Medical Center, Jackson, MS; 5Baystate Medical Center, Springfield, MA Background: Increasing the threshold to define a positive D-dimer in low-risk patients could reduce unnecessary computed tomographic pulmonary angiography (CTPA) for suspected PE. This strategy might increase rates of missed PE and missed pneumonia, the most common non-thromboembolic finding on CTPA that might not otherwise be diagnosed. Objectives: Measure the effect of doubling the standard D-dimer threshold for ‘‘PE unlikely’’ Revised Geneva (RGS) or Wells’ scores on the exclusion rate, frequency, and size of missed PE and missed pneumonia. Methods: Prospective enrollment at four academic US hospitals. Inclusion criteria required patients to have at least one symptom or sign and one risk factor for PE, and have 64-channel CTPA completed. Pretest probability data were collected in real time and the D-dimer was measured in a central laboratory. Criterion standard for PE or pneumonia consisted of CPTA interpretation by two independent radiologists combined with necessary treatment plan. Subsegmental PE was defined as total vascular obstruction <5%. Patients were followed for outcome at 30 days. Proportions were compared with 95% CIs. Results: Of 678 patients enrolled, 126 (19%) were PE+ and 93 (14%) had pneumonia. With RGS£6 and standard threshold (<500 ng/mL), D-dimer was negative in 110/678 (16%, 95% CI: 13–19%), and 4/110 were PE+ (posterior probability 3.8%, 95% CI: 1–9.3%). With RGS£6 and a threshold <1000 ng/mL, D-dimer was negative in 208/678 (31%, 27–44%) and 11/208 (5.3%, 2.8– 9.3%) were PE+, but 10/11 missed PEs were subsegmental, and none had concomitant DVT. The posterior probability for pneumonia among patients with RGS≤6 and D-dimer<500 was 9/110 (8.2%, 4–15%) which compares favorably to the posterior probability of 12/208 (5.4%, 3–10%) observed with RGS& #8804;6 and D-dimer<1000 ng/mL. Of the 200 (35%) patients who also had plain film CXR, radiologists found an infiltrate in only 58. Use of Wells£4 produced similar results as the RGS≤6 for exclusion rate and posterior probability of both PE and pneumonia. Conclusion: Doubling the threshold for a positive D-dimer with a PE unlikely pretest probability can significantly reduce CTPA scanning with a slightly increased risk of missed isolated subsegmental PE, and no increase in rate of missed pneumonia. S22 33 2012 SAEM ANNUAL MEETING ABSTRACTS A Randomized Trial of N-Acetyl Cysteine and Saline versus Normal Saline Alone to Prevent Contrast Nephropathy in Emergency Department Patients Undergoing Contrast Enhanced Computed Tomography Stephen Traub1, Alice Mitchell2, Alan E. Jones3, Aimee Tang1, Jennifer O’Connor1, John Kellum4, Nathan Shapiro1 1 Beth Israel Deaconess Medical Center, Boston, MA; 2Carolinas Medical Center, Charlotte, NC; 3 University of Mississippi, Jackson, MS; 4 University of Pittsburgh Medical Center, Pittsburgh, PA Background: Prior studies have suggested that both N-acetylcysteine (NAC) and intravenous (IV) fluids confer renal protection in patients exposed to radiocontrast media. However, few studies have focused on CT scans in ED patients. Objectives: To test the hypothesis that NAC plus normal saline (NS) is more effective than saline alone in the prevention of radiocontrast nephropathy (RCN). Methods: Double blind, randomized, controlled trial conducted in two tertiary care, urban EDs. Inclusion Criteria: Patients undergoing a clinically indicated CT scan with IV contrast who were 18 years or older and had one or more RCN risk factors. Exclusion Criteria: end-stage renal disease, pregnancy, clinical instability. Intervention: treatment group: 3 grams of NAC in 500 cc NS as an IV bolus, and then 200 mg per hour in 67 cc NS per hour for up to 24 hours; placebo group: 500 cc NS bolus and then 67 cc/hr NS as above. Primary outcome: RCN defined as an increase in serum creatinine by 25% or 0.5 mg/dl measured 48–72 hours after CT. Follow-up: as inpatient or by outpatient phlebotomy. Statistical Methods: comparisons made by Fisher’s exact test and logistic regression for modeling. Power calculation: assuming a 20% baseline event rate, to find a 10% absolute risk reduction, alpha = 0.05 and 90% power, we estimated needing 588 patients. O’Brien-Flemming stopping rules were specified a priori. Results: We enrolled 399 patients, of whom 357 (89%) completed follow-up and were included. The study was stopped early due to futility/equivalence. The population was well matched between groups. The RCN rate in the NAC plus NS group was 14/185 (7.6%), versus 12/172 (6.9%) in the saline only group, p = 0.83. However, we did find a strong correlation between IV fluid administration and reduced incidence of RCN. The rate of RCN in patients who received <1 liter of IV fluids was 19/158 (12.0%), compared to 7/199 (3.5%) in those who received ‡ 1 liter (p < 0.001). Our final adjusted model found that NAC, age, and CHF were not associated with RCN, but there was a 69% risk reduction (OR 0.41; 95% CI: 0.21–0.80) per liter of IVF fluids administered. Conclusion: We found no benefit in the routine administration of NAC in ED patients undergoing contrast enhanced CT above IV fluids alone, but a strong association between volume of IV fluids administered in the ED and a reduction in RCN was present. 34 Patients Transferred from an Outpatient Clinic in Rural Haiti: An Evaluation of Reason for Transfer Benjamin D. Nicholson, Harinder S. Dhindsa, Zachary W. Lipsman, Benjamin Yoon, Renee D. Reid Virginia Commonwealth University, Richmond, VA Background: The limitations of developing world medical infrastructure require that patients are transferred from health clinics only when the patient care needs exceed the level of care at the clinic and the receiving hospital can provide definitive therapy. Objectives: To determine what type of definitive care service was sought when patients were transferred from a general outpatient clinic operating Monday through Friday from 8:00 AM to 3:00 PM in rural Haiti to urban hospitals in Port-au-Prince. Methods: Design - Prospective observational review of all patients for whom transfer to a hospital was requested or for whom a clinic ambulance was requested to an off-site location to assist with patient care. Setting Weekday, daytime only clinic in Titanyen, Haiti. Participants/Subjects - Consecutive series of all patients for whom transfer to another health care facility or for whom an ambulance was requested during the time period of 11/22/2010–12/14/2010 and 3/28/2011–5/13/2011. Results: Between 11/22/2010–12/14/2010 and 3/28/2011– 5/13/2011, 37 patients were identified who needed to be transferred to a higher level of care. Sixteen patients (43.2%) presented with medical complaints, 12 (32.4%) were trauma patients, 6 (16.2%) were surgical, and 3 (8.1%) were in the obstetric category. Within these categories, 6 patients were pediatric and 4 non-trauma patients required blood transfusion. Conclusion: While trauma services are often focused on in rural developing world medicine, the need for obstetric care and blood transfusion constituted six (16.2%) cases in our sample. These patients raise important public health, planning, and policy questions relating to access to prenatal care and the need to better understand transfusion medicine utilization among rural Haitian patients with non-trauma related transfusion needs. The data set is limited by sample size and single location of collection. Another limitation of understanding the needs is that many patients may not present to the clinic for their health care needs in certain situations if they have knowledge that the resources to provide definitive care are unavailable. 35 Competency-Based Measurement of EM Learner Performance in International Training Programs James Kwan1, Cherri Hobgood2, Venkataraman Anantharaman3, Glen Bandiera4, Gautam Bodiwala5, Peter Cameron6, Pinchas Halpern7, C. James (Jim) Holliman8, Nicholas Jouriles9, Darren Kilroy10, Terrence Mulligan11, Andrew Singer12 1 Sydney Medical School, Westmead, Australia; 2 Indiana University School of Medicine, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Indianapolis, IN; 3Singapore General Hospital, Singapore, Singapore; 4St Michael’s Hospital, University of Toronto, Toronto, ON, Canada; 5 Leicester Royal Infimary, Leicester, United Kingdom; 6The Alfred Hospital Emergency and Trauma Centre, Monash University, Melbourne, Australia; 7Tel Aviv Medical Center, Tel Aviv, Israel; 8Uniformed Services University of the Health Sciences, Bethesda, MD; 9Akron Medical Center, Akron, OH; 10College of Emergency Medicine, London, United Kingdom; 11University of Maryland School of Medicine, Baltimore, MD; 12 Australian Government Department of Health and Ageing, Australian National University Medical School, The Canberra Hospital, Canberra, Australia Background: The International Federation for Emergency Medicine (IFEM) has previously defined the minimum standards for EM specialist training in a competency-based curriculum. This curriculum provides a foundation for reforming and standardizing educational practice within the international community. However, all educators regardless of their nations’ existing academic EM infrastructures require methods to assess the progress of learners, in both a formative and summative manner. Objectives: To develop recommendations for assessment for an IFEM model curriculum for EM specialists. Methods: An expert panel of medical educators from IFEM member nations was convened. Each member provided detailed information regarding training standards, assessment methods, and metrics for their host nations. In addition, an extensive literature search was performed to identify best practices that might not have been identified by the consensus process. EM curricular content items were mapped to both the CanMEDS and ACGME curricular frameworks. Similarly, assessment methods used by member nations for specific curricular elements were mapped, tabulated, and compared for determination of international consensus. Results: A range of assessment methods was identified as currently being used in member nations with established academic EM infrastructure. As there is a wide variability in the academic EM infrastructure in many member nations, feasibility was an important consideration in determining the utility of different assessment methods in these nations. The portfolio was considered to be potentially useful for assessing the performance of a trainee, allowing the triangulation of aggregated information across multiple sources of information, assessment methods, and different time points. Conclusion: These assessment recommendations provide educators, medical professionals, and experts in EM, with (i) an overview of the basic principles of designing assessment programs, (ii) a process for mapping the curriculum onto existing competency frameworks which are currently being utilized in IFEM member nations, (iii) an overview of methods of assessment and standards of assessment currently being • www.aemj.org S23 utilized in IFEM member nations, and (iv) a description of a process for aligning the curriculum with assessment methods best suited to measure the different competencies. 36 US Model Emergency Medicine in Japan Seung Young Huh1, Masaru Suzuki2, Seikei Hibino3, Takashi Shiga4, Takeshi Shimazu5, Shoichi Ohta6 1 Aizawa Hospital, Matsumoto, Japan; 2Keio University School of Medicine, Tokyo, Japan; 3 University of Minnesota Medical Center, Fairview, Minneapolis, MN; 4Tokyo Bay Urayasu Ichikawa Medical center, Urayasu, Japan; 5Osaka University School of Medicine, Osaka, Japan; 6 Tokyo Medical University, Tokyo, Japan Background: The practice of emergency medicine in Japan has been unique in that emergency physicians are mostly engaged in critical care and trauma with a multi-specialty model. For the last decade with progress in medicine, an aging population with complicated problems, and institution of postgraduate general clinical training, the US model emergency medicine with single-specialty model has been emerging throughout Japan. However, the current status is unknown. Objectives: The objective of this study was to investigate the current status of implementation of the US model emergency medicine at emergency medicine training institutions accredited by the Japanese Association for Acute Medicine (JAAM). Methods: The ER Committee of the JAAM, the most prestigious professional organization in Japanese emergency medicine, conducted the survey by sending questionnaires to 499 accredited emergency medicine training institutions. Results: Valid responses obtained from 299 facilities were analyzed. US model EM was provided in 211 facilities (71% of 299 facilities), either in full time (24 hours a day, seven days a week; 123 facilities) or in part time (less than 24 hours a day; 88 facilities). Among these 211 US model facilities, 44% have a number of beds between 251–500. The annual number of ED visits was less than 20,000 in 64%, and 37% have ambulance transfers between 2,001–4,000 per year. The number of emergency physicians was less than 5 in 60% of the facilities. Postgraduate general clinical training was offered at US model ED in 199 facilities, and ninety hospitals adopted US model EM after 2004, when a 2-year period of postgraduate general clinical training became mandatory for all medical graduates. Sixty-four facilities provided a residency program to be a US model emergency physician, and another 9 institutions were planning to establish it. Conclusion: US model EM has emerged and become commonplace in Japan. The background including advance in medicine, aging population, and mandatory postgraduate general clinical training system are considered to be contributing factors. S24 37 2012 SAEM ANNUAL MEETING ABSTRACTS Occupational Upper Extremity Injuries Treated At A Teaching Hospital In Turkey Erkan Gunay, Ersin Aksay, Ozge Duman Atilla, Nilay Zorbalar, Savas Sezik Tepecik Research and Training Hospital, Izmir, Turkey Background: Workplace safety and occupational health problems are increasing issues especially in developing countries as a result of the industrial automatisation and technologic improvements. Occupational injuries are preventable but they can occasionally cause morbidity and mortality resulting in work day loss and financial problems. Hand injuries are one-third of all traumatic injuries and are the most injured parts after occupational accidents. Objectives: We aim to evaluate patients with occupational upper extremity injuries for demographic characteristics, injury types, and work day loss. Methods: Trauma patients over 15 years old admitted to our emergency department with an occupational upper extremity injury were prospectively evaluated from 15.04.2010 to 30.04.2011. Patients with one or more of digit, hand, forearm, elbow, humerus, and shoulder injuries were included. Exclusion criteria were multitrauma, patient refusal to participate, and insufficient data. Patients were followed up from the hospital information system and by phone for work day loss and final diagnosis. Results: During the study period there were 570 patients with an occupational upper extremity injury. Total of 521 (91.4%) patients were included. Patients were 92.1% male, 36.5% between the age 25 to 34, and mean age was calculated 32.9 ± 9.6 years. 43.8% of the patients were from the metal and machinery sector, and primary education was the highest education level for the 74.7% of the patients. Most injured parts were fingers with the highest rate for index finger and thumb. Crush injury was the most common injury type. 96.3% (n = 502) of the patients were discharged after treatment in the emergency department. Tendon injuries, open fractures, and high degree burns were the reasons for admission to clinics. Mean work day loss was 12.8 ± 27.2 days and this increases for the patients with laboratory or radiologic studies, consultant evaluation, or admission. The 15–24 age group had a significantly lower work day loss average. Conclusion: Evaluating occupational injury characteristics and risks is essential for identifying preventive measures and actions. With the guidance of this study preventive actions focusing on high-risk sectors and patients may be the key factor for avoiding occupational injuries and creating safer workplace environments in order to reduce financial and public health problems. 38 Mixed Methods Evaluation of Emergency Physician Training Program in India Erika D. Schroeder1, Anne Daul2, Kate Douglass1, Punidha Kaliaperumal3, Mary Pat McKay1 1 George Washington University, Washington, DC; 2Emory University, Atlanta, GA; 3Max Hospital, New Delhi, India Background: As emergency medicine (EM) gains increased recognition and interest in the international arena, a growing number of training programs for emergency health care workers have been implemented in the developing world through international partnerships. Objectives: To evaluate the quality and appropriateness of an internationally implemented emergency physician training program in India. Methods: Physicians participating in an internationally implemented EM training program in India were recruited to participate in a program evaluation. A mixed methods design was used including an online anonymous survey and semi-structured focus groups. The survey assessed the research, clinical, and didactic training provided by the program. Demographics and information on past and future career paths were also collected. The focus group discussions centered around program successes and challenges. Results: Fifty of 59 eligible trainees (85%) participated in the survey. Of the respondents, the vast majority were Indian; 16% were female, and all were between the ages of 25 and 45 years (mean age 31 years). All but two trainees (96%) intend to practice EM as a career. One-third listed a high-income country first for preferred practice location and half listed India first. Respondents directly endorsed the program structure and content, and they demonstrated gains in self-rated knowledge and clinical confidence over their years of training. Active challenges identified include: (1) insufficient quantity and inconsistent quality of Indian faculty, (2) administrative barriers to academic priorities, and (3) persistent threat of brain drain if local opportunities are inadequate. Conclusion: Implementing an international emergency physician training program with limited existing local capacity is a challenging endeavor. Overall, this evaluation supports the appropriateness and quality of this partnership model for EM training. One critical challenge is achieving a robust local faculty. Early negotiations are recommended to set educational priorities, which includes assuring access to EM journals. Attrition of graduated trainees to high-income countries due to better compensation or limited in-country opportunities continues to be a threat to long-term local capacity building. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 39 An Analysis of Proposed Core Curriculum Elements for International Emergency Medicine and Global Health Fellowships Gabrielle Jacquet1, Alexander Vu1, Bill Ewen2, Bhakti Hansoti3, Steven Andescavage4, David Price5, Robert Suter2, Jamil Bayram1 1 Johns Hopkins University, Baltimore, MD; 2 University of Texas Southwestern, Dallas, TX; 3 University of Chicago, Chicago, IL; 4George Washington University, Washington, DC; 5 Gwinnett Medical Center, Lawrenceville, GA Background: With an increasing frequency and intensity of manmade and natural disasters, and a corresponding surge in interest in international emergency medicine (IEM) and global health (GH), the number of IEM and GH fellowships is constantly growing. There are currently 34 IEM and GH fellowships, each with a different curriculum. Several articles have proposed the establishment of core curriculum elements for fellowship training. To the best of our knowledge, no study has examined whether IEM and GH fellows are actually fulfilling these criteria. Objectives: This study sought to examine whether current IEM and GH fellowships are consistently meeting these core curricula. Methods: An electronic survey was administered to current IEM and GH fellowship directors, current fellows, and recent graduates of a total of 34 programs. Survey respondents stated their amount of exposure to previously published core curriculum components: EM System Development, Humanitarian Assistance, Disaster Response, and Public Health. A pooled analysis comparing overall responses of fellows to those of program directors was performed using two-sampled t-test. Results: Response rates were 88% (n = 30) for program directors and 53% (n = 17) for current and recent fellows. Programs varied significantly in terms of their emphasis on and exposure to six proposed core curriculum areas: EM System Development, EM Education Development, Humanitarian Aid, Public Health, EMS, and Disaster Management. Only 43% of programs reported having exposure to all four core areas. As many as 67% of fellows reported knowing their curriculum only somewhat or not at all prior to starting the program. Conclusion: Many fellows enter IEM and GH fellowships without a clear sense of what they will get from their training. As each fellowship program has different areas of curriculum emphasis, we propose not to enforce any single core curriculum. Rather, we suggest the development of a mechanism to allow each fellowship program to present its curriculum in a more transparent manner. This will allow prospective applicants to have a better understanding of the various programs’ curricula and areas of emphasis. 40 • www.aemj.org S25 Predicting ICU Admission and Mortality at Triage using an Automated Computer Algorithm Steven Horng1, David A. Sontag2, David T. Chiu1, Joshua W. Joseph1, Nathan I. Shapiro1, Larry A. Nathanson1 1 Beth Israel Deaconess Medical Center / Harvard Medical School, Boston, MA; 2New York University, New York, NY Background: Advance warning of probable intensive care unit (ICU) admissions could allow the bed placement process to start earlier, decreasing ED length of stay and relieving overcrowding conditions. However, physicians and nurses poorly predict a patient’s ultimate disposition from the emergency department at triage. A computerized algorithm can use commonly collected data at triage to accurately identify those who likely will need ICU admission. Objectives: To evaluate an automated computer algorithm at triage to predict ICU admission and 28-day in-hospital mortality. Methods: Retrospective cohort study at a 55,000 visit/ year Level I trauma center/tertiary academic teaching hospital. All patients presenting to the ED between 12/16/2008 and 10/1/2010 were included in the study. The primary outcome measure was ICU admission from the emergency department. The secondary outcome measure was 28-day all-cause in-hospital mortality. Patients discharged or transferred before 28 days were considered to be alive at 28 days. Triage data includes age, sex, acuity (emergency severity index), blood pressure, heart rate, pain scale, respiratory rate, oxygen saturation, temperature, and a nurse’s free text assessment. A Latent Dirichlet Allocation algorithm was used to cluster words in triage nurses’ free text assessments into 500 topics. The triage assessment for each patient is then represented as a probability distribution over these 500 topics. Logistic regression was then used to determine the prediction function. Results: A total of 94,973 patients were included in the study. 3.8% were admitted to the ICU and 1.3% died within 28 days. These patients were then randomly allocated to train (n = 75,992; 80%) and test (n = 18,981; 20%) data sets. The area under the receiver operating characteristic curve (AUC) when predicting ICU S26 2012 SAEM ANNUAL MEETING ABSTRACTS admission at triage was 0.91 (derivation) and 0.90 (validation). The AUC for predicting 28-day mortality was 0.94 (derivation) and 0.92 (validation). Conclusion: Triage computer algorithms can accurately identify patients who will require ICU admission using entirely automated techniques. These predictions can be made at the earliest point in a patient’s ED stay, several hours before a formal bed request is made. Predicting future resource utilization is essential in systems modeling and other approaches to systematically decrease ED crowding. 41 Replacing Traditional Triage with a Rapid Evaluation Unit Decreases Left-WithoutBeing-Seen Rate at a Community Emergency Department Dominic Ruocco1, Jeffrey P. Green2, Gladys Sillero1, Tony Berger2 1 Palisades Medical Center, Palisades, NJ; 2UC Davis Medical Center, Davis, CA Background: Many patients who leave the emergency department (ED) without being seen (LWBS) are seriously ill and would benefit from immediate evaluation. Objectives: To determine whether replacing traditional ED triage with a Rapid Evaluation Unit (REU) would decrease the proportion of patients who LWBS at a community ED. Methods: This was a pre/post interventional study design. Setting: A 155-bed community hospital with 35,000 annual ED visits. Intervention: On March 15, 2008 traditional ED triage (focused history & vital sign assessment with patient return to waiting room) was replaced with a REU, which entailed immediate placement of ambulatory patients into a treatment area. This allowed for parallel task performance by registration, triage, nursing, and providers. No significant changes to the ED or hospital physical structure, staffing, or other patient processing changes occurred during the study period. The primary outcome was change in monthly average number of patients who LWBS as a proportion of monthly ED census for a 10-month period prior to (May 2007- February 2008) and after REU initiation (May 2008–February 2009). ED throughput times for the same time period were also analyzed. These time periods represented the longest contiguous periods before and after REU initiation not involving other staffing/infrastructure changes or the immediate intervention period (March-April 2008). Results: The average monthly ED census increased by 102 visits (95%CI )11 to 215 visits) from the pre-REU to the post-REU study-period. In spite of this increase, the average monthly proportion of patients who LWBS decreased from 3.6% in the 10-month pre-REU period to 1.9% in the 10-month post-REU period, representing a 1.7% absolute decrease (95%CI 0.64, 2.72%) and a 53% relative decrease. Similarly, in the post-REU period, time from ED arrival to placement in a treatment area decreased by 35 minutes (95%CI 30.9, 39.3 min.), time from ED arrival to provider evaluation decreased by 28 minutes (95%CI 22.1, 33.9 min.), and ED length-of-stay decreased by 43 minutes (95%CI 20.1, 65.0 min.) compared to the pre-REU time period. Conclusion: Replacing traditional ED triage with a REU decreased the proportion of ED patients who LWBS as well as throughput times at a community ED. This improvement occurred without significantly changing ED or hospital physical structure, staffing, or other process changes. 42 Failure to Validate Hospital Admission Prediction Models Adding Coded Chief Complaint to Demographic, Emergency Department Operational, and Patient Acuity Data Available at ED Triage Neal Handly1, Arvind Venkat2, Jiexun Li3, David A. Thompson4, David M. Chuirazzi2 1 Drexel University College of Medicine, Philadelphia, PA; 2Allegheny General Hospital, Pittsburgh, PA; 3Drexel University School of Information Science and Technology, Philadelphia, PA; 4Northwestern University School of Medicine, Chicago, IL Background: At the 2011 SAEM Annual Meeting, we presented the derivation of two hospital admission prediction models adding coded chief complaint (CCC) data from a published algorithm (Thompson et al. Acad Emerg Med 2006;13:774–782) to demographic, ED operational, and acuity (Emergency Severity Index (ESI)) data. Objectives: We hypothesized that these models would be validated when applied to a separate retrospective cohort, justifying prospective evaluation. Methods: We conducted a retrospective, observational validation cohort study of all adult ED visits to a single tertiary care center (census: 49,000/yr) (4/1/09–12/31/10). We downloaded from the center’s clinical tracking system demographic (age, sex, race), ED operational (time and day of arrival), ESI, and chief complaint data on each visit. We applied the derived CCC hospital admission prediction models (all identified CCC categories and CCC categories with significant odds of admission from multivariable logistic regression in the derivation cohort) to the validation cohort to predict odds of admission and compared to prediction models that consisted of demographic, ED operational, and ESI data, adding each category to subsequent models in a stepwise manner. Model performance is reported by areaunder-the-curve (AUC) data and 95%CI. Results: 85,144 ED visits were included (51.2% female, 22.6% age>65, 66.1% Caucasian) with 22,363 visits resulting in hospital admission. 15.8% of visits were at night (12 am–8 am) and 27.3% on weekends. ESI percentages were as follows: Level 1: 2.3, 2: 9.0, 3: 55.5, 4: 25.6 and 5: 7.6. All descriptive characteristics were comparable to the derivation cohort. The demographics and ED operational variable model had an AUC of 0.712 (95%CI 0.708–0.715). Adding ESI data, the AUC was 0.815 (95%CI 0.812–0.819). Adding all 213 identified CCC categories to the demographics, ED operational, and ESI model, the AUC was 0.698 (95%CI 0.694– 0.702). Adding the 87 CCC categories with significant odds of admission to the demographics, ED ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 operational, and ESI acuity model, the AUC was 0.812 (95%CI 0.811–0.818). Conclusion: In this application to a retrospective validation cohort of previously derived hospital admission prediction models adding CCC to demographic, ED operational, and ESI data, we did not find CCC models’ performance strength that would justify prospective evaluation. 43 A Decision Tree Algorithm to Assist Pre-ordering Diagnostics on Emergency Department Patients During Triage Gerald Maddalozzo, Greg Neyman, Greg Sorkin, Joseph Calabro St Michael’s Medical Center, Newark, NJ Background: With more patients accessing emergency departments (ED) nationally, waiting times and lengths of stay are on the rise. Patients are lingering longer times with unappreciated critical illness, and patient dissatisfaction and walk out rates are increasing. Preordering diagnostic blood work or radiology, prior to physician evaluation, can decrease length of stay (LOS), by reducing a secondary delay when the physician waits for results after having evaluated the patient. Over-ordering inappropriately adds to costs, creates results which may be lost to proper follow-up, and tips patients towards more expensive tests or admission. Objectives: To see if a Decision Tree Algorithm (DTA), generated from a national dataset, can appropriately predict physician test orders, using inputs that are available at triage. Methods: The National Hospital Ambulatory Medical Care Survey - Emergency Department (NHAMCS ED) 2009 database was used for algorithm generation. A C4.5 Decision DTA was trained on demographics, nursing home residence, ambulance arrival, vital • www.aemj.org S27 signs, pain level, triage level, 72-hour return, number of past visits in the previous year, injury, and one of 122 chief complaint codes (representing 90% of all visits in the database). Outputs for training included ordering of a complete blood count, basic chemistry (electrolytes, blood urea nitrogen, creatinine), cardiac enzymes, liver function panel, urinalysis, electrocardiogram, x-ray, computed tomography, or ultrasound. Once trained, it was used on the NHAMCS-ED 2008 database, and predictions were generated. Predictions were compared with documented physician orders. Outcomes included the percent of total patients who were correctly pre-ordered, sensitivity (the percent of patients who had an order that were correctly predicted), and the percent over-ordered. Waiting time for correctly pre-ordered patients was highlighted, to represent a potential reduction in length of stay achieved by preordering. LOS for patients overordered was highlighted to see if over-ordering may cause an increase in LOS for those patients. Unit cost of the test was also highlighted, as taken from the 2011 Medicare fee schedule. Results: Patients correctly pre-ordered may experience a median LOS reduction of 30 minutes. Conclusion: A DTA can assist in pre-ordering tests for ED patients upon triage of the patient. 44 Reducing ED Length Of Stay For Dischargeable Patients: Advanced Triage And Now Advanced Disposition Jeff Dubin1, H. Joseph Blumenthal2, David Roseman2, David Milzman1 1 MedStar Washington Hospital Center, 2 Washington, DC; MedStar Institute for Innovation, Washington, DC Background: Advanced triage by a physician/nurse team reduces left without being seen rates and door to Table - Abstract 43: Results All Ordered All Predicted Correctly Pre-ordered Sensitivity Complete Blood Count Basic Chemistry 36.57% 37.65% 21.81% 59.64% 26.50% 25.78% 12.29% 46.38% Cardiac Enzymes 14.17% 10.23% 3.55% 25.06% Liver Function Panel Urinalysis 9.73% 10.96% 1.92% 19.75% 23.22% 22.32% 9.71% 41.82% Electrocardiogram 18.06% 15.34% 7.03% 38.95% X-ray 34.63% 32.89% 17.32% 50.03% Computed Tomography Ultrasound 13.78% 13.96% 4.38% 31.81% 3.31% 1.89% 0.41% 12.39% Test Median wait time in minutes, correctly pre-ordered (Q1–Q2) 32.08 (15.41–63.53) 31.06 (14.7–62.28) 26.79 (12.82–53.22) 30.81 (15.4–58.95) 35.64 (17.32–69.84) 27.69 (13.11–55.5) 32.59 (16.12–62.92) 30.49 (14.39–61.33) 38.66 (18.72–76.02) Overordered 143.32% 150.90% 147.14% 192.87% 154.30% 146.02% 144.97% 169.55% 144.78% Median LOS in minutes, over-ordered (Q1–Q3) 140.26 (82.32–232.61) 167.26 (96.04–282.94) 194.74 (112.19–328.41) 189.21 (107.16–324.01) 172.25 (97.44–295.31) 182.02 (103.13–311.6) 147.74 (83.48–253.52) 165.06 (93.4–282.86) 200.92 (124.42–317.44) Unit cost $12.31 $16.09 $53.07 $15.53 $6.02 $22.96 $43.06 $296.36 $226.42 S28 2012 SAEM ANNUAL MEETING ABSTRACTS physician times. However, during peak ED census times, many patients with completed tests and treatment initiated by triage await discharge by the next assigned physician. Objectives: Determine if a physician-led discharge disposition (DD) team can reduce the ED length of stay (LOS) for patients of similar acuity who are ultimately discharged compared to standard physician team assignment. Methods: This prospective observational study was performed from 10/2010 to 10/2011 at an urban tertiary referral academic hospital with an annual ED volume of 87,000 visits. Only Emergency Severity Index Level 3 patients were evaluated. The DD team was scheduled weekdays from 14:00 until 23:00. Several ED beds were allocated to this team. The team was comprised of one attending physician and either one nurse and a tech or two nurses. Comparisons were made between LOS for discharged patients originally triaged to the main ED side who were seen by the DD team versus the main side teams. Time from triage physician to team physician, team physician to discharge decision time, and patient age were compared by unpaired t-test. Differences were studied for number of patients receiving x-rays, CT scan, labs, and medications. Results: DD team mean LOS in hours for discharged patients was shorter at 3.4 (95% CI: 3.3–3.6, n = 1451) compared to 6.4 (95% CI: 6.3–6.5, n = 4601) on the main side, p < 0.01. The mean time from triage physician to DD team physician was 1.4 hours (95% CI: 1.4–1.5, n = 1447) versus to 2.7 hours (95% CI: 2.7–2.8, n = 4568) to main side physician, p < 0.01. The DD team physician mean time to discharge decision was 1.0 hour (95% CI: 1.0–1.1, n = 1432) compared to 2.5 hours (95% CI: 2.4– 2.6, n = 4590) for main side physician, p < 0.01. The DD team patients’ mean age was 42.6 years (95% CI: 41.9– 43.6, n = 1454) compared to main side patients’ mean age of 49.1 years (95% CI: 48.5–49.6, n = 4621.) The DD team patients (n = 1454) received fewer x-rays (40% vs. 59%), CT scans (13% vs. 23%), labs (64% vs. 85%), and medications (63% vs. 68%) than main side patients (n = 4621), p < 0.01 for all compared. Conclusion: The DD team complements the advanced triage process to further reduce LOS for patients who do not require extended ED treatment or observation. The DD team was able to work more efficiently because its patients tended to be younger and had fewer lab and imaging tests ordered by the triage physician compared to patients who were later seen on the ED main side. 45 ED Boarding is Associated with Increased Risk of Developing Hospital-Acquired Pressure Ulcers Candace McNaughton, Wesley H. Self, Keith Wrenn, Stephan Russ Vanderbilt University, Nashville, TN Background: Hospital-acquired pressure ulcers (HAPUs) are reportable hospital-acquired conditions and ‘‘never events’’ according to Centers for Medicaid and Medicare Services (CMS). Patients boarded in the ED for extended periods of time may not receive the same preventive care for HAPU that in-patients do, such as frequent repositioning, padding vulnerable skin, and specialized mattresses. Objectives: To evaluate the association between ED boarding time and the risk of developing HAPU. Methods: We conducted a retrospective cohort study using administrative data from an academic medical center with an adult ED with 55,000 annual patient visits. All patients admitted into the hospital through the ED 6/30/2008–2/28/2011 were included. Development of HAPU was determined using the standardized, national protocol for CMS reporting of HAPU. ED boarding time was defined as the time between an order for inpatient admission and transport of the patient out of the ED to an in-patient unit. We used a multivariate logistic regression model with development of a HAPU as the outcome variable, ED boarding time as the exposure variable, and the following variables as covariates: age, sex, initial Braden score, and admission to an intensive care unit (ICU) from the ED. The Braden score is a scale used to determine a patient’s risk for developing a HAPU based on known risk factors. A Braden score is calculated for each hospitalized patient at the time of admission. We included Braden score as a covariate in our model to determine if ED boarding time was a predictor of HAPU independent of Braden Score. Results: Of 46,704 patients admitted to the hospital through the ED during the study period, 243 developed a HAPU during their hospitalization. Clinical characteristics are presented in the table. Per hour of ED boarding time, the adjusted OR of developing a HAPU was 1.02 (95% CI 1.01–1.04, p = 0.007). A median of 40 patients per day were admitted through the ED, accumulating 144 hours of ED boarding time per day, with each hour of boarding time increasing the risk of developing a HAPU by 2%. Conclusion: In this single-center, retrospective study, longer ED boarding time was associated with increased risk of developing a HAPU. Table - Abstract 45: Clinical characteristics Age, median years (IQR) Male, no. (%) Braden Score, median (IQR) Developed HAPU, No. (%) Admitted to an ICU, No. (%) ED boarding time hrs, median (IQR) 46 51 (34,66) 22,741 (48) 20 (18, 22) 243 (0.52) 13,163 (28) 3.6 (1.7, 9.7) Nursing Attitudes Regarding Boarding of Admitted ED Patients John Richards, Bryce Pulliam UC Davis Medical Center, Sacramento, CA Background: Boarding admitted patients in the emergency department (ED) is a major cause of overcrowding and access block. One solution is boarding admitted patients in inpatient ward hallways. This study ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 queried ED and inpatient nurses and compared their opinions toward inpatient boarding. It also assessed their preferred boarding location if they were patients. Objectives: This study queried ED and inpatient nurses and compared their opinions toward inpatient boarding. Methods: A survey was administered to a convenience sample of ED and ward nurses. It was performed in a 631-bed academic medical center (30,000 admissions/yr) with a 68-bed ED (60,000 visits/yr). Nurses were identified as ED or ward and whether they had previously worked in the ED. The nurses were asked if there were any circumstances where admitted patients should be boarded in the ED or inpatient hallways. They were also asked their preferred location if they were admitted as a patient. Six clinical scenarios were then presented and their opinions on boarding queried. Results: Ninety nurses completed the survey; 35 (39%) were current ED nurses (cED), 40 (44%) had previously worked in the ED (pED). For the entire group 46 (52%) believed admitted patients should board in the ED. Overall, 52 (58%) were opposed to inpatient boarding, with 20% of cED versus 83% of current ward (cW) nurses (P < 0.0001) and 28% of pED versus 85% of nurses never having worked in the ED (nED) opposed (P < 0.001). If admitted as patients themselves, overall 43 (54%) preferred inpatient boarding, with 82% of cED versus 33% of cW nurses (P < 0.0001) and 74% of pED versus 34% nED nurses (P = 0.0007) preferring inpatient boarding. For the six clinical scenarios, significant differences in opinion regarding inpatient boarding existed in all but two cases: a patient with stable COPD but requiring oxygen and an intubated, unstable sepsis patient. Conclusion: Ward nurses and those who have never worked in the ED are more opposed to inpatient boarding than ED nurses and nurses who have worked previously in the ED. Nurses admitted as patients seemed to prefer not being boarded where they work. ED and ward nurses seemed to agree that unstable or potentially unstable patients should remain in the ED. 47 Randomized Controlled Trial of Volume-based Staffing Brian H. Rowe1, Trevor Lashyn1, Mira Singh1, Stephanie Couperthwaite1, Cristina Villa-Roel1, Michael Bullard1, William Sevcik1, Karen Latoszek2, Brian R. Holroyd1 1 University of Alberta, Edmonton, AB, Canada; 2 Alberta Health Services, Edmonton, AB, Canada Background: Emergency department (ED) overcrowding is a common and growing problem. Among throughput interventions, volume-based staffing has been described infrequently. Objectives: This study evaluated the effect of adding an additional shift in a moderate case-complexity area of a typical crowded urban, high-volume, and academic centre with severe ED overcrowding. Methods: This un-blinded, parallel group, controlled trial took place between 24/06 and 24/08/2011 at a Canadian hospital experiencing growing and long-standing ED overcrowding. Computerized, block-randomized sequences (2-week blocks) were generated for a total of • www.aemj.org S29 8 weeks. Staff satisfaction was evaluated through pre/ post-shift and study surveys; administrative data (physician initial assessment (PIA), length of stay (LOS), patients leaving without being seen (LWBS) and against medical advice [LAMA]) were collected from an electronic, real-time ED information system. Data are presented as proportions and medians with interquartile ranges (IQR); bivariable analyses were performed. Results: ED physicians and nurses expected the intervention to reduce the LOS of discharged patients only. PIA decreased during the intervention period (68 vs 74 minutes; p < 0.001). No statistically/clinically significant differences were observed in the LOS; however, there was a significant reduction in the LWBS (4.7% to 3.5% p = 0.003) and LAMA (0.7% to 0.4% p = 0.028) rates. While there was a reduction of approximately 5 patients seen per physician in the affected ED area, the total number of patients seen on that unit increased by approximately 10 patients/day. Overall, compared to days when there was no extra shift, 61% of emergency physicians stated their workload decreased and 73% felt their stress level at work decreased. Conclusion: While this study didn’t demonstrate a reduction in the overall LOS, it did reduce PIA times and the proportion of LWBS/LAMA patients. While physicians saw fewer patients during the intervention study period, the overall patient volume increased and satisfaction among ED physicians was rated higher. 48 Provider- and Hospital-Level Variation In Admission Rates And 72-Hour Return Admission Rates Jameel Abualenain1, William Frohna2, Robert Shesser1, Ru Ding1, Mark Smith2, Jesse M. Pines1 1 The George Washington University, Washington, DC; 2Washington Hospital Center, Washington, DC Background: Decisions for inpatient versus outpatient management of ED patients are the most important and costliest decision made by emergency physicians, but there is little published on the variation in the decision to admit among providers or whether there is a relationship between a provider’s admission rate and the proportion of their patients who return within 72 hours of the initial visit and are subsequently admitted (72HRA). Objectives: We explored the variation in provider-level admission rates and 72H-RA rates, and the relationship between the two. Methods: A retrospective study using data from three EDs with the same information system over varying time periods: Washington Hospital Center (WHC) (2008–10), Franklin Square Hospital Center (FSHC) (2006–9), and Union Memorial Hospital (UMH) (2005–9). Patients were excluded if left without being seen, left against medical advice, fast-track, psychiatric patients, and aged <18 years. Physicians with <500 ED encounters or an admission rate <15% were excluded. Logistic regression was used to assess the relationship between physician-level 72H-RA and admission rates, adjusting for patient age, sex, race, and hospital. S30 2012 SAEM ANNUAL MEETING ABSTRACTS Results: 389,120 ED encounters were treated by 90 physicians. Mean patient age was 50 years SD 20, 42% male, and 61% black. Admission rates differed between hospitals (WHC = 40%, UMH = 37%, and FSHC = 28%), as did the 72H-RA (WHC = 0.9%, UMH = 0.6%, and FSHC = 0.6%). Across all hospitals, there was great variation in individual physician admission rates (15.4%–50.0%). The 72H-RA rates were quite low, but demonstrated a similar magnitude of individual variation (0.3%–1.2%). Physicians with the highest admission rate quintile had lower odds of 72H-RA (OR 0.8 95% CI 0.7–0.9) compared to the lowest admission rate quintile, after adjusting for other factors. No intermediate admission rate quintiles (2nd, 3rd, or 4th) were significantly different from the lowest admission rate quintile with regard to 72H-RA. Conclusion: There is more than three-fold variation in individual physician admission rates indicating great variation among physicians in hospital admission rates and 72H-RA. The highest admitters have the lowest 72H-RA; however, evaluating the causes and consequences of such significant variation needs further exploration, particularly in the context of health reform efforts aimed at reducing costs. 49 Emergency Medicine Resident Physician Attitudes about the Introduction of a Scribe Program at an Academic EM Training Site Mia Tanaka, Jordan Marshall, Christopher Verdick, Richard C. Frederick, George Z. Hevesy, Huaping Wang, John W. Hafner University of Illinois College of Medicine at Peoria, Peoria, IL Background: ED scribes have become an effective means to assist emergency physicians (EPs) with clinical documentation and improve physician productivity. Scribes have been most often utilized in busy community EDs and their utility and functional integration into an academic medical center with resident physicians is unknown. Objectives: To evaluate resident perceptions of attending physician teaching and interaction after introduction of scribes at an EM residency training program, measured through an online survey. Residents in this study were not working with the scribes directly, but were interacting indirectly through attending physician use of scribes during ED shifts. Methods: An online ten question survey was administered to 31 residents of a Midwest academic emergency medicine residency program (PGY1–PGY3 program, 12 annual residents), 8 months after the introduction of scribes into the ED. Scribes were introduced as EMR documentation support (Epic 2010, Epic Systems Inc.) for attending EPs while evaluating primary patients and supervising resident physicians. Questions investigated EM resident demographics and perceptions of scribes (attending physician interaction and teaching, effect on resident learning, willingness to use scribes in the future), using Likert scale responses (1 minimal, 9 maximum) and a graduated percentage scale used to quantify relative values, where applicable. Data were analyzed using Kruskal-Wallis and Mann-Whitney U tests. Results: Twenty-one of 31 EM residents (68%) completed the survey (81% male; 33% PGY1, 29% PGY2, 38% PGY3). Four residents had prior experience with scribes. Scribes were felt to have no effect on attending EPs direct resident interaction time (mean score 4.5, SD 1.2), time spent bedside teaching (4.8, SD 0.9), or quality of teaching (4.9, SD 0.8), as well as no effect on residents’ overall learning process (4.6, SD 1.1). However, residents felt positive about utilizing scribes at their future occupation site (6.0, SD 2.7). No response differences were noted for prior experience, training level, or sex. Conclusion: When scribes are introduced at an EM residency training site, residents of all training levels perceive it as a neutral interaction, when measured in terms of perceived time with attending EPs and quality of the teaching when scribes are present. 50 The Effect of Introduction of an Electronic Medical Record on Resident Productivity in an Academic Emergency Department Shawn London, Christopher Sala University of Connecticut School of Medicine, Farmington, CT Background: There are little available data which describe the effect of implementation of an electronic medical record (EMR) on provider productivity in the emergency department, and no studies which, to our knowledge, address this issue pertaining to housestaff in particular. Objectives: We seek to quantify the changes in provider productivity pre- and post-EMR implementation to support our hypothesis that resident clinical productivity based on patients seen per hour will be negatively affected by EMR implementation. Methods: The academic emergency department at Hartford Hospital, the principle clinical site in the University of Connecticut Emergency Medicine Residency, sees over 95,000 patients on an annual basis. This environment is unique in that pre-EMR, patient tracking and orders were performed electronically using the Sunrise system (Eclipsys Corp) for over 8 years prior to conversion to the Allscripts ED EMR in October, 2010 for all aspects of ED care. The investigators completed a random sample of days/evening/night/weekend shift productivity to obtain monthly aggregate productivity data (patients seen per hour) by year of training. Results: There was an initial 4.2% decrease of in productivity for PGY-3 residents on average from 1.44 patients per hour on average in the three blocks preceding activation of the EMR to 1.38 patients seen per hour compared in the subsequent three prior blocks. PGY 3 performance returned to baseline in the subsequent three months to 1.48 patients per hour. There was no change noted in patients seen per hour of PGY1 and PGY-2 residents. Conclusion: While many physicians tend to assume that EMRs pose a significant barrier to productivity in ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 the ED, in our academic emergency department, there was no lasting change on resident productivity based on the patients seen per hour metric. The minor decrease which did occur in PGY-3 residents was transient and was not apparent 3 months after the EMR was implemented. Our experience suggests that decrease in the rate of patients seen per hour in the resident population should not be considered justification to delay or avoid implementation of an EMR in the emergency department. Block Block Block Block 51 2–4 5–7 7–9 10–13 PGY-2 PGY-3 0.83 0.89 0.91 1.01 1.25 1.30 1.32 1.44 1.44 1.38 1.48 1.52 Physician Feedback Reduces Resource Use in the Emergency Department Shabnam Jain1, Gary Frank2, Baohua Wu1, Brent Johnson1 1 Emory University, Atlanta, GA; 2Children’s Healthcare of Atlanta, Atlanta, GA Background: Variation in physician practice is widely prevalent and highlights an opportunity for quality improvement and cost containment. Monitoring resources used in the management of common pediatric emergency department (ED) conditions has been suggested as an ED quality metric. Objectives: To determine if providing ED physicians with severity-adjusted data on resource use and outcomes, relative to their peers, can influence practice patterns. Methods: Data on resource use by physicians were extracted from electronic medical records at a tertiary pediatric ED for four common conditions in mid-acuity (Emergency Severity Index level 3): fever, head injury, respiratory illness, and gastroenteritis. Condition-relevant resource use was tracked for lab tests (blood count, chemistry, CRP), imaging (chest x-ray, abdominal x-ray, head CT scan, abdominal CT scan), intravenous fluids, parenteral antibiotics, and intravenous ondansetron. Outcome measures included admission to hospital and ED length of stay (LOS); 72-hr return to ED (RR) was used as a balancing measure. Scorecards were constructed using box plots to show physicians their practice patterns relative to peers (the figure shows an example of the scorecard for gatroenteritis for one physician, showing resources use rates for IV fluids and labs). Blinded scorecards were distributed quarterly for five quarters using rolling-year averages. A pre/post-intervention analysis was performed with Sep 1, 2010 as the intervention date. Fisher’s exact and Wilcoxon rank sum tests were used for analysis. Results: We analyzed 45,872 patient visits across two hospitals (24,834 pre- and 21,038 post-intervention), comprising 17.6% of the total ED volume during the study period. Patients were seen by 100 physicians www.aemj.org S31 A Table - Abstract 50: Average patients seen per hour by PGY year PGY-1 • B (mean 462 patients/physician). The table shows overall physician practice in the pre- and post-intervention periods. Significant reduction in resource use was seen for abdominal/pelvic CT scans, head CT scan, chest x-rays, IV ondansetron, and admission to hospital. ED LOS decreased from 129 min to 126 min (p = 0.0003). There was no significant change in 72-hr return rate during the study period (2.2% pre-, 2.0% post-intervention). Conclusion: Feedback on comprehensive practice patterns including resource use and quality metrics can influence physician practice on commonly used resources in the ED. Table - Abstract 51: Change in Resource Use, Length of Stay, and Return Rate Before and After Feedback Resource/Outcome Abdomen/Pelvic CT (%) Head CT (%) Chest x-ray (%) Abdominal x-ray (%) Lab tests (%) IV antibiotics (%) IV fluids (%) IV ondansetron (%) admission (%) 52 Preintervention Postintervention p-value 1.2 0.5 <0.0001 26.0 31.7 15.7 71.1 12.0 37.8 11.6 7.4 18.9 28.9 16.3 70.4 11.1 38.8 8.1 6.7 <0.0001 0.0004 ns ns ns ns <0.0001 <0.0001 Publicly Posted Emergency Department Wait Times: How Accurate Are They? Nicholas Jouriles, Erin L. Simon, Peter L. Griffin, Jo Williams Akron General Medical Center, Akron, OH Background: Hospitals are advertising their emergency departments (ED) wait times on the internet, S32 2012 SAEM ANNUAL MEETING ABSTRACTS billboards, via iPhone application, Twitter, and text messaging. There is a paucity of data describing the accuracy of publically posted ED wait times. Objectives: To examine the accuracy of publicly posted wait times of four emergency departments within one hospital system. Methods: A prospective analysis of four ED-posted wait times in comparison to the wait times for actual patients. The main hospital system calculated and posted ED wait times every twenty minutes for all four system EDs. A consecutive sample of all patients who arrived 24/7 over a 4-week period during July and August 2011 was included. An electronic tracking system identified patient arrival date and the actual incurred wait time. Data consisted of the arrival time, actual wait time, hospital census, budgeted hospital census, and the posted ED wait time. For each ED the difference was calculated between the publicly posted ED wait time at the time of patient’s arrival and the patient’s actual ED wait time. The average wait times and average wait time error between the ED sites were compared using a two-tailed Student’s t-test. The correlation coefficient between the differences in predicted/ actual wait times was also calculated for each ED. Results: There were 8890 wait times within the four EDs included in the analysis. The average wait time (in minutes) at each facility was: 64.0 (±62.4) for the main ED, 22.0 (±22.1) for freestanding ED (FED) #1, 25.0 (±25.6) for FED #2, and 10.0 (±12.6) for the small community ED. The average wait time error (in minutes) for each facility was 31(±61.2) for the main ED, 13 (±23.65) for FED #1, 17 (±26.65) for FED #2, and 1 (±11.9) for the community hospital ED. The results from each ED were statistically significant for both average wait time and average wait time error (p < 0.0001). There was a positive correlation between the average wait time and average wait time error, with R-values of 0.84, 0.83, 0.58, and 0.48 for the main ED, FED #1, FED #2, and the small community hospital ED, respectively. Each correlation was statistically significant; however, no correlation was found between the number of beds available (budgeted-actual census) and average wait times. Conclusion: Publically posted ED wait times are accurate for facilities with less than 2000 ED visits per month. They are not accurate for EDs with greater than 4000 visits per month. 53 Reduction of Pre-analytic Laboratory Errors in the Emergency Department Using an Incentive-Based System Benjamin Katz, Daniel Pauze, Karen Moldveen Albany Medical Center, Albany, NY Background: Over the last decade, there has been an increased effort to reduce medical errors of all kinds. Laboratory errors have a significant effect on patient care, yet they are usually avoidable. Several studies suggest that up to 90% of laboratory errors occur during the pre- or post-analytic phase. In other words, errors occur during specimen collection and transport or reporting of results, rather than during laboratory analysis itself. Objectives: In an effort to reduce pre-analytic laboratory errors, the ED instituted an incentive-based program for the clerical staff to recognize and prevent specimen labeling errors from reaching the patient. This study sought to demonstrate the benefit of this incentive-based program. Methods: This study examined a prospective cohort of ED patients over a three year period in a tertiary care academic ED with annual census of 72,000. As part of a continuing quality improvement process, laboratory specimen labeling errors are screened by clerical staff by reconciling laboratory specimen label with laboratory requisition labels. The number of ‘‘near-misses’’ or mismatched specimens captured by each clerk was then blinded to all patient identifiers and was collated by monthly intervals. Due to poor performance in 2009, an incentive program was introduced in early 2010 by which the clerk who captured the most mismatched specimens would be awarded a $50 gift card on a quarterly basis. The total number of missed laboratory errors was then recorded on a monthly basis. Investigational data were analyzed using bivariate statistics. Results: In 2009 and the first month of 2010, 80,339 patients were treated in the ED with 89 errors found in the laboratory. From February 2010 through the end of the year, 65,411 patients were seen with 35 errors reaching the laboratory. The institution of the incentive program was found to have a risk reduction for preanalytic laboratory error of 0.48 (0.33, 0.71). Conclusion: The institution of an incentive program was associated with a marked reduction in preanalytic laboratory errors. 54 Comparison of Emergency Department Operation Metrics by Annual Volume Over 7 Years Daniel Handel1, James J. Augustine2, Heather L. Farley3, Charles M. Shufflebarger4, Benjamin C. Sun1, Robert E. O’Connor5 1 Oregon Health & Science University School of Medicine, Portland, OR; 2EMP, Cincinnati, OH; 3 Christiana Care Health System, Wilmington, DE; 4Indiana University, Indianapolis, IN; 5 University of Virginia, Charlottesville, VA Background: Most studies on operational research have been focused in academic medical centers, which typically have larger volumes of patients and are located in urban metropolitan areas. As CMS core measures in 2013 begin to compare emergency departments (EDs) on treatment time intervals, especially length of stay (LOS), it is important to explore if any differences exist inherent to patient volume. Objectives: The objective of this study is to look at differences in operational metrics based on annual patient census. The hypothesis is that treatment time intervals and operational metrics differ amongst these different categories. Methods: The ED Benchmarking Alliance has collected yearly operational metrics since 2004. As of 2010, there are 499 EDs providing data across the United States. EDs are stratified by annual volume for comparison in ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S33 Table - Abstract 54: Volume of ED (visits/year) >80K 60–80K 40–60K 20–40K <20K High Acuity (% CPT 4, 5, or Critical Care) Pediatric (%) Admitted (%) Transferred to Another Hospital (%) 57.2* 61.2* 61.0* 60.9* 53.0 23.8 20.9* 19.4* 21.8* 23.7 21.8* 21.1* 19.7* 16.8* 11.2 0.8* 0.9* 1.2* 1.7* 2.7 Arrived by EMS (%) Median LOS for Discharged Patients (minutes) Median LOS for Admitted Patients (minutes) Left Before Treatment Complete (%) 21.4* 18.4* 19.0* 15.7* 11.5 213.1* 180.9* 167.0* 140.1* 107.4 386.0* 350.7* 322.7* 273.4* 215.7 3.4* 3.5* 2.9* 2.1* 1.3 *p < 0.05 compared to <20K the following categories: <20K, 20–40K, 40–60K, and over 80K. In this study, metrics for EDs with <20K visits per year were compared to those of different volumes, averaged from 2004–2010. Mean values were compared to <20K visits as a reference point for statistical difference using t-tests to compare means with a p-value < 0.05 considered significant. Results: As seen in the table, a greater percentage of high acuity of patients was seen in higher volume EDs than in <20K EDs. The percentage of patients transferred to another hospital was higher in <20K EDs. A higher percentage arrived by EMS and a higher percentage were admitted in higher volume EDs when compared to <20K visits. In addition, the median LOS for both discharged and admitted patients and percentage who left before treatment was complete (LBTC) were higher in the higher volume EDs. Conclusion: Lower volume EDs have lower acuity when compared to higher volume EDs. Lower volume EDs have shorter median LOS and left before treatment complete percentages. As CMS core measures require hospitals to report these metrics, it will be important to compare them based on volume and not in aggregate. 55 Does the Addition of a Hands-Free Communication Device Improve ED Interruption Times? Amy Ernst, Steven J. Weiss, Jeffrey A. Reitsema University of New Mexico, Albuquerque, NM Background: ED interruptions occur frequently. Recently a hands-free communication device (Vocera) was added to a cell phone and a pager in our ED. Objectives: The purpose of the present study was to determine whether this addition improved interruption times. Our hypothesis was that the device would significantly decrease length of time of interruptions. Methods: This study was a prospective cohort study of attending ED physician calls and interruptions in a Level I trauma center with EM residency. Interruptions included phone calls, EKG interpretations, pages to resuscitation, and other miscellaneous interruptions (including nursing issues, laboratory, EMS, and radiology). We studied a convenience sampling intended to include mostly evening shifts, the busiest ED times. Length of time the interruption lasted was recorded. Data were collected for a comparison group pre-Vocera. Three investigators collected data including seven different addendings’ interruptions. Data were collected on a form, then entered into an Excel file. Data collectors’ agreement was determined during two additional four hour shifts to calculate a kappa statistic. SPSS was used for data entry and statistical analysis. Descriptive statistics were used for univariate data. Chi-square and Mann Whitney U nonparametric test were used for comparisons. Results: Of the total 511 interruptions, 33% were phone calls, 24% were EKGs to be read, 18% were pages to resuscitation, and 25% miscellaneous. There were no significant differences in types of interruptions pre- vs. post-Vocera. Pre-Vocera we collected 40 hours of data with 65 interruptions with a mean 1.6 per hour. Post-Vocera, 180 hours of data were collected with 446 interruptions with a mean 2.5 per hour. There was a significant difference in length of time of interruptions with an average of 9 minutes pre-Vocera vs. 4 minutes post-Vocera (p = 0.012, diff 4.9, 95% CI 1.8–8.1). Vocera calls were significantly shorter than non-Vocera calls (1 vs 6 minutes, p < 0.001). Comparing data collectors for type of interruption during the same 4-hour shift resulted in a kappa (agreement) of 0.73. Conclusion: The addition of a hands-free communication device may improve interruptions by shortening call length. 56 ‘‘Talk-time’’ In The Emergency Department: The Duration Of Patient-provider Interactions During An ED Visit Danielle M. McCarthy, Kenzie A. Cameron, Francisco Acosta, Jennifer Stancati, Victoria E. Forth, Barbara A. Buckley, Michael Schmidt, James G. Adams, Kirsten G. Engel Northwestern University, Chicago, IL Background: Analyses of patient flow through the ED typically focus on metrics such as wait time, total length of stay (LOS), or boarding time. However, little is known about how much interaction a patient has with clinicians after being placed in a room, or what proportion of the in-room visit is also spent ‘‘waiting,’’ rather than directly interacting with care providers. Objectives: The objective was to assess the proportion of time, relative to the time in a patient care area, that a patient spends actively interacting with providers during an ED visit. S34 2012 SAEM ANNUAL MEETING ABSTRACTS Methods: A secondary analysis of 29 audiotaped encounters of patients with one of four diagnoses (ankle sprain, back pain, head injury, laceration) was performed. The setting was an urban, academic ED. ED visits of adult patients were recorded from the time of room placement to discharge. Audiotapes were edited to remove all downtime and non-patient-provider conversations. LOS and door-to-doctor times were abstracted from the medical record. The proportion of time the patient spent in direct conversation with providers (‘‘talk-time’’) was calculated as the ratio of the edited audio recording time to the time spent in a patient care area (talk-time = [edited audio time/(LOS - door-to-doctor)]). Multiple linear regression controlling for time spent in patient care area, age, and sex was performed. Results: The sample was 31% male with a mean age of 37 years. Median LOS: 133 minutes (IQR: 88–169), median door-to-doctor: 42 minutes (IQR: 29–67), median time spent in patient care area: 65 minutes (IQR: 53– 106). Median time spent in direct conversation with providers was 16 minutes (IQR: 12–18), corresponding to a talk-time percentage of 19.2% (IQR: 14.7–24.6%). There were no significant differences based on diagnosis. Regression analysis showed that those spending a longer time in a patient care area had a lower percentage of talk time (b = )0.11, p = 0.002). Conclusion: Although limited by sample size, these results indicate that approximately 80% of a patients’ time in a care area is spent not interacting with providers. While some of the time spent waiting is out of the providers’ control (e.g. awaiting imaging studies), this significant ‘‘downtime’’ represents an opportunity for both process improvement efforts to decrease downtime as well as the development of innovative patient education efforts to make the best use of the remaining downtime. 57 Degradation of Emergency Department Operational Data Quality During Electronic Health Record Implementation Michael J. Ward, Craig Froehle, Christopher J. Lindsell University of Cincinnati, Cincinnati, OH Background: Process improvement initiatives targeted at operational efficiency frequently use electronic timestamps to estimate task and process durations. Errors in timestamps hamper the use of electronic data to improve a system and may result in inappropriate conclusions about performance. Despite the fact that the number of electronic health record (EHR) implementations is expected to increase in the U.S., the magnitude of this EHR-induced error is not well established. Objectives: To estimate the change in the magnitude of error in ED electronic timestamps before and after a hospital-wide EHR implementation. Methods: Time-and-motion observations were conducted in a suburban ED, annual census 35,000, after receiving IRB approval. Observation was conducted 4 weeks pre- and 4 weeks post-EHR implementation. Patients were identified on entering the ED and tracked until exiting. Times were recorded to the nearest second using a calibrated stopwatch, and are reported in minutes. Electronic data were extracted from the patient-tracking system in use pre-implementation, and from the EHR post-implementation. For comparison of means, independent t-tests were used. Chi-square and Fisher’s t-tests were used for proportions, as appropriate. Results: There were 263 observations; 126 before and 137 after implementation. The differences between observed times and timestamps were computed and found to be normally distributed. Post-implementation, mean physician seen times along with arrival to bed, bed to physician, and physician to disposition intervals occurred before observation. Physician seen timestamps were frequently incorrect and did not improve postimplementation. Significant discrepancies (ten minutes or greater) from observed values were identified in timestamps involving disposition decision and exit from the ED. Calculating service time intervals resulted in every service interval (except arrival to bed) having at least 15% of the times with significant discrepancies. It is notable that missing values were more frequent postEHR implementation. Conclusion: EHR implementation results in reduced variability of timestamps but reduced accuracy and an increase in missing timestamps. Using electronic timestamps for operational efficiency assessment should recognize the magnitude of error, and the compounding of error, when computing service times. Table - Abstract 57: Difference Between Observed versus Electronic Timestamps Following EHR Implementation Pre-Implementation Period Bed Placement Bed Placement Physician Seen Disposition Decision Exit from ED Arrival-to-Bed Bed-to-Physician Physician-toDisposition Disposition-to-Exit Post-Implementation Period Mean, min Prop. >10 Min Diff from Obs N (Missing) 120 (6) 119 (7) 98 (28) 0.38 after 0.34 before 6.59 after 6.7% 25.2% 28.6% 104 119 119 98 2.42 0.68 0.72 7.47 after before before after 19.2% 4.2% 27.7% 38.8% 0.71 before 28.4% N (Missing) (22) (7) (7) (28) 95 (31) Mean, min Prop. >10 Min Diff from Obs p-value (mean) p-value (Proportion) 70 (67) 115 (22) 107 (30) 0.19 before 6.48 before 11.21 after 2.9% 27.8% 16.8% 0.419 <0.001 0.289 0.329 0.650 0.044 81 70 69 106 4.25 2.86 8.28 18.12 after before before after 8.6% 11.4% 39.1% 39.6% 0.512 <0.001 <0.001 0.016 0.043 0.075 0.106 0.901 6.76 before 17.2% 0.112 0.061 (56) (67) (68) (31) 99 (38) ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 58 Factors Associated With Excessive Emergency Department Length Of Stay For Treated & Released Patients in an Urban Academic Medical Center Jeremy D. Sperling1, Michael J. McMahon1, Kabir Rezvankhoo2, Neal E. Flomenbaum1, Peter W. Greenwald1, Rahul Sharma1 1 Weill Cornell Medical College / NewYorkPresbyterian Hospital, New York, NY; 2NewYorkPresbyterian Hospital / Columbia University, New York, NY Background: Prolonged ED length of stay (LOS) negatively affects patient satisfaction and consumes overall ED resources. ED treated and released patients (TRP) are the patients later surveyed for satisfaction with their overall stay. EDs have more control over LOS for TRPs than they do for admitted patients. Identifying modifiable variables for TRPs, especially those with extreme LOS (greater than 90th percentile), may prove beneficial in improving both patient care and patient satisfaction scores. Objectives: To identify factors associated with extreme LOS for TRPs in a 76,000 visit urban academic tertiary care center. Methods: ED electronic tracking systems prospectively recorded demographics, chief complaints, diagnostics, and consultations obtained, and ED census. Researchers blinded to study hypotheses collated one month of consecutive adult data (excluding AMAs, walkouts). Associations between variables and extreme LOS were analyzed by single and multiple logistic and linear regression. Results: In June 2010, out of 5265 adult ED visits, 3463 (65.8%) met inclusion criteria. Mean LOS for TRPs was 4.92 hours (95%CI 4.8–5.02, range 0.1–29.5). Mean LOS for TRPs with extreme LOS was 11.8 hours (95%CI 11.5–12.1, range 9.1–29.5). Significant associations with extreme LOS included age, sex, day of week, time of day, chief complaint, ED census, diagnostic studies, and consultant or social work involvement. In multiple • www.aemj.org S35 logistic regression, only diagnostic studies (imaging and labs) and consultation (social work and specialists) variables remained significant (table). By multiple linear regression, hourly delays were (95%CI): 3.6 h (3.1–4.1) for PO contrast CT, 1.54 h (1.3–1.8) for noncontrast CT, and 4.1 h (3.4–4.8) for MRI. Conclusion: Diagnostic imaging (e.g. CT, MRI, ultrasound), labs, and social work involvement had the strongest associations with extreme LOS while patientspecific factors (e.g. age, sex) were not associated with extreme LOS after adjustment for other factors. A single-center study is unlikely to be generalizable as system inefficiencies vary by institution. However, this type of analysis can be implemented in any center and help EDs identify variables to focus on in order to reduce LOS for TRPs, especially for those with extreme LOS. Our next step will be to improve processes associated with extreme LOS and evaluate its effect on overall ED LOS. 59 Ketamine-Propofol Combination (Ketofol) versus Propofol Alone for Emergency Department Procedural Sedation and Analgesia: A Prospective Randomized Trial Gary Andolfatto1, Riyad B. Abu-Laban2, Peter J. Zed3, Sean M. Staniforth1, Sherry Stackhouse1, Susanne Moadebi1, Elaine Willman4 1 Lions Gate Hospital, North Vancouver, BC, 2 Vancouver General Hospital, Canada; Vancouver, BC, Canada; 3Queen Elizabeth II Health Sciences Centre, Halifax, NS, Canada; 4 University of British Columbia, Vancouver, BC, Canada Background: Procedural sedation and analgesia is used in the ED in order to efficiently and humanely perform necessary painful procedures. The opposing physiological effects of ketamine and propofol suggest the potential for synergy, and this has led to interest in their combined use, commonly termed ‘‘ketofol’’, to facilitate ED procedural sedation. Objectives: To determine if a 1:1 mixture of ketamine and propofol (ketofol) for ED procedural sedation results in a 13% or more absolute reduction in adverse respiratory events compared to propofol alone. Methods: Participants were randomized to receive either ketofol or propofol in a double-blind fashion according to a weight-based dosing protocol. Inclusion criteria were age 14 years or greater, and ASA Class 1–3 status. The primary outcome was the number and proportion of patients experiencing an adverse respiratory event according to pre-defined criteria (the ‘‘Quebec Criteria’’). Secondary outcomes were sedation consistency, sedation efficacy, induction time, sedation time, procedure time, and adverse events. Results: A total of 284 patients were enrolled, 142 per group. Forty-three (30%) patients experienced an adverse respiratory event in the ketofol group compared to 46 (32%) in the propofol group (difference 2%; 95% CI )9% to 13%; p = 0.798). Thirty-eight (27%) patients receiving ketofol and 36 (25%) receiving propofol developed hypoxia, of whom three (2%) S36 2012 SAEM ANNUAL MEETING ABSTRACTS Table 1 - Abstract 59: Respiratory Events and Interventions Patients experiencing a respiratory event Oxygen desaturation Central apnea Partial upper airway obstruction Complete upper airway obstruction Received airway positioning / stimulation Also received oxygen Also received bag-valve-mask Ketofol, No., (%)[95% CI] N = 142 Propofol, No., (%)[95% CI] N = 142 Difference, % (95% CI) 43 (30) [23 to 38] 38 (27) [20 to 35] 15 (11) [7 to 17] 11 (8) [4 to 13] 6 (4) [2 to 9] 5 (4) [2 to 8] 35 (25) [18 to 32] 3 (2) [0.7 to 6) 46 (32) [25 to 41] 36 (25) [19 to 33] 13 (9) [6 to 15] 11 (8) [4 to 13] 4 (3) [1 to 7] 14 (10) [6 to 16] 31 (22) [16 to 29] 1 (1) [0.1 to 4] 2 2 2 0 1 6 3 1 (-9 to 13) p = 0.798 (-9 to 12) (-5 to 9) (-3 to 6) (0.4 to 13) (-7 to 13) (-2 to 5) Table 2 - Abstract 59: Drug Dosage and Sedation Time Intervals Total medication dose mg/Kg Total medication dose mL/Kg Sedation time, min Procedure time, min Recovery time, min Time to Ramsay Sedation Score 5, min Number doses required to reach Ramsay Sedation Score 5 Number patients Ramsay Sedation Score <5 or requiring repeat dosing during procedure Efficacious sedation Ketofol, (n = 142) Median (IQR) [range] Propofol, (n = 142) Median (IQR) [range] 0.7 (0.6 to 0.9) [0.4 to 1.6] 0.14 (0.12 to 0.18)[0.08 to 0.32] 7 (4 to 9)[1 to 29] 4 (2 to 7) [1 to 28] 8 (7 to 10) [1 to 26] 2 (1 to 3) [1 to 6] 2 (2 to 3) [1 to 7] 1.5 (1.1 to 2.0) [0.7 to 5.1] 0.15 (0.11 to 0.20) [0.07 to 0.51] 7 (4 to 9) [1 to 18] 5 (2 to 7) [1 to 13] 6 (2 to 8) [2 to 13] 2 (1 to 3) [1 to 10] 2 (1 to 3) [1 to 9] 65 (46) [38 to 54] 93 (65) [57 to 73] difference 19% (8% to 31%) p = 0.001 129 (91) [85 to 95] 126 (89) [83 to 93] ketofol patients and 1 (1%) propofol patient received bag-valve-mask ventilation. Sixty-five (46%) patients receiving ketofol and 93 (65%) receiving propofol required repeat medication dosing or lightened to a Ramsay Sedation Score of 4 or less during their procedure (difference 19%; 95% CI 8% to 31%; p = 0.001). Procedural agitation occurred in 5 patients (3.5%) receiving ketofol compared to 15 (11%) receiving propofol (difference 7.5%, 95% CI 1% to 14%). Recovery agitation requiring treatment occurred in six patients (4%, 95% CI 2.0% to 8.9%) receiving ketofol. Other secondary outcomes were similar between the groups. Patients and staff were highly satisfied with both agents. Conclusion: Ketofol for ED procedural sedation does not result in a reduced incidence of adverse respiratory events compared to propofol alone. Induction time, efficacy, and sedation time were similar; however, sedation depth appeared to be more consistent with ketofol. 60 The Effect of CMS Guideline on Deep Sedation withPropofol Lindsay Harmon1,AnthonyJ.Perkins1, Beth Sandford2 1 Indiana University School of Medicine, Indianapolis, IN; 2Wishard Hospital, Indianapolis, IN Background: Emergency physicians routinely perform emergency department procedural sedation (EDPS) with propofol and its safety is well established. However, in 2010 CMS enacted guidelines defining propofol as deep sedation and requiring administration by a physician. Common EDPS practice had been one physician performing both the sedation and procedure. EDPS has proven safe under this one-physician practice. However, the 2010 guidelines mandated separate physicians perform each. Objectives: The study hypothesis was that one-physician propofol sedation complication rates are similar to two-physician. Methods: Before and after, observational study of patients >17 years of age consenting to EDPS with propofol. EDPS completed with one physician were compared to those completed with two (separate physicians performing the sedation and the procedure). All data were prospectively collected. The study was completed at an urban Level I trauma center. Standard monitoring and procedures for EDPS were followed with physicians blinded to the objectives of this research. The frequency and incremental dosing of medication was left to the discretion of the treating physicians. The study protocol required an ED nurse trained in data collection to be present to record vital signs and assess for any prospectively defined complications. We used chi-square tests to compare the binary outcomes and ASA scores across the time periods, and two-sample t-tests to test for differences in age between the two time periods. Results: During the 2-year study period we enrolled 481 patients: 252 one-physician EDPS sedations and 229 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 two-physician. All patients meeting inclusion criteria were included in the study. Total adverse event rates were 4.4% and 3.1%, respectively (p = 0.450). The most common complications were hypotension and oxygen desaturation, and they respectively showed one-physcian rates of 2.0% and 0.8% and two-physician rates of 1.8% and 0.9% (p = 0.848 and 0.923.) The unsuccessful procedure rates were 4.0% vs 3.9% (p = 0.983). Conclusion: This study demonstrated no significant difference in complication rates for propofol EDPS completed by one physician as compared to two. The Use of End-Tidal CO2 Monitoring in Patients Undergoing Observation for Sedative Overdose in the Emergency Department James Miner, Johanna Moore, Jon B. Cole Hennepin County Medical Center, Minneapolis, MN Background: Overdose patients are often monitored using pulse oximetry, which may not detect changes in patients on high-flow oxygen. Objectives: To determine whether changes in end-tidal carbon dioxide (ETCO2) detected by capnographic monitoring are associated with clinical interventions due to respiratory depression (CRD) in patients undergoing evaluation for a decreased level of consciousness after a presumed drug overdose. Methods: This was a prospective, observational study of adult patients undergoing evaluation for a drug overdose in an urban county ED. All patients received supplemental oxygen. Patients were continuously monitored by trained research associates. The level of consciousness was recorded using the Observer’s Assessment of Alertness/Sedation scale (OAA/S). Vital signs, pulse oximetry, and OAA/S were monitored and recorded every 15 minutes and at the time of occurrence of any CRD. Respiratory rate and ETCO2 were measured at five second intervals using a CapnoStream20 monitor. CRD included an increase in supplemental oxygen, the use of bag-valve-mask ventilations, repositioning to improve ventilation, and physical or verbal stimulus to induce respiration, and were performed at the discretion of the treating physicians and nurses. Changes from baseline in ETCO2 values and waveforms among patients who did or did have a clinical intervention were compared using Wilcoxon rank sum tests. Results: 100 patients were enrolled in the study (age 35, range 18 to 67, 62% male, median OAAS 4, range 1 to 5). Suspected overdoses were due to opioids in 34, benzodiazepines in 14, an antipsychotic in 14, and others in 38. The median time of evaluation was 165 minutes (range 20 to 725). CRD occurred in 47% of patients, including an increase in O2 in 38%, repositioning in 14%, and stimulation to induce respiration in 23%. 16% had an O2 saturation of <93% (median 88, range 73 to 92) and 8% had a loss of ETCO2 waveform at some time, all of whom had a CRD. The median www.aemj.org S37 change in ETCO2 from baseline was 5 mmHg, range 1 to 30. Among patients with CRD it was 14 mmHg, range 10 to 30, and among patients with no CRD it was 5 mmHg, range 1 to 13 (p = 0.03). Conclusion: The change in ETCO2 from baseline was larger in patients who required clinical interventions than in those who did not. In patients on high-flow oxygen, capnographic monitoring may be sensitive to the need for airway support. 62 61 • How Reliable Are Health Care Providers in Reporting Changes in ETCO2 Waveform Anas Sawas1, Scott Youngquist1, Troy Madsen1, Matthew Ahern1, Camille BroadwaterHollifield1, Andrew Syndergaard1, Jared Phelps2, Bryson Garbett1, Virgil Davis1 1 University of Utah, Salt Lake City, UT; 2 Midwestern University, Glendale, AZ Background: ETCO2 changes have been used in procedural sedation analgesia (PSA) research to evaluate subclinical respiratory depression associated with sedation regiments. Objectives: To evaluate the accuracy of bedside clinician reporting of changes in ETCO2. Methods: This was a prospective, randomized, singleblind study conducted in ED setting from June 2010 until the present time. This study took place at an academic adult ED of a 405-bed (21 in the ED) and a Level I trauma center. Subjects were randomized to receive either ketamine-propofol or propofol according to a standardized protocol. Loss of ETCO2 waveforms for ‡ 15 sec were recorded. Following sedation, questionnaires were completed by the sedating physicians. Digitally recorded ETCO2 waveforms were also reviewed by an independent physician and a trained research assistant (RA). To ensure the reliability of trained research assistants, we compared their analyses with the analyses of an independent physician for the first 41 recordings. The target enrollment was 65 patients in each group (N = 130 total). Statistics were calculated using SAS statistical software. Results: 91 patients were enrolled; 53 (58.2%) are males and 38 (41.8%) are females. Mean age was 44.93 ± 17.93 years. Most participants did not have major risk factors for apnea or for further complications (86.3% were ASA class 1 or 2). ETCO2 waveforms were reviewed by 87 (95.6%) sedating physicians and 84 (92.3%) nurses at the bedside. There were 70 (76.9%) ETCO2 waveforms recordings, 42 (60.0%) were reviewed by an independent physician and 70 (100%) were reviewed by an RA. A kappa test for agreement between independent physicians and RAs was conducted on 41 recordings and there were no discordant pairs (kappa = 1). Compared to sedating physicians, the independent physician was more likely to report ETCO2 wave losses (OR 1.37, 95% CI 1.08–1.73). Compared to sedating physicians, RAs were more likely to report ETCO2 wave losses (OR 1.39, 95% CI 1.14–1.70). S38 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: Compared to sedating physicians at the bedside, independent physicians and RAs were more likely to note ETCO2 waveform losses. An independent review of recorded ETCO2 waveform changes will be more reliable for future sedation research. 63 Effectiveness and Safety in Rapid Sequence Intubation Versus Non-Rapid Sequence intubation in the Emergency Department: Multi-center Prospective Observational Study in Japan Masashi Okubo1, Yusuke Hagiwara2, Kohei Hasegawa3 1 Okinawa Chubu Hospital, Okinawa, Japan; 2 Kawaguchi Medical Center, Kawaguchi, Japan; 3 Massachusetts General Hospital, Boston, MA Background: Comprehensive studies evaluating current practices of ED airway management in Japan are lacking. Many emergency physicians in Japan still experience resistance regarding rapid sequence intubation (RSI). Objectives: We sought to describe the success and complication rate of RSI with non-RSI. Methods: Design and Setting: We conducted a multicenter prospective observational study using the JEAN registry of EDs at 11 academic and community hospitals in Japan during between 2010 and 2011. Data fields include ED characteristics, patient and operator demographics, method of airway management, number of attempts, and adverse events. We defined non-RSI as intubation with sedation only, neuromuscular blockade only, and without medication. Participants: All patients undergoing emergency intubation in ED were eligible for inclusion. Cardiac arrest encounters were excluded from the analysis. Primary analysis: We described RSI with non-RSI in terms of success rate on first attempt, within three attempts, and complication rate. We present descriptive data as proportions with 95% confidence intervals (CIs). We report odds ratios (OR) with 95% CI via chi-square testing. Results: The database recorded 2710 intubations (capture rate 98%) and 1670 met the inclusion criteria. RSI was the initial method chosen in 489 (29%) and non-RSI in 1181 (71%). Use of RSI varied among institutes from 0% to 79%. Success cases of RSI on first and within three attempts are 353 intubations (72%, 95%CI 68%– 76%) and 474 intubations (97%, 95%CI 95%–98%), respectively. The success cases of non-RSI on first and within three attempts are 724 intubations (61%, 95%CI 58%–64%) and 1105 intubations (94%, 95%CI 92%– 95%). Success rates of RSI on first and within three attempts are higher than non-RSI (OR 1.64, 95%CI 1.30–2.06 and OR 2.14, 95% CI 1.22–3.77, respectively). We recorded 67 complications in RSI (14%) and 165 in non-RSI (14%). There is no significant difference of complication rate between RSI and non-RSI (OR 0.98, 95% CI 0.72–1.32). Conclusion: In this multi-center prospective study in Japan, we demonstrated a high degree of variation in use of RSI for ED intubation. Additionally we found that success rate of RSI on first and within attempts were both higher than non-RSI. study has the limitation of reporting bias and founding by indication. (Originally submitted ‘‘late-breaker.’’) 64 three This conas a How Can Technology Help Us Further Interpret ETCO2 Changes? Anas Sawas, Scott Youngquist, Troy Madsen, Matthew Ahern, Camille Broadwater-Hollifield, Andrew Syndergaard, Bryson Garbett, Virgil Davis University of Utah, Salt Lake City, UT Background: ETCO2 changes have been used in procedural sedation analgesia research to evaluate clinical respiratory depression (CRP). Objectives: To determine if the number of episodes and the duration of lost ETCO2 are better predictors of CRP compared to any loss of ETCO2. Methods: This was a prospective, randomized, singleblind study conducted in the ED setting from June 2010 until the present time. This study took place at an academic adult ED of a 405-bed (21 in the ED) and a Level I trauma center. Subjects were randomized to receive either ketamine-propofol or propofol according to a standardized protocol. ETCO2 waveforms were digitally recorded. ETCO2 changes were evaluated by the sedating physicians at the bedside. Recorded waveforms were reviewed by an independent physician and a trained research assistant (RA). To ensure the reliability of trained RAs, we computed a kappa test for agreement between the analysis of independent physicians and RAs for the first 41 recordings. A post-hoc analysis of the association between any loss, the number of losses, and total duration of loss of ETCO2 waveform and CRP was performed. On review we recorded the absence or presence of loss of ETCO2 and the total duration in seconds of all lost ETCO2 episodes ‡15 seconds. ORs were calculated using SAS statistical software. Results: 91 patients were enrolled; 53 (58.2%) are males and 38 are (41.8%) females. 86.3% participants were ASA class 1 or 2. Waveforms were reviewed by 87 (95.6%) sedating physicians. There were 70 (76.9%) waveforms recordings, 42 (60.0%) were reviewed by an independent physician and 70 (100%) were reviewed by RAs, where there were no discordant pairs (kappa = 1). There were 24 (26.4%) CRP events. Any loss of ETCO2 was associated with a non-significant OR of 4.06 (95% CI 0.75–21.9) for CRP. However, the duration of ETCO2 loss was significantly associated with CRP with an OR of 1.38 (95% CI 1.08–1.76) for each 30 second interval of lost ETCO2. The number of losses was significantly associated with the outcome (OR 1.48, 95% CI 1.15–1.91). Conclusion: Defining subclinical respiratory depression as present or absent may be less useful than quantitative measurements. This suggests that risk is cumulative over periods of loss of ETCO2, and the duration of loss may be a better marker of sedation depth and risk of complications than classification of any loss. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 65 One-Year Peer Violence Outcomes Following a Brief Motivational Interviewing Intervention for Violence and Alcohol among Teens Rebecca M. Cunningham, Lauren K. Whiteside, Stephen T. Chermack, Marc A. Zimmerman, Maureen A. Walton University of Michigan, Ann Arbor, MI Background: ED visits present an opportunity to deliver brief interventions (BIs) to reduce violence and alcohol misuse among urban adolescents at risk for future injury. Previous analyses demonstrated that a brief intervention resulted in reductions in violence and alcohol consequences up to 6 months. Objectives: This paper describes findings examining the efficacy of BIs on peer violence and alcohol misuse at 12 months. Methods: Patients (14–18 yrs) at an ED reporting past year alcohol use and aggression were enrolled in the RCT, which included computerized assessment, and randomization to control group or BI delivered by a computer (CBI) or therapist assisted by a computer (TBI). Baseline and 12 months included violence (peer aggression, peer victimization, violence related consequences) and alcohol (alcohol misuse, binge drinking, alcohol-related consequences). Results: 3338 adolescents were screened (88% participation). Of those, 726 screened positive for violence and alcohol use and were randomized; 84% completed 12-month follow-up. As compared to the control group, the TBI group showed significant reductions in peer aggression (p < 0.01) and peer victimization (p < 0.05) at 12 months. BI and control groups did not differ on alcohol-related variables at 12 months. Conclusion: Evaluation of the SafERteens intervention one year following an ED visit provides support for the efficacy of computer-assisted therapist brief intervention for reducing peer violence. 66 Violence Against ED Health Care Workers: A 9-Month Experience Terry Kowalenko1, Donna Gates2, Gordon Gillespie2, Paul Succop2 1 University of Michigan, Ann Arbor, MI; 2 University of Cincinnati, Cincinnati, OH Background: Health care (HC) support occupations have an injury rate nearly 10 times that of the general sector due to assaults, with doctors and nurses nearly 3 times greater. Studies have shown that the ED is at greatest risk of such events compared to other HC settings. Objectives: To describe the incidence of violence in ED HC workers over 9 months. Specific aims were to 1) identify demographic, occupational, and perpetrator factors related to violent events; 2) identify the predictors of acute stress response in victims; and 3) identify predictors of loss of productivity after the event. Methods: Longitudinal, repeated methods design was used to collect monthly survey data from ED HC • www.aemj.org S39 workers (W) at six hospitals in two states. Surveys assessed the number and type of violent events, and feelings of safety and confidence. Victims also completed specific violent event surveys. Descriptive statistics and a repeated measure linear regression model were used. Results: 213 ED HCWs completed 1795 monthly surveys, and 827 violent events were reported. The average per person violent event rate per 9 months was 4.15. 601 events were physical threats (3.01 per person in 9 months). 226 events were assaults (1.13 per person in 9 months). 501 violent event surveys were completed, describing 341 physical threats and 160 assaults with 20% resulting in injuries. 63% of the physical threats and 52% of the assaults were perpetrated by men. Comparing occupational groups revealed significant differences between nurses and physicians for all reported events (p = 0.0048), with the greatest difference in physical threats (p = 0.0447). Nurses felt less safe than physicians (p = 0.0041). Physicians felt more confident than nurses in dealing with the violent patient (p = 0.013). Nurses were more likely to experience acute stress than physicians (p < 0.001). Acute stress significantly reduced productivity in general (p < 0.001), with a significant negative effect on ‘‘ability to handle/ manage workload’’ (p < 0.001) and ‘‘ability to handle/ manage cognitive demands’’ (p < 0.05). Conclusion: ED HCWs are frequent victims of violence perpetrated by visitors and patients. This violence results in injuries, acute stress, and loss of productivity. Acute stress has negative consequences on the workers’ ability to perform their duties. This has serious potential consequences to the victim as well as the care they provide to their patients. 67 A Randomized Controlled Feasibility Trial of Vacant Lot Greening to Reduce Crime and Increase Perceptions of Safety Eugenia C. Garvin, Charles C. Branas Perelman School of Medicine at the University of Pennsylvania, Philadelphia, PA Background: Vacant lots, often filled with trash and overgrown vegetation, have been associated with intentional injuries. A recent quasi-experimental study found a significant decrease in gun crimes around vacant lots that had been greened compared with control lots. Objectives: To determine the feasibility of a randomized vacant lot greening intervention, and its effect on police-reported crime and perceptions of safety. Methods: For this randomized controlled feasibility trial of vacant lot greening, we partnered with the Pennsylvania Horticulture Society (PHS) to perform the greening intervention (cleaning the lots, planting grass and trees, and building a wooden fence around the perimeter). We analyzed police crime data and interviewed people living around the study vacant lots (greened and control) about perceptions of safety before and after greening. Results: A total of 5200 sq ft of randomly selected vacant lot space was successfully greened. We used a master database of 54,132 vacant lots to randomly S40 2012 SAEM ANNUAL MEETING ABSTRACTS select 50 vacant lot clusters. We viewed each cluster with the PHS to determine which were appropriate to send to the City of Philadelphia for greening approval. The vacant lot cluster highest on the random list to be approved by the City of Philadelphia was designated the intervention site, and the next highest was designated the control site. Overall, 29 participants completed baseline interviews, and 21 completed follow-up interviews after 3 months. 59% of participants were male, 97% were black or African American, and 52% had a household income less than $25,000. Unadjusted difference-in-differences estimates showed a decrease in gun assaults around greened vacant lots compared to control. Regression-adjusted estimates showed that people living around greened vacant lots reported feeling safer after greening compared to those who lived around control vacant lots (p < 0.01). Conclusion: Conducting a randomized controlled trial of vacant lot greening is feasible. Greening may reduce certain gun crimes and make people feel safer. However, larger prospective trials are needed to further investigate this link. 68 Screening for Violence Identifies Young Adults at Risk for Return ED Visits for Injury Abigail Hankin-Wei, Brittany Meagley, Debra Houry Emory University, Atlanta, GA Background: Homicide is the second leading cause of death among youth ages 15–24. Prior studies, in nonhealth care settings, have shown associations between violent injury and risk factors including exposure to community violence, peer behavior, and delinquency. Objectives: To assess whether self-reported exposure to violence risk factors can be used to predict future ED visits for injuries. Methods: We conducted a prospective cohort study in the ED of a Southeastern US Level I trauma center. Patients aged 15–24 presenting for any chief complaint were included unless they were critically ill, incarcerated, or could not read English. Recruitment took place over six months, by a trained research assistant (RA). The RA was present in the ED for 3–5 days per week, with shifts scheduled such that they included weekends and weekdays, over the hours from 8 am–8 pm. Patients were offered a $5 gift card for participation. At the time of initial contact in the ED, patients completed a written questionnaire which included validated measures of the following risk factors: a) aggression, b) perceived likelihood of violence, c) recent violent behavior, d) peer behavior, e) community exposure to violence, and f) positive future outlook. At 12 months following the initial ED visit, the participants’ medical records were reviewed to identify any subsequent ED visits for injury-related complaints. Data were analyzed with chi-square and logistic regression analyses. Results: 332 patients were approached, of whom 300 patients consented. Participants’ average age was 21.1 years, with 57% female, and 86% African American. Return visits for injuries were significantly associated with hostile/aggressive feelings (RR 3.7, CI 1.42,9), self-reported perceived likelihood of violence (RR 5.16, CI 1.93, 13.78), recent violent behavior (RR 3.16, CI 1.01, 9.88), and peer group violence (RR 4.4, CI 1.72, 11.25). These findings remained significant when controlling for participant sex. Conclusion: A brief survey of risk factors for violence is predictive of return visit to the ED for injury. These findings identify a potentially important tool for primary prevention of violent injuries among young adults visiting the ED for both injury and non-injury complaints. 69 Firearm Possession among Adolescents and Young Adults presenting to an Urban Emergency Department for Assault Patrick M. Carter1, Manya Newton1, Lauren Whiteside1, Kevin Loh2, Maureen A. Walton3, Marc Zimmerman4, Rebecca M. Cunningham1 1 University of Michigan, School of Medicine, Department of Emergency Medicine; University of Michigan Injury Center, Ann Arbor, MI; 2 University of Michigan School of Public Health, University of Michigan Injury Center, Ann Arbor, MI; 3University of Michigan Department of Psychiatry, Ann Arbor, MI; 4University of Michigan, School of Public Health, Department of Health Behavior and Health Education, Ann Arbor, MI Background: Violence is the leading cause of death among African American youth. Firearms are a leading cause of death in adolescents, and youth illicit gun carriage and possession is a risk factor for violent injury. Identification of assault-injured youth who own a gun is an important component of violence prevention. Objectives: 1) To determine rates and correlates of gun possession, how and why firearms are obtained, and type of firarms common among assaulted youth seeking care; 2) To understand differences in risk factors (prior violence, substance use) among these youth seeking care with and without possession of a gun. Methods: 14–24 yr old patients presenting to a Level I ED with violent injury over a 12 mo period were administered a computerized screening survey. Validated instruments were administered measuring demographics, attitudes towards aggression, substance use, and prior violence history. Results: 718 assault-injured youth completed the survey (84.6% participation); 163 (28%) possessed a gun, 117 (71.8%) were male, mean 19.8 yrs, 103 (63.2%) on public assistance. Bivariate analysis found, compared to those without a gun, patients possessing a gun were more likely involved in a recent physical altercation (85.9% vs. 78.9% p < 0.05), use alcohol before fights (33.7% vs 19.5%, p < 0.01), or have caused injury requiring treatment (46.0% vs 20.5%, p < 0.001). They were more likely to have shot or stabbed another person (8.0% vs 2.5%, p < 0.01) or used a gun in a fight with a dating partner (6.8% vs 0.7%, p < 0.001). Most patients kept a gun ‘‘for protection’’ (37.1%), and obtained guns illegally from friends (23.3%). Among ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 assaulted youth possessing a gun, more believed ‘‘revenge was a good thing’’ (2.9 vs. 3.1, p < 0.05) and it was ‘‘ok to hurt people if they hurt you first’’ (2.6 vs. 2.8, p < 0.01). Logistic regression results indicated male sex (AOR 3.55; 95%CI 2.32–5.44), illicit drug use (AOR 1.51, 95% CI 1.01–2.24), recent involvement in a serious fight (AOR 1.71; 95%CI 1.01–2.90), any previous dating violence (AOR 2.58; 95%CI 1.67–4.00), and attitudes favoring retaliation (AOR 1.56; 95%CI 1.08–2.22) predicted gun possession among assault-injured youth. Conclusion: Among assaulted youth patients seeking care, 28% have gun possession. Future violence prevention efforts should address risk of future injury or assault and substance use in this high risk population. Funding: NIDA RO1 024646 70 Does Cellulitis Belong in an Observation Unit? Louisa Canham, Kathryn A. Volz, Emily Kaplan, Leon D. Sanchez, Christopher Fischer, Shamai A. Grossman BIDMC, Boston, MA Background: Cellulitis is a common patient diagnosis in emergency department (ED) observation units (OBS). Prior studies suggest that as many as 38% of ED OBS cellulitis patients may ultimately need full hospital admission, versus a national average of 15% for other diseases. There are little data to help predict patients who are likely to fail a trial of ED OBS. Objectives: To identify characteristics in patients with cellulitis that are predictive of OBS failure. Methods: Retrospective cohort review of 405 consecutive patients who were seen in an urban, academic ED with >55,000 visits and admitted to OBS with a diagnosis of skin infection/cellulitis. Data analysis used t-test and chi square. Results: 377 patients met study criteria; 29.2% were admitted to the inpatient unit after a stay in OBS. There was no significant difference in average age of admitted vs discharged patients (47.2 vs 45.6, p = 0.39), nor did sex have an effect on admission (28.3% females admitted vs 30.6% of males, p = 0.74). There was a higher admission rate in skin infections of the hand (43.4%) vs other body locations (p < 0.001). Other rates by location were torso (33.3%), head/neck (30.6%), arm (23.2%), foot (20%), leg (19.5%), and buttock (0%). Diabetes was not predictive, with a 26.7% admission rate vs 29.5% in non-diabetics (p = 0.69). Patients with IVDU history were no more likely to require admission than those who reported no drug use (41.7% vs 28.8%, p = 0.33). 133 patients received some oral antibiotics as outpatients prior to their ED OBS stay; these patients were not more likely to be admitted (31% rate of admission vs 27.9% in those who had not received oral antibiotics, p = 0.45). Of patients treated with prior antibiotics, 61% had received MRSA coverage. Patients treated with PO MRSA coverage did not have different admission rates than those without MRSA coverage (28.6% vs 36%, p = 0.37). Conclusion: In this study group, almost one-third of patients admitted to ED OBS with cellulitis ultimately • www.aemj.org S41 were admitted to an inpatient unit. We found no significant difference in admission rates based on age, sex, diabetes, IVDU, or prior course of oral antibiotics. Patients with hand infections had the highest admission rates compared to other body locations. Cellulitis is more likely than other diagnoses to require inpatient care, decisions to observe patients based on location of infection may result in a more judicious use of ED OBS. 71 Comparison of a Novel Clinical Prediction Rule, MEDS, SIRS, and CURB-65 in the Prediction of Hospital Mortality for Septic Patients Visiting the Emergency Department Kuan-Fu Chen, Chun-Kuei Chen, Sheng-Che Lin, Peng-Hui Liu, Jih-Chang Chen, Te-Fa Chiu Chang-Gung Memorial Hospital, Taoyuan County, Taiwan Background: Sepsis is a commonly encountered disease in ED, with high mortality. While several clinical prediction rules (CPR) including MEDS, SIRS, and CURB-65 exist to facilitate clinicians in early recognition of risk of mortality for sepsis, most are of suboptimal performance. Objectives: To derive a novel CPR for mortality of sepsis utilizing clinically available and objective predictors in ED. Methods: We retrospectively reviewed all adult septic patients who visited the ED at a tertiary hospital during the year 2010 with two sets of blood cultures ordered by physicians. Basic demographics, ED vital signs, symptoms and signs, underlying illnesses, laboratory findings, microbiological results, and discharge status were collected. Multivariate logistic regressions were used to obtain a novel CPR using predictors with <0.1 p-value tested in univariate analyses. The existing CPRs were compared with this novel CPR using AUC. Results: Of 8699 included patients, 7.6% died in hospital, 51% had diabetes, 49% were older than 65 years of age, 21% had malignancy, and 16% had positive blood bacterial culture tests. Predisposing factors including history of malignancy, liver disease, immunosuppressed status, chronic kidney disease, congestive heart failure, and older than 65 years of age were found to be associated with mortality (all p < 0.05). Patients who developed mortality tended to have lower body temperature, narrower pulse pressure, higher percentage of red cell distribution width (RDW) and bandemia, higher blood urea nitrogen (BUN), ammonia, and C-reactive protein level, and longer prothrombin time and activated partial thromboplastin time (aPTT) (all p < 0.05). The most parsimonious CPR incorporating history of malignancy (OR 2.3, 95% CI 1.9–2.7), prolonged aPTT (3.0, 2.4–3.8), presence of bandemia (1.7, 1.4–2.0), higher BUN level (2.0, 1.7–2.4), and wider RDW (2.6, 2.2–3.1) were developed with better AUC (0.77, 95% CI 0.75–0.79), compared to CURB-65 (0.68, 0.66–0.70), MEDS (0.66, 0.64–0.68), and SIRS (0.54, 0.52–0.56) (all p < 0.05). Conclusion: We derived a novel clinical prediction rule for mortality owing to sepsis using a history of malignancy, prolonged aPTT, presence of bandemia, S42 2012 SAEM ANNUAL MEETING ABSTRACTS Objectives: To quantify the overuse of CT in MTBI in the ED based upon current guideline recommendations. HYPOTHESIS: Physicians are not obtaining imaging consistent with guidelines, leading to unnecessary CTs. Methods: DESIGN: Retrospective analysis of secondary data from a prospective observational study. SETTING: Urban, Level I ED with >90,000 visits per year. SUBJECTS: Adult patients with blunt MTBI, Glasgow Coma Scale score 13–15, non-focal neurologic exam, and receiving CT imaging for trauma at the discretion of the treating physician from March 2006-April 2007. OBSERVATIONS: Proportion of cases meeting criteria for CT based on the CCTHR, ACEP Clinical Policy, and NOC were to be reported using descriptive statistics. Results: All 346 patients enrolled in the original study were included in the analysis. The proportion of cases meeting criteria for CT were: CCTHR 64.7% (95% CI 0.60–0.70), ACEP 74.3% (95% CI 0.70–0.79), and NOC 90.5% (95% CI 0.87–0.94). Sensitivities of the guidelines were: CCTHR 79.2% (95% CI 0.58–0.93), ACEP 95.8% (95% CI 0.80–0.99), and NOC 91.7% (95% CI 0.74–0.98). Conclusion: 10–35% of CTs obtained in the ED for MTBI are not guideline-compliant. Successful implementation of these guidelines could decrease CT use in MTBI by up to 35%. Performing CT according to guidelines would save up to $394 million annually and prevent roughly 36 radiation-induced cancers. LIMITATIONS: This analysis was limited by the data collected in the original study. To account for this limitation, guideline compliance was overestimated. 73 higher BUN level, and wider RDW with better performance of existing CPRs. Further validation study is merited. 72 Overuse of CT for Mild Traumatic Brain Injury Edward R. Melnick1, Christopher M. Szlezak1, Suzanne K. Bentley2, Lori A. Post1 1 Yale School of Medicine, New Haven, CT; 2 Mount Sinai School of Medicine, New York, NY Background: 1.4 million patients with mild traumatic brain injury (MTBI) are treated and released from EDs annually in the United States. To differentiate MTBI from clinically important brain injury and to prevent the overuse of CT, multiple evidence-based guidelines exist to direct appropriate use of CT. The sensitivity and specificity of the Canadian CT Head Rule (CCTHR) and New Orleans Criteria (NOC) have been previously validated. Despite conventional implementation strategies, CT use is growing rapidly with up to 75% of these patients receiving imaging. Over-testing exposes patients to undue radiation risk and cost. Doctor Knows Best: Published Guidelines vs. ED Physicians’ Predictions Of ACS Amisha Parekh, Robert Birkhahn, Paris Ayana Datillo New York Methodist Hospital, Brooklyn, NY Background: Guidelines are meant to encompass the needs of the majority of patients in many situations. While they may be generalizable to a primary care setting, in the unique environment of an emergency department (ED), the feasibility of incorporating AHA/ ACC/ACEP guidelines into the clinical workflow remains in question. Often the varied patient presentation and acuity of presentation requires a physician’s insight for accurate diagnosis, rendering guidelines ineffective for use in the ED. Objectives: To compare an ACS risk stratification rating assigned by ED physicians to the AHA/ACC/ACEP guidelines for acute coronary syndrome (ACS), and assess each for accuracy in predicting a definitive diagnosis of ACS. Methods: We conducted a prospective observational cohort study on all patients evaluated for ACS, over a 14-week time period in an urban teaching hospital ED. The patient’s risk stratification for ACS classified by the evaluating ED physician was compared to AHA/ACC/ACEP guidelines. Patients were contacted at 48 hours and 30 days following the index ED visit ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 to determine all cause mortality, unscheduled hospital/ED revisits, and objective cardiac testing performed. Results: There was poor agreement between the physician’s unstructured assessment used in clinical practice and the guidelines put forth by the AHA/ACC/ACEP task force. ED physicians were more likely to assess a patient as low risk (42%), while AHA guidelines were more likely to classify patients as intermediate (50%) or high (40%) risk. However, when comparing the patient’s final ACS diagnosis and the relation to the risk assessment value, ED physicians proved better predictors of high-risk patients who in fact had ACS, while the AHA/ACC/ACEP guidelines proved better at correctly identifying low-risk patients who did not have ACS. Conclusion: In the ED, physicians are far more efficient at correctly placing patients with underlying ACS into a high-risk category, while established criteria may be overly conservative when applied to an acute care population. Further research is indicated to look at ED physicians’ risk stratification and ensuing patient care to assess for appropriate decision making and ultimate outcomes. 74 Compartative Accuracy of the Wells Score and AMUSE Score for the Detection of Acute Lower Limb Deep Vein Thrombosis Gabriel E. Blecher1, Ian G. Stiell1, Michael Y. Woo1, Paul Pageau1, Phil Wells1, Matthew D.F McInnes1, George A. Wells2 1 University of Ottawa, Ottawa, ON, Canada; 2 University of Ottawa Heart Institute, Ottawa, ON, Canada Background: The Wells clinical decision rule is commonly used to risk-stratify patients with potential lower limb deep vein thrombosis (DVT). Recently another rule that incorporates a d-dimer was developed to address the reported lowered sensitivity amongst primary care patients. Objectives: We evaluated the test characteristics of both the Wells and AMUSE clinical decision rules in a cohort of emergency department patients who presented with suspected acute lower limb DVT. Methods: We conducted a prospective cohort study on a sample of adult patients who presented to an academic ED with suspected acute lower limb DVT. Demographic and clinical data were collected and d-dimer was measured on all patients. The outcome of interest was any fatal or non-fatal venous thromboembolism at 90 days, as determined by radiology database and medical record review, as well as structured telephone interview. The sensitivity, specificity, positive and negative predictive values were calculated for patients classified as high risk for DVT by the Wells and AMUSE scores. Results: The 148 study patients had the following characteristics: mean age 58.6 and male sex 45.3%; 1.4% had congestive heart failure, 2.7% peripheral vascular disease, and 0.7% systemic lupus erythematosis. 19.6% were taking antiplatelet agents and 5.4% anticoagulants. • www.aemj.org S43 41.7% had an elevated d-dimer and 33.3% of the cohort were classified as high-risk by the Wells score and 45.5% by the AMUSE score. The Wells score had a sensitivity of 81.0% (95% CI 58.1–94.6) and specificity of 76.3%, (95% CI 65.2–85.3) whereas the AMUSE score had a sensitivity of 90.5% (95% CI 69.6–98.8) and specificity of 64.0% (95% CI 52.1–74.8). Conclusion: The AMUSE score was more specific, but the Wells score was more sensitive for acute lower limb DVT in this cohort. There is no significant advantage in using the AMUSE over the Wells score in ED patient with suspected DVT. Table - Abstract 74: Sensitivity Specificity Positive predictive value Negative predictive value 75 Wells instrument AMUSE instrument 81.0% (95% CI 58.1–94.6) 76.3% (95% CI 65.2–85.3) 48.6% (95% CI 31.4–66.0) 93.6% (95% CI 84.3–98.2) 90.5% (95% CI 69.6–98.8) 64.0% (95% CI 52.1–74.8) 41.3% (95% CI 27.0–56.8) 96.0% (95% CI 86.3–99.5) Facial Action Coding to Enhance The Science of Pretest Probability Assessment (the FACES initiative) Melissa Haug, Virginia Krabill, David Kammer, Jeffrey Kline Carolinas Medical Center, Charlotte, NC Background: Pretest probability scoring systems do not explicitly measure appearance sickness or wellness as a predictor variable. However, emergency physicians frequently employ the observation of ‘‘(s)he looks good’’ to make real-time decisions. Objectives: Test if patient facial expressions can be quantized and used to improve pretest probability assessment. Methods: This was a prospective, non-interventional, proof of concept study of diagnostic accuracy. We programmed a laptop computer to simultaneously videotape the patient’s face while viewing an automated, 15-second presentation containing three visual stimuli slides, two humorous, and one sad. Patients were adults with a chief complaint of chest pain or dyspnea. Criterion standard: 1. D-: No significant cardiopulmonary disease within 7 d, or 2. D+: Acute disease requiring hospital care for >24 hours (PE, ACS, aortic disaster, or pneumonia, CHF, COPD or arrhythmia requiring IV medication). Two observers analyzed the videos and for each slide, they computed the Facial Action Coding System (FACS) score, based upon action units (AU, 0–5) of 43 individual facial muscle actions. IOV assessed with weighted kappa, medians with Mann-Whitney U, and diagnostic accuracy with ROC and sensitivity (sens) and specificity (spec). Results: We enrolled 50 patients (52 ± 16 years, 38% female, 46% Caucasian, 94% triage code 3 or 4 [4 highest possible]). Complete FACS scores were obtained S44 2012 SAEM ANNUAL MEETING ABSTRACTS from all 50. 12/50 (24%) were D+ and their median total AU score (7.5, IQR 1–36) was significantly lower than D- (31.5, IQR 14–62, P = 0.21 MWU) with ROC = 0.71. FACS scores were clustered for expressions of Smile, Surprise, or Frown, each with AU scores ranging from 0 (completely absent, test negative) to 22 (strongest expression). Using best response to appropriate slide, Smile had ROC = 0.85 (0.6 to 1.0), sens = 91%, spec = 64%; Surprise had ROC = 0.72 (0.40 to 1.0), sens = 82%, spec = 59%, and Frown had ROC = 0.54, sens = 82%, spec = 77%. For total AU score, weighted kappa = 0.23 (0.05 to 0.40) and for Smile, interobserver agreement within ±1 AU was 80% and weighted kappa = 0.28 (95% CI 0.1 to 0.47). Conclusion: Type and magnitude of facial expression contains useful diagnostic information. These data specifically support the diagnostic value of documenting presence of patient smile on physical examination for patients with chest complaints. dependent on the type of treatment being offered. Our study found substantially more agreement for standard treatments (scenarios 1 and 5) than for experimental/ research protocols (scenarios 2, 3, and 4). The adaptive clinical trial was more acceptable than standard RCT, although there was substantially more patient/surrogate disagreement regarding participation in the adaptive trial, potentially due to the increased complexity of the design. Further research is needed into optimal consent approaches in time sensitive, high-stakes diseases such as stroke and other acute neurological conditions. Acute Stroke Research and Treatment Consent: The Accuracy of Surrogate Decision Makers Jessica Bryant University of Michigan Medical School, Ann Arbor, MI Background: Emergency patients are often treated with weight-based fluids and medications with only a guess as to their actual weight. The patients who are too ill for a triage-obtained weight are those most likely to receive weight-based, life-saving medications such as heparin, low molecular weight heparins, thrombolytics, vasopressors, and anti-arrhythmics. There are no validated methods by which to estimate adult patient weight in the emergency department. We have previously derived a rule for converting mid-arm circumference (MAC) to weight in adult emergency patients in a mostly Hispanic population in southwest Texas. This formula is: Weight Kg = (MAC cm x 3.5) – 30. Objectives: To validate the previously derived formula for converting MAC to weight in a different patient population. Methods: This was an IRB-approved study conducted at a large, university-affiliated hospital in Kentucky seeing a predominately Caucasian population. Triaged patients had bilateral MAC measured. Age and measured triage weight were also collected. Using regression analysis, a plot and formula for the conversion of MAC to weight has been derived. This conversion factor will be compared to the previously derived conversion formula. Results: 294 patients have been evaluated. A formula converting MAC to weight has been derived and compared favorably to the previously derived formula. Age does not appear to affect the conversion formula. Conclusion: The formula for converting MAC to weight, originally derived in a mostly Hispanic population in 76 Background: Surrogate consent for treatment and research options on behalf of mentally incapacitated patients with acute stroke is currently the standard of practice in emergency departments across the nation. Many studies have, however, indicated the failure of surrogates to accurately predict patient preferences in a variety of other clinical settings. Objectives: This study is designed to investigate the hypothesis that such a failure extends to the acute stroke setting as well–that significant discrepancies exist between surrogate decisions and the preferences of the patients they represent. Methods: We performed a cross-sectional verbal survey of 200 patients in the University of Michigan ED without stroke, and the family member, friend, or significant other who had accompanied the patient, resulting in a total enrollment of 400. The patient was presented with five scenarios for treatment decisions in the event of an acute stroke or cardiac arrest. After each scenario, he or she indicated the likelihood of consenting to the treatment/protocol for each scenario. The same procedure was then performed separately on the surrogate. Results: Overall, surrogates predicted patients’ treatment preferences with 80.2% accuracy. Patient/surrogate agreement for scenarios 1, 2, and 5 was 96%, 87%, and 95% respectively. Scenarios 3 and 4–regarding a standard pharmacotherapy RCT and an adaptive RCT– gave rise to the vast majority of disagreement between patients and surrogates. In scenario 3, 69.5% of pairs refused the trial while 4.5% of pairs consented to the trial, resulting in an agreement rate of 74%. The adaptive clinical trial (scenario 4) represented that lowest rate of agreement at 49%. Conclusion: The accuracy with which surrogates were able to predict patient preferences was highly 77 Validation of Mid-Arm Circumference for a Rapid Estimate of Emergency Patient Weight Matthew N. Graber1, Christopher Belcher1, Matthew Mart1, Roba Rowe1, Patrick Jenkins1, Melchor Ortiz2 1 University of Kentucky, Lexington, KY; 2Texas Tech, El Paso, TX ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Southwest Texas, has an excellent fit to the data from a mostly Caucasian population in Kentucky. The formula is therefore validated in varying populations and can be used to rapidly estimate a patient’s weight in the emergency department. WeightKg = (MACcm x 3.5) – 30 78 Hospice and Palliative Care in the Emergency Department Afzal Beemath1, Robert Zalenski2, Cheryl Courage2, Alexandra Edelen2 1 Detroit Medical Center and Seasons Hospice and Palliative Care, Detroit, MI; 2Division of Palliative Medicine, Department of Emergency Medicine, Wayne State University School of Medicine, Detroit, MI Background: An estimated 2.4–3.6 million visits are made to hospital EDs across the United States each year by nursing home (NH) residents. Little is known regarding the frequency at which NH patients are eligible for a hospice and palliative care (HPC) referral and how many actually receive an appropriate HPC intervention during their hospitalizations. Objectives: The goal of the investigation was to examine and quantify hospice eligibility and code status in a cohort of patients transferred from nursing homes and other extended care facilities to EDs who are subsequently admitted to the hospital. Additionally, it was hypothesized that demographic variables would influence whether an HPC intervention was received and that hospital course would differ for those patients who received an HPC intervention. Methods: A retrospective chart review of a random sample (n = 97) of non-cancer terminally ill NH patients admitted to a Level I trauma center via the ED was performed. Nursing home residents who presented to the ED were excluded if they were transferred back to the NH after a period of observation of less than 24 hours. Eligibility for HPC referral was assessed using the Hospice Referral Guidelines for Non-Cancer Diagnoses, a checklist of markers of severity of diagnosis. Demographics, hospital course, and changes to code status during hospital stay were also examined. • www.aemj.org S45 Results: Almost all patients were eligible for an HPC referral (97.9%), and only 8.2% (n = 8) of patients actually received an HPC intervention. Patients who received the intervention did not differ in terms of demographics or hospital course from those who did not receive the HPC intervention. Most patients were ‘‘full code’’ at admission (72.1%), and 18.6% of patients changed their status to DNR while admitted. Those who had a goal of care discussion were significantly more likely (X2 (1, 81) = 6.99, p = 0.007) to change their code status to DNR than those who did not have a discussion. Those who had HPC referrals had higher rates of goals of care discussions, DNR status, and higher mortality. Conclusion: These findings suggest that health care providers should consider an HPC referral for all nursing home patients who enter the ED with a terminal diagnosis. Providing goals of care discussions and HPC services to patients is likely to prevent unwanted medical intervention and improve quality of life, and may also be cost effective. 79 Prospective Validation of a Prediction Instrument for Endocarditis in Febrile Injection Drug Users Hangyul M. Chung-Esaki1, Robert M. Rodriguez2, Jonathan Fortman2, Bitou Cisse3 1 UCSF, San Francisco, CA; 2San Francisco General Hospital, San Francisco, CA; 3Alameda County Medical Center, Oakland, CA Background: Diagnosis of infectious endocarditis (IE) among injection drug users (IDUs) in the ED is challenging due to poor sensitivity and specificity of individual clinical signs and laboratory tests. Our pilot derivation study derived a decision instrument (DI) based on tachycardia, lack of skin infection, and cardiac murmur, which yielded 100% sensitivity (95% CI 84– 100) and 100% negative predictive value (95% CI 88– 100) for IE in the derivation cohort. Objectives: To evaluate and validate the diagnostic performance of a previously derived DI to rule out IE in febrile IDUs using recursive partitioning and multiple logistic regression models. Table - Abstract 79: Recursive partitioning model results Tachycardia (HR >= 100) Sensitivity (95% CI) Specificity (95% CI) Positive predictive value (95% CI) Negative predictive value (95% CI) +Likelihood ratio (95% CI) )Likelihood ratio (95% CI) Odds ratio (95% CI) 0.889 0.428 0.109 0.980 1.554 0.260 6.107 (0.654–0.986) (0.363–0.495) (0.064–0.171) (0.930–0.998) (1.275–1.894) (0.070–0.967) (1.345–26.637) Cardiac murmur 0.722 0.706 0.160 0.970 2.453 0.394 6.232 (0.465–0.903) (0.642–0.764) (0.088–0.259) (0.932–0.990) (1.730–3.489) (0.186–0.833) (2.139–18.161) Absence of skin infection 0.944 0.433 0.115 0.990 1.665 0.128 12.977 (0.727–0.999) (0.368–0.500) (0.068–0.178) (0.946–1.000) (1.421–1.952) (0.019–0.867) (1.698–99.156) Decision instrument (All three) 1.000 (0.815–1.000) 0.134 (0.089–0.180) 0.083 (0.049–0.127) 1.000 (0.884–1.000) 1.149 (1.093–1.208) 0.000 Infinite S46 2012 SAEM ANNUAL MEETING ABSTRACTS Methods: Febrile IDUs admitted from two urban EDs were prospectively enrolled in June 2007 to March 2011 if they were admitted for rule out endocarditis. Clinical prediction data from the first 6 hours of ED presentation and outcome data from inpatient records were prospectively collected. Diagnosis of IE was based on Modified Duke criteria and discharge diagnosis of endocarditis. In this new cohort, we determined the screening performance of the previously derived recursive partitioning model, and derived and validated a multiple logistic regression model with the same criteria. Results: Of the 249 subjects, 18 (7.3%) had IE. The recursive partitioning derived DI had 100% sensitivity (95% CI 83.7–100), 100% negative predictive value (95% CI 90.6–100), but low specificity (13.4%, 95% CI 12.2– 13.4). The logistic regression model with the three predictors using a probability threshold of 0.13 yielded 94.4% sensitivity (95% CI 72.7–99.9) with higher specificity (58.4%, 95% CI 51.8–64.9). Conclusion: In this internal validation study, a DI using tachycardia, cardiac murmur, and absence of skin infection ruled out IE with high sensitivity (94.4–100%) and low to moderate specificity (13.4–58.4%), using recursive partitioning and logistic regression models. If external validation demonstrates similar screening performance, the DI may decrease admissions for ‘‘rule out endocarditis’’ in febrile IDU. 80 Early Predictors of Post-Concussive Syndrome in Mild Traumatic Brain Injuries Presenting to the ED Boris Vidri1, Syed Ayaz2, Patrick Medado2, Scott Millis2, Brian J. O’Neil2 1 Wayne State, Detroit, MI; 2Detroit Medical Center, Detroit, MI Background: Negative head CT does not exclude intracranial injury and does not predict post-concussive syndrome (PCS) in mild traumatic brain injury (mTBI). Objectives: Concussion Symptom Inventory (CSI) is a self-reported symptom scoring using a Likert scale that has been validated in mTBI. Standard Assessment of Concussion (SAC) is another validated scale to test neurocognitive impairment. We studied whether CSI or SAC obtained in the emergency department (ED) were predictive of PCS. Methods: The CSI and SAC surveys were assessed upon ED presentation, 3 hours post, and at 7 and 45 days after presentation. Patients presenting with potential mTBI and a Conley Score = 10 were enrolled. Controls consisted of ED patients with pain, but without head injury. PCS was defined as a CSI>12 at 45 days. Data were analyzed by simple statistics, ROC, and a linear regression model. Results: 174 patients were enrolled (55 controls, 119 mTBI) over 18 months. ROC analysis revealed a CSI cutoff >12 differentiated control from TBI patients. Of the individual CSI variables, lack of headache and light sensitivity appropriately classified 73% of mTBI and 76% of controls. The baseline CSI showed significance (p < 0.0001) in predicting 7-day CSI. Baseline CSI had moderate correlation to PCS (r = 0.39), while 7-day CSI was highly correlated to PCS (r = 0.54). The baseline SAC did not differentiate TBI from controls, mean for controls was 24.75 ± 3 and for TBI was 23.77 ± 4 (p = 0.08). The initial SAC score was not able to predict the 7-day CSI (p = 0.25). A ROC curve analysis of SAC gave c statistics for detection of PCS of: initial SAC = 0.44, 3-hour SAC = 0.39, and 45-day SAC = 0.41. Conclusion: A CSI of >12 best predicted head injury from controls. Lack of headache and light sensitivity are the best symptoms to differentiate the head injured from sham control. Baseline CSI predicts symptom retention at day 7. 7-day CSI is a better predictor of PCS than baseline CSI. The SAC does not differentiate head-injured from control and does not predict PCS. Patients with a CSI >12 would probably benefit from outpatient referral, particularly if CSI remains high at day 7. 81 Clinical Characteristics Associated with Moderate to Severe Carotid Stenosis or Dissection in Transient Ischemic Attack Patients Presenting to the Emergency Department Heather Heipel1, Mukul Sharma1, Ian G. Stiell1, Jane Sutherland2, Marco L.A. Sivilotti3, Marcel Emond4, Andrew Worster5, Grant Stotts1, Jeffrey J. Perry1 1 University of Ottawa, Ottawa, ON, Canada; 2 Ottawa Hospital Research Institute, Ottawa, ON, Canada; 3Queen’s University, Kingston, ON, Canada; 4Laval University, Laval, QC, Canada; 5 McMaster University, Hamilton, ON, Canada Background: The recent emphasis on early surgical intervention for symptomatic carotid stenosis >50% highlights the need to prioritize imaging of transient ischemic attack (TIA) patients. While clinical decision rules have sought to identify patients at high risk for stroke, there are limited data on clinical characteristics associated with significant carotid findings on imaging, including moderate to severe stenosis and dissection. Objectives: We compared clinical features of TIA patients with significant carotid findings on carotid artery imaging to those with normal or insignificant abnormalities. Methods: We prospectively enrolled adult TIA patients at eight tertiary-care academic EDs over 5 years as part of a larger TIA decision rule study. We actively followed patients with health record review and telephone calls to identify interventions and outcomes for 90 days following enrolment. Significant carotid findings were defined as stenosis >50% or dissection by catheter angiography, CTA, MRA, or carotid doppler at the discretion of the treating physician. Characteristics were recorded on standardized data forms prior to review of carotid imaging results. We conducted logistic regression for adjusted odds ratios (OR). Results: We enrolled 3,704 patients in the main TIA clinical decision rule study (mean 78.2 years; 50.5% male), of whom 3,069 (82.9%) underwent carotid imaging and represent the study population for this analysis. Overall, 2,580 (69.7%) had normal or <50% stenosis, 41 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 (1.1%) had near to complete stenosis, 435 (11.7%) had 50–99% stenosis, and 13 (0.4%) had dissection. Conclusion: TIA patients with aphasia, SBP ‡ 160 mmHg, and a history of hypertension, coronary artery disease, dyslipidemia, or previous carotid stenosis are more likely to have carotid artery stenosis or dissection. Patients with confusion and vertigo were at lower risk. Despite the size and power of this study, many patient characteristics are only weakly associated with positive imaging. For now, urgent carotid imaging should remain an essential component of early risk stratification of all TIA patients. 82 Boarding Costs: ED Care Is More Expensive Richard Martin, James Miranda Temple University, Philadelphia, PA Background: The direct cost of medical care is not accurately reflected in charges or reimbursement. The cost of boarding admitted patients in the ED has been studied in terms of opportunity costs, which are indirect. The actual direct effect on hospital expenses has not been well defined. Objectives: We calculate the difference to the hospital in the cost of caring for an admitted patient in the ED and in a non-critical care in-patient unit. Methods: Time-directed activity-based costing (TDABC) has recently been proposed as a method of determining the actual cost of providing medical services. TDABC was used to calculate the cost per patient bed-hour both in the ED and for an in-patient unit. The costs include nursing, nursing assistants, clerks, attending and resident physicians, supervisory salaries, and equipment maintenance. Boarding hours were determined from placement of admission order to transfer to in-patient unit. A convenience sample of 100 consecutive non-critical care admissions was assessed to find the degree of ED physician involvement with boarded patients. Results: The overhead cost per patient bed-hour in the ED was $60.80. The equivalent cost per bed-hour inpatient was $23.39, a differential of $37.41. There were 27,618 boarding hours for medical-surgical patients in 2010, a differential of $1,033,189.38 for the year. For the short-stay unit (no residents), the cost per patient • www.aemj.org S47 hour was $11.36 and the boarding hours were 11,804. This resulted in a differential cost of $583,389.76, a total direct cost to the hospital of $1,616,579.14. Review of 100 consecutive admissions showed no orders placed by the ED physician after decision-toadmit. Conclusion: Concentration of resources in the ED means considerably higher cost per unit of care as compared to an in-patient unit. Keeping admitted patients boarding in the ED results in expensive underutilization. This is exclusive of significant opportunity costs of lost revenue from walk-out and diverted patients. This study includes the cost of teaching attendings and residents (ED and in-patient). In a non-teaching setting, the differential would be less and the cost of boarding would be shared by a fee-for-service ED physician group as well as the hospital. 83 Improving Identification of Frequent Emergency Department Users Using a Regional Health Information Exchange William Fleischman1, John Angiollilo2, Gilad Kuperman3, Arit Onyile1, Jason S. Shapiro1 1 Mount Sinai School of Medicine, New York, NY; 2 Columbia University College of Physicians and 3 Surgeons, New York, NY; New York Presbyterian Hospital, New York, NY Background: Frequent ED users consume a disproportionate amount of health care resources. Interventions are being designed to identify such patients and direct them to more appropriate treatment settings. Because some frequent users visit more than one ED, a health information exchange (HIE) may improve the ability to identify frequent ED users across sites of care. Objectives: To demonstrate the extent to which a HIE can identify the marginal increase in frequent ED users beyond that which can be detected with data from a single hospital. Methods: Data from 6/1/10 to 5/31/11 from the New York Clinical Information Exchange (NYCLIX), a HIE in New York City that includes ten hospitals, were analyzed to calculate the number of frequent ED users (‡4 visits in 30 days) at each site and across the HIE. S48 2012 SAEM ANNUAL MEETING ABSTRACTS 84 Results: There were 10,555 (1% of total patients) frequent ED users, with 7,518 (71%) of frequent users having all their visits at a single ED, while 3,037 (29%) frequent users were identified only after counting visits to multiple EDs (Table 1). Site-specific increases varied from 7% to 62% (SD 16.5). Frequent ED users accounted for 1% of patients, but for 6% of visits, averaging 9.74 visits per year, versus 1.55 visits per year for all other patients. 28.5% of frequent users visited two or more EDs during the study period, compared to 10.6% of all other patients. Conclusion: Frequent ED users commonly visited multiple NYCLIX EDs during the study period. The use of a HIE helped identify many additional frequent users, though the benefits were lower for hospitals not located in the relative vicinity of another NYCLIX hospital. Measures that take a community, rather than a single institution, into account may be more reflective of the care that the patient experiences. Indocyanine Green Dye Angiography Effective at Early Prediction of Second Degree Burn Outcome Mitchell S. Fourman, Brett T. Phillips, Laurie Crawford, Filippo Romainelli, Fubao Lin, Adam J. Singer, Richard A. Clark Stony Brook University Medical Center, Stony Brook, NY Background: Due to their complex nature and high associated morbidity, burn injuries must be handled quickly and efficiently. Partial thickness burns are currently treated based upon visual judgment of burn depth by the clinician. However, such judgment is only 67% accurate and not expeditious. Laser Doppler Imaging (LDI) is far more accurate - nearly 96% after 3 days. However, it is too cumbersome for routine clinical use. Laser Assisted Indocyanine Green Angiography (LAICGA) has been indicated as an alternative for diagnosing the depth of burn injuries, and possesses greater utility for clinical translation. As the preferred outcome of burn healing is aesthetic, it is of interest to determine if wound contracture can be predicted early in the course of a burn by LAICGA. Objectives: Determine the utility of early burn analysis using LAICGA in the prediction of 28-day wound contracture. Methods: A prospective animal experiment was performed using six anesthetized pigs, each with 20 standardized wounds. Differences in burn depth were created by using a 2.5 · 2.5 cm aluminum bar at three exposure times and temperatures: 70 degrees C for 30 seconds, 80 degrees C for 20 seconds, and 80 degrees C for 30 seconds. We have shown in prior validation experiments that these burn temperatures and times create distinct burn depths. LAICGA scanning, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 using Lifecell SPY Elite, took place at 1 hour, 24 hours, 48 hours, 72 hours, and 1 week post burn. Imaging was read by a blinded investigator, and perfusion trends were compared with day 28 post-burn contraction outcomes measured using ImageJ software. Biopsies were taken on day 28 to measure scar tissue depth. Results: Deep burns were characterized by a blue center indicating poor perfusion while more superficial burns were characterized by a yellow-red center indicating perfusion that was close to that of the normal uninjured adjacent skin (see figure). A linear relationship between contraction outcome and burn perfusion could be discerned as early as 1 hour post burn, peaking in strength at 24–48 hours post-burn. Burn intensity could be effectively identified at 24 hours post-burn, although there was no relationship with scar tissue depth. Conclusion: Pilot data indicate that LAICGA using Lifecell SPY has the ability to determine the depth of injury and predict the degree of contraction of deep dermal burns within 1–2 days of injury with greater accuracy than clinical scoring. 85 The Sepsis Alert: Real-time Electronic Health Record Surveillance in the Emergency Department and Sepsis Outcomes Thomas Yeich, Suzan Brown, Kristin Thomas, Robin Walker York Hospital, York, PA Background: Early intervention (e.g. fluid resuscitation, antibiotics, and goal directed therapy) improves outcomes in ED patients with severe sepsis and septic shock. Clinical definitions of severe sepsis and septic shock such as SIRS criteria, hypotension, elevated lactate, and end-organ damage (with the exception of mental status changes) are objectively defined and quantifiable. Objectives: We hypothesize that real-time monitoring of an integrated electronic medical records system and the subsequent firing of a ‘‘sepsis alert’’ icon on the electronic ED tracking board results in improved mortality for patients who present to the ED with severe sepsis or septic shock. Methods: We retrospectively reviewed our hospital’s sepsis registry and included all patients diagnosed with severe sepsis or septic shock presenting to an academic community ED with an annual census of 73,000 visits and who were admitted to a medical ICU or stepdown ICU bed between June 2009 and October 2011. In May 2010 an algorithm was added to our integrated medical records system that identifies patients with two SIRS criteria and evidence of endorgan damage or shock on lab data. When these criteria are met, a ‘‘sepsis alert’’ icon (prompt) appears next to that patient’s name on the ED tracking board. The system also pages an in-house, specially trained ICU nurse who can respond on a PRN basis and assist in the patient’s management. 18 months of intervention data are compared with 11 months of baseline data. Statistical analysis was via z-test for proportions. • www.aemj.org S49 Results: For ED patients with severe sepsis, the preand post-alert mortality was 19 of 125 (15%) and 34 of 378 (9%), respectively (p = 0.084; n = 503). In the septic shock group, the pre- and post-alert mortality was 27 of 92 (29%) and 48 of 172 (28%), respectively (p = 0.977). With ED and inpatient sepsis alerts combined, the severe sepsis subgroup mortality was reduced from 17% to 9% (p = 0.013; n = 622). Conclusion: Real-time ED EHR screening for severe sepsis and septic shock patients did not improve mortality. A positive trend in the severe sepsis subgroup was noted, and the combined inpatient plus ED data suggests statistical significance may be reached as more patients enter the registry. Limitations: retrospective study, potential increased data capture post intervention, and no ‘‘gold standard’’ to test the sepsis alert sensitivity and specificity. 86 Emergency Department Physician Experience With A Real-time, Electronic Pneumonia Decision Support Tool Caroline G. Vines1, Dave S. Collingridge2, Barbara E. Jones1, Leanne Struck2, Todd L. Allen2, Nathan C. Dean2 1 University of Utah, Salt Lake City, UT; 2 Intermountain Medical Center, Salt Lake City, UT Background: Variability in the decision to admit patients with pneumonia has been documented. We developed a real-time, electronic decision support tool to aid ED physicians in triage, diagnostic studies, and antibiotic selection for patients presenting with pneumonia. The tool identifies patients likely to have pneumonia and alerts the physician via an electronic tracker board (ETB). Utilization of the tool in the first 2 months was low. Objectives: To describe experience and impressions of physicians regarding the tool. Methods: An online survey was sent to all ED physicians and EM residents who work at one or more of four EDs in Salt Lake Valley 6 months after tool implementation. Surveys were confidential through REDCap (Research Electronic Data Capture). The survey utilized discrete questions with continuous, binary, or polychotomous response formats (see table). Descriptive statistics were calculated. Principal component analysis was used to determine questions with continuous response formats that could be aggregated. Aggregated outcomes were regressed onto predictor demographic variables using multiple linear regression. Results: 80/100 physicians completed the survey. Physicians had a mean of 9.8 ± 9.0 years experience in the ED. 23.8% were female. Eight physicians (10%) reported never having used the tool, while 70.8% of users estimated having used it more than five times. 75% of users cited the ‘‘P’’ alert on the ETB as the most common notification method. Most felt the ‘‘P’’ alert did not help them identify patients with pneumonia earlier (mean = 2.5 ± 1.2), but found it moderately useful in reminding them to use the tool (3.5 ± 1.3). Physicians found the tool helpful in making decisions regarding S50 2012 SAEM ANNUAL MEETING ABSTRACTS triage, diagnostic studies, and antibiotic selection for outpatients and inpatients (3.7 ± 1.0, 3.6 ± 1.1, 3.6 ± 1.1, and 4.2 ± 0.9, respectively). They did not feel it negatively affected their ability to perform other tasks (1.6 ± 0.9). Using multiple linear regression, neither age, sex, years experience, nor tool use frequency significantly predicted responses to questions about triage and antibiotic selection, technical difficulties, or diagnostic ordering. Conclusion: ED physicians perceived the tool to be helpful in managing patients with pneumonia without negatively affecting workflow. Perceptions appear consistent across demographic variables and experience. Table - Abstract 87: Table - Abstract 86: Sample of continuous response format survey questions* 9. Does the ‘‘P’’ alert on the PTS board help you to identify pneumonia patients earlier? 10. Does the ‘‘P’’ alert on the PTS board remind you to start and use the tool? 13. Do you feel the tool helped you to appropriately triage your patients? 14. Do you feel the tool helped you to order the appropriate diagnostic studies for your patients? 15. How useful was th etool in helping you to prescribe appropriate antibiotics for OUTPATIENTS? 16. How useful was th etool in helping you to prescribe appropriate antibiotics for INPATIENTS? 17. How frequently do you experience technical difficulties when using the tool? 18. Did the tool negatively impact your ability to perform other tasks? * response options were 1 through 5, with 1 indicating ‘‘not at all’’ and 5 ‘‘very much so’’ 87 Improving Identification of Hospital Readmissions Using a Regional Health Information Exchange John Angiollilo1, William Fleischman2, Gilad Kuperman3, Arit Onyile2, Jason S. Shapiro2 1 Columbia University College of Physicians and Surgeons, New York, NY; 2Mount Sinai School of Medicine, New York, NY; 3New York Presbyterian Hospital, New York, NY Background: Hospital readmissions within 30 days, many of which occur via the ED, are proposed as a target for improvement via ‘‘payment incentives’’ by the Center for Medicare and Medicaid Services. A portion of readmissions, however, occur away from the discharging hospital, making it difficult for hospitals to identify part of this population. Objectives: To demonstrate the extent to which a health information exchange system (HIE) can identify the marginal increase in 30-day readmissions and ED visits of recently discharged patients, which potentially lead to these readmissions. Methods: Data from 5/1/10 to 4/30/11 from the New York Clinical Information Exchange (NYCLIX), a HIE in New York City that includes ten hospitals with a total of 7,717 inpatient beds, were analyzed to calculate hospital index discharges and subsequent readmissions and ED visits to the discharging hospital versus other hospitals. Results: There were 320,967 inpatient admissions/discharges by 271,460 patients. There were 41,630 (13% of total discharges) readmissions within 30 days of discharge, with 37,829 readmissions occurring at the same hospital, while 3,801 patients (9.1% of total readmissions and 1.2% of total discharges) were readmitted to a different hospital. Site-specific increases in identification of readmissions ranged from 3%–25%, SD 8.7 (see table). There were 37,697 ED visits within 30 days of discharge, with 34,143 (91%) visits to the discharging hospital’s ED and 3,554 (9%) visits to a different hospital’s ED. Conclusion: Readmissions and ED visits within thirty days of discharge to non-discharging hospitals were common among this group of hospitals and patients during the study period. The use of a HIE helped identify many additional readmissions, though the benefits were lower for hospitals not located in the relative vicinity of another NYCLIX hospital. Measures that take a community, rather than a single institution, into account may be more reflective of the care that the patient experiences. 88 A Comparison Of Outcomes In Post-cardiac Arrest Patients With And Without Significant Intracranial Pathology On Head CT Sean Doran1, Sameer Syed1, Chris Martin2, Shelley McLeod1, Matthew Strauss1, Neil Parry1, Bryan Young1 1 The University of Western Ontario, London, ON, Canada; 2Royal Victoria Hospital, Barrie, ON, Canada Background: When survivors of sudden cardiac arrest arrive to the emergency department (ED) with return of spontaneous circulation (ROSC), physicians should ascertain the etiology of the cardiac arrest (cardiac, neurologic, respiratory, etc). Consequently, some clinicians advocate for post-cardiac arrest head CT in all patients with ROSC for diagnostic, treatment, and prognostic purposes. Objectives: To explore clinical practice regarding the use of head CT in the immediate post-cardiac arrest period in patients with ROSC. A secondary objective was to compare outcomes of patients with significant ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 (SAH, edema, or infarct) versus non-significant findings on head CT. Methods: A retrospective medical record review was conducted for all adult (‡ 18 years) post-cardiac arrest (initial rhythm VT/VF, PEA, or asystole) patients admitted to the intensive care unit (ICU) of an academic tertiary care centre (annual ED census 150,000) from 2006–2007. Data were extracted using a standardized data collection tool by trained research personnel. Results: 200 patients were enrolled. Mean (SD) age was 66 (16) years and 56.5% were male. 79 (39.5%) had a head CT within 24 hours of ICU admission. 14 (17.7%) had significant findings on head CT. Of these, 1 (7.1%) patient survived to ICU discharge, compared to 11/65 (16.9%) patients with non-significant findings (D )9.8; 95% CI )22.2%, 15.6%). Of those with significant findings on head CT, median (IQR) ICU length of stay was 1 (1, 2.5) day compared to 4 (2, 4) days for patients with non-significant findings. Survival to hospital discharge was not different for patients with significant findings on head CT (1; 7.1%) compared to those with non-significant (9; 13.8%) findings (D )6.7; 95% CI )18.7%, 18.5%). No patients with significant head CT findings survived to 1 year, compared to 9 (13.8%) patients with non-significant findings (D )13.8; 95% CI )24.3%, 8.6%). Conclusion: The use of post-cardiac arrest head CT was variable with less than half of patients undergoing head CT within 24 hours. There were no differences in outcomes between those with significant versus nonsignificant findings on head CT in this small pilot study. Further research is required to more accurately determine the prognostic utility of this imaging modality and determine if there is a difference in other outcomes such as ICU or hospital length of stay. 89 Bandemia Does Not Predict Mortality, Positive Cultures, Or Source Location In Septic Patients Presenting To The ED Scott Teanu Mataoa, David Conner, Charles R. Wira Yale New Haven Hospital, New Haven, CT Background: Bandemia is believed to be associated clinically as a predictor of mortality in septic patients and has been incorporated into severity of illness scoring systems. To date there are very little data looking specifically at whether bandemia predicts mortality in this patient population. Objectives: Early identification and risk stratification of septic patients in the emergency department (ED) is important when implementing early clinical interventions to prevent mortality. In this study we evaluated whether bandemia >10% is associated with mortality in patients with severe sepsis or septic shock presenting to the ED. Secondary analyses included evaluating bandemia as a predictor of culture positivity, gram-negative source of infection, or source location. Methods: Retrospective cross-sectional study utilizing patients identified in the Yale-New Haven Hospital sepsis registry. • www.aemj.org S51 Results: 521 patients from the sepsis registry were included in the study with 19.2% (n = 100) meeting criteria of bandemia. The in-hospital 28-day mortality rate in patients with bandemia was 18% (18/100), compared to 14.5% (61/421) in patients without bandemia (p = 0.43). The rate of culture positivity defined as any positive culture (blood, urine, sputum, wound, catheter, abscess) was 50% (50/100) in patients with bandemia and 40% (170/421) in patients without bandemia (p = 0.09). With respect to blood cultures specifically the rate of positive cultures was 20% (20/ 100) in patients with bandemia and 18% (76/421) in patients without bandemia (p = 0.56). Additionally the rate of gram-negative organisms from positive blood cultures was 25% (25/100) in patients with bandemia and 24.7% (105/421) in patients without bandemia (p = 1.00). Regarding source location the respective rates of incidence among patients with bandemia were pneumonia 19.6% (31/158), GU 19.5% (16/82), abdominal 25.4% (14/55), soft tissue 23.5% (8/ 34) with no significant difference between the groups (p = NS). Conclusion: In septic patients presenting to the ED with bandemia there was no observed difference in rates of mortality, positive cultures, or source location. 90 Impact of ED Volumes on Sepsis Resuscitation Bundle Compliance at an Urban Level I Trauma Center Hima Rao, Manu Malhotra, Howard A. Klausner, Joseph Tsao, Emanuel Rivers Henry Ford Hospital, Detroit, MI Background: The Early Goal Directed Therapy (EGDT) Resuscitation Bundle (RB) has been proven to reduce mortality in patients with severe sepsis or septic shock. Universal implementation, however, has proved elusive, resulting in what we consider a preventable loss of life. Objectives: The purpose of this study is to determine if there is an association between ED patient volumes and RB compliance. Methods: Our study was a retrospective chart review performed at an urban, Level I trauma center teaching hospital with an annual volume of >90,000 patients. All patients ‡ 18 years old who presented to the ED between July 1, 2010 and December 31, 2010 with diagnoses of severe sepsis or septic shock were included. ED volume data for patients who received complete compliance with the RB were compared with those who did not. The electronic medical record and sepsis registry were used to obtain data regarding bundle compliance, daily census, new arrivals during stay, and nursing and physician to patient ratios. Wilcox rank sum tests were used to compare differences between the RB compliance and control groups. Results: During the review period 224 eligible patients presented (112 compliance group, 112 control group). Average daily ED census was comparable (245.2 compliance, 249.86 control, p = 0.206), as was Category 1 (ED high acuity area) census (39.94 compliance, 40.86 control, p = 0.302). New Category 1 patients (10.1 S52 2012 SAEM ANNUAL MEETING ABSTRACTS compliance, 10.39 control, p = 0.737) and new resuscitation patients during ED stay (1.73 compliance, 2.12 control, p = 0.117) did not show significant differences. Finally, nursing to patient (2.2 compliance, 2.19 control, p = 0.843) and physician to patient ratios (4.33 compliance, 4.27 control, p = 0.076) were also similar between groups. Conclusion: This study did not show a statistically significant difference between ED volume data for patients receiving complete resuscitation bundle compliance and those who did not. Limitations include the retrospective design of the study and that our sample size may be too small to detect a threshold of overcrowding over which bundle compliance is adversely affected. Given the effect on mortality of this intervention, further study is needed to identify barriers to RB compliance. 91 An Experimental Comparison of Endotracheal Intubation Using a Blind Supraglottic Airway Device During Ongoing CPR With Manual Compression versus Automated Compression Bob Cambridge, Amy Chelin, Austin Lamb, John Hafner OSF St. Francis Medical Center, Peoria, IL Background: There are a variety of devices on the market, designed to quickly provide an airway during resuscitation efforts. A new device, the Supraglottic Airway Laryngopharyngeal Tube (SALT; Life Assist) is designed for blind tracheal placement of an airway. Use of the device in previous studies has been in static models. Objectives: We seek to examine whether use of the SALT device can provide reliable tracheal intubation during ongoing CPR. The dynamic model tested the device with human powered CPR (manual) and with an automated chest compression device (Physio Control Lucas 2). The hypothesis is that the predictable movement of an automated chest compression device will make tracheal intubation easier than the random movement from manual CPR. Methods: The project was an experimental controlled trial and took place in the ED at a tertiary referral center in Peoria, Illinois. This project was an expansion arm of a similarly structured study using traditional laryngoscopy. Emergency medicine residents, attending physicians, paramedics, and other ACLS-trained staff were eligible for participation. In randomized order, each participant attempted intubation on a mannequin using the SALT device with no CPR ongoing, during CPR with a manual compression, and during CPR with an automatic chest compression. Participants were timed in their attempt and success was determined after each attempt. Results: There were 43 participants in the trial. The success rates in the control group and the automated CPR group were both 86% (37/43) and the success rate in the manual CPR group was 79% (34/43). The difference in success rates were not statistically significant (p = 0.99 and p = 0.41). The automated CPR group had the fastest average time but the difference was not significant (8.051 sec; p = 0.144). The mean time for intubation with manual CPR and no CPR were not statistically different (9.042 sec, 8.770 sec; p = 0.738). Conclusion: Using the SALT device, the success rate of tracheal intubation with ongoing chest compression was similar to the success rate of intubation without CPR. The SALT device did not guarantee a secure airway every time in a dynamic scenario and it had the drawback of not providing direct visualization during placement. The SALT device is a helpful adjunct but should not replace a direct visualization method for a first line intubation attempt in the ED. 92 Comparison of Baseline Aortic Velocity Profiles and Response to Weight-Based Volume Loading in Fasting Subjects: A Randomized, Prospective Double-Blinded Trial Anthony J. Weekes, Margaret R. Lewis, Zachary Kahler, Donald Stader, Dale P. Quirke, Courtney Almond, H. James Norton, Dawn Middleton, Vivek S. Tayal Carolinas Medical Center, Charlotte, NC Background: Cardiac output increases have been used to define fluid responsiveness in mechanically ventilated septic patients. The aortic velocity profile is an important variable in cardiac output calculations. We evaluated noninvasive measurements of aortic velocity using bedside cardiac ultrasound. Objectives: Our primary hypothesis was that in fasting, asymptomatic subjects, larger fluid boluses would lead to proportional aortic velocity changes. Our secondary endpoints were to determine inter- and intra-subject variation in aortic velocity measurements. Methods: The authors performed a prospective randomized double-blinded trial using healthy volunteers. We measured the velocity time integral (VTI) and maximal velocity (Vmax) with an estimated 0–20 pulsed wave Doppler interrogation of the left ventricular outflow in the apical-5 cardiac window. Three physicians reviewed optimal sampling gate position, Doppler angle and verified the presence of an aortic closure spike. Angle correction technology was not used. Subjects with no history of cardiac disease or hypertension fasted for 12 hours and were then randomly assigned to receive a normal saline bolus of 2 ml/kg, 10 ml/kg or 30 ml/kg over 30 minutes. Aortic velocity profiles were measured before and after each fluid bolus. Results: Forty-two subjects were enrolled. Mean age was 33 ± 10 (range 24 to 61) and mean body mass index 24.7 ± 3.2 (range 18.7 to 32). Mean volume (in ml) for groups receiving 2 ml/kg, 10 ml/kg, and 30 ml/kg were 151, 748, and 2162, respectively. Mean baseline Vmax (in cm/s) of the 42 subjects was 108.4 ± 12.5 (range 87 to 133). Mean baseline VTI (in cm) was 23.2 ± 2.8 (range 18.2 to 30.0). Pre- and post-fluid mean differences for Vmax were )1.7 (± 10.3) and for VTI 0.7 (± 2.7). Aortic velocity changes in groups receiving 2 ml/kg, 10 ml/kg, and 30 ml/kg were not statistically significant (see table). Heart rate changes were not significant. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Conclusion: Aortic velocity changes associated with fluid loading were neither proportional nor statistically significant within the different groups. Pulsed wave Doppler values showed temporal variations that were likely unrelated to volume loading. Angle of Doppler interrogation was difficult to standardize and may have influenced variability. There are technical limitations to serial cardiac output calculations and comparisons with echocardiography. 93 Identification of Critical Areas for Improvement in ED Severe Sepsis Resuscitation Utilizing In Situ Simulation Emilie Powell, David H. Salzman, Susan Eller, Lanty O’Connor, John Vozenilek Northwestern University, Chicago, IL Background: Clinicians recognize that septic shock is a highly prevalent, high mortality disease state. Evidence supports early ED resuscitation, yet care delivery is often inconsistent and incomplete. The objective of this study was to discover latent critical barriers to successful ED resuscitation of septic shock. Objectives: Clinicians recognize that septic shock is a highly prevalent, high mortality disease state. Evidence supports early ED resuscitation, yet care delivery is often inconsistent and incomplete. The objective of this study was to discover latent critical barriers to successful ED resuscitation of septic shock. Methods: We conducted five 90-minute risk-informed in-situ simulations. ED physicians and nurses working in the real clinical environment cared for a standardized patient, introduced into their existing patient workload, with signs and symptoms of septic shock. Immediately after case completion clinicians participated in a 30minute debriefing session. Transcripts of these sessions were analyzed using grounded theory, a method of qualitative analysis, to identify critical barrier themes. Results: Fifteen clinicians participated in the debriefing sessions: four attending physicians, five residents, five nurses, and one nurse practitioner. The most prevalent critical barrier themes were: anchoring bias and difficulty with cognitive framework adaptation as the patient progressed to septic shock (n = 26), difficult interactions between the ED and ancillary departments (n = 22), difficulties with physician-nurse commu- • www.aemj.org S53 nication and teamwork (n = 18), and delays in placing the central venous catheter due to perceptions surrounding equipment availability and the desire to attend to other competing interests in the ED prior to initiation of the procedure (n = 17 and 14). Each theme was represented in at least four of the five debriefing sessions. Participants reported the in-situ simulations to be a realistic representation of ED sepsis care. Conclusion: In-situ simulation and subsequent debriefing provides a method of identifying latent critical areas for improvement in a care process. Improvement strategies for ED-based septic shock resuscitation will need to address the difficulties in shock recognition and cognitive framework adaptation, physician and nurse teamwork, and prioritization of team effort. 94 The Presenting Signs and Symptoms of Ruptured Abdominal Aortic Aneurysms: A Meta-analysis of the Literature Utpal Inamdar, Charles R. Wira Yale New Haven Hospital Department Emergency Medicine, New Haven, CT of Background: Ruptured abdominal aortic aneurysms (AAA) are the 15th leading cause of death in the United States causing more than 10,000 deaths annually. Misdiagnosis rates can range as high as 30–60%. The classic clinical presentation involves abdominal pain, back pain, syncope, hypotension, and a pulsatile abdominal mass. However, the frequency of these signs and symptoms has never been evaluated in a meta-analysis. Objectives: To identify the frequency of presenting signs and symptoms of patients with ruptured AAAs. Methods: A review of the literature from 1980 to present was performed using the MeSH heading of ‘‘Ruptured Abdominal Aortic Aneurysm.’’ Results: 1547 articles were identified. After content analysis and the application of inclusion and exclusion criteria, 28 articles were identified regarding the presenting symptoms of abdominal pain, back pain, or syncope (n = 3600), 30 articles were identified regarding hypotension (n = 3419), and 13 articles were identified regarding the presence of a pulsatile mass (n = 1926). 58% (1669 of 2857, 95% CI 0.46 to 0.71) of patients presented with abdominal pain, 47% had back pain (1090 of 2299, 95% CI 0.38 to 0.63), and 26% had syncope (476 of 1797, 95% CI 0.20 to 0.37). 52% were initially hypotensive with a systolic blood pressure less than 80– 100 mmHg (1771 of 3419, 95% CI 0.43 to 0.55), and 66% had a palpable abdominal pulsatile mass (1262 of 1926, 95% CI 0.46 to 0.79). Conclusion: The diagnosis of ruptured AAAs can be elusive and requires a high index of suspicion. The classic presenting signs and symptoms may not always be present. If a ruptured AAA is suspected, further diagnostic imaging is mandatory for confirmation of diagnosis. S54 95 2012 SAEM ANNUAL MEETING ABSTRACTS Normal Initial Blood Sugar Level and History of Diabetes Might Reduce In-hospital Mortality of Septic Patients Visiting the Emergency Department Hsiao-Yun Chao1, Sheng-Che Lin1, Chun-Kuei Chen1, Peng-Hui Liu1, Jih-Chang Chen1, Yi-Lin Chan1, Kuan-Fu Chen2 1 Chang-Gung Memorial Hospital, Taoyuan county, Taiwan; 2Chang-Gung Memorial Hospital, Keelung City, Taiwan Background: The association between blood glucose level and mortality in critically ill patients is highly debated. Several studies have investigated the association between history of diabetes, blood sugar level, and mortality of septic patients; however, no consistent conclusion could be drawn so far. Objectives: To investigate the association between diabetes and initial glucose level and in-hospital mortality in patients with suspected sepsis from the ED. Methods: We conducted a retrospective cohort study that consisted of all adult septic patients who visited the ED at a tertiary hospital during the year 2010 with two sets of blood cultures ordered by physicians. Basic demographics, ED vital signs, symptoms and signs, underlying illnesses, laboratory findings, microbiological results, and discharge status were collected. Logistic regressions were used to evaluate the association between risk factors, initial blood sugar level, and history of diabetes and mortality, as well as the effect modification between initial blood sugar level and history of diabetes. Results: A total of 4997 patients with available blood sugar levels were included, of whom 48% had diabetes, 46% were older than 65 years of age, and 56% were male. The mortality was 6% (95% CI 5.3–6.7%). Patients with a history of diabetes tended to be older, female, and more likely to have chronic kidney disease, lower sepsis severity (MEDS score), and positive blood culture test results (all p < 0.05). Patients with a history of diabetes tended to have lower in-hospital mortality after ED visits with sepsis, controlling for initial blood sugar level (aOR 0.72, 95% CI 0.56–0.92, p = 0.01). Initial normal blood sugar seemed to be beneficial compared to lower blood sugar level for in-hospital mortality, controlled history of diabetes, sex, severity of sepsis, and age (aOR 0.61, 95% CI 0.44–0.84, p = 0.002). The effect modification of diabetes on blood sugar level and mortality, however, was found to be not statistically significant (p = 0.09). Conclusion: Normal initial blood sugar level in ED and history of diabetes might be protective for mortality of septic patients who visited the ED. Further investigation is warranted to determine the mechanism for these effects. 96 Sedation and Paralytic Use During Hypothermia After Cardiac Arrest William A. Knight1, Shaun Keegan2, Opeolu Adeoye1, Jordan Bonomo1, Kimberly Hart1, Lori Shutter1, Christopher Lindsell1 1 University of Cincinnati, Cincinnati, OH; 2UC Health - University Hospital, Cincinnati, OH Background: Therapeutic hypothermia improves outcomes in comatose survivors of cardiac arrest, yet fewer than half of successfully resuscitated patients survive to hospital discharge. Deep sedation and/or paralytics may be used to avoid shivering and maintain target temperature, but how often this occurs is unknown. Objectives: We aim to determine the frequency of deep sedation and paralytic usage in patients undergoing hypothermia. Methods: This IRB-approved retrospective chart review included all patients treated with therapeutic hypothermia after cardiac arrest during 2010 at an urban, academic teaching hospital. Every patient undergoing therapeutic hypothermia is treated by neurocritical care specialists. Patients were identified by review of neurocritical care consultation logs. Clinical data were dually abstracted by trained clinical study assistants using a standardized data dictionary and case report form. Medications reviewed during hypothermia were midazolam, lorazepam, propofol, fentanyl, cisatracurium, and vecuronium. Results: There were 33 patients in the cohort. Median age was 57 (range 28–86 years), 67% were white, 55% were male, and 49% had a history of coronary artery disease. Seizures were documented by continuous EEG in 11/33 (33%), and 20/33 (61%) died during hospitalization. Most, 30/33 (91%), received fentanyl, 21/33 (64%) received benzodiazepine pharmacotherapy, and 23/33 (70%) received propofol. Paralytics were administered to 23/33 (68%) patients, 14/33 (42%) with cisatracurium and 9/33 (27%) with vecuronium. Of note, one patient required pentobarbital for seizure management. Conclusion: Sedation and neuromuscular blockade are common during management of patients undergoing therapeutic hypothermia after cardiac arrest. Patients in this cohort often received analgesia with fentanyl, and sedation with a benzodiazepine or propofol. Given the frequent use of sedatives and paralytics in survivors of cardiac arrest undergoing hypothermia, future studies should investigate the potential effect of these drugs on prognostication and survival after cardiac arrest. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 97 The Implementation of Therapeutic Hypothermia in the Emergency Department: A Multi-Institution Case Review Sara W. Johnson1, Daniel Joseph1, Dina Seif1, Melissa Joseph1, Meena Zareh1, Christine Kulstad2, Mike Nelson3, David Barnes4, Christine Riguzzi5, Adrian Elliot6, Eric Kochert7, David Slattery8, Sean O. Henderson1 1 Keck School of Medicine of the University of Southern California, Los Angeles, CA; 2Advocate Christ Medical Center, Oak Lawn, IL; 3Cook County Hospital, Chicago, IL; 4University of California Davis, Davis, CA; 5Highland General Hospital, Oakland, CA; 6Shands Jacksonville Medical Center, Jacksonville, FL; 7York Hospital, York, ME; 8University of Nevada School of Medicine, Las Vegas, NV Background: The use of therapeutic hypothermia (TH) is a burgeoning treatment modality for post-cardiac arrest patients. Objectives: We performed a retrospective chart review of patients who underwent post cardiac arrest TH at eight different institutions across the United States. Our objective was to assess how TH is currently being implemented in emergency departments and assess the feasibility of conducting more extensive TH research using multi-institution retrospective data. Methods: A total of 94 charts with dates from 2008– 2011 were sent for review by participating institutions of the Peri-Resuscitation Consortium. Of those reviewed, eight charts were excluded for missing data. Two independent reviewers performed the review and the results were subsequently compared and discrepancies resolved by a third reviewer. We assessed patient demographics, initial presenting rhythm, time until TH initiation, duration of TH, cooling methods and temperature reached, survival to hospital discharge, and neurological status on discharge. Results: The majority of cases of TH had initial cardiac rhythms of asystole or pulseless electrical activity (55.2%), followed by ventricular tachycardia or fibrillation (34.5%), and in 10.3% the inciting cardiac rhythm was unknown. Time to initiation of TH ranged from 0–783 minutes with a mean time of 99 min (SD 132.5). Length of TH ranged from 25–2171 minutes with a mean time of 1191 minutes (SD 536). Average minimum temperature achieved was 32.5C, with a range from 27.6–36.7 C (SD 1.5 C). Of the charts reviewed, 29 (33.3%) of the patients survived to hospital discharge and 19 (21.8%) were discharged relatively neurologically intact. Conclusion: Research surrounding cardiac arrest has always been difficult given the time and location span from pre-hospital care to emergency department to intensive care unit. Also, as witnessed cardiac arrest events are relatively rare with poor survival outcomes, very large sample sizes are needed to make any meaningful conclusions about TH. Our varied and inconsistent results show that a multi-center retrospective review is also unlikely to provide useful information. A prospective multi-center trial with a uniform TH protocol is needed if we are ever to make any evidence- • www.aemj.org S55 based conclusions on the utility of TH for post-cardiac arrest patients. 98 Serum Lactate as a Screening Tool and Predictor of Outcome in Pediatric Patients Presenting to the Emergency Department with Suspected Infection Myto Duong, Loren Reed, Jennifer Carroll, Antonio Cummings, Stephen Markwell, Jarrod Wall Southern Illinois University, Springfield, IL Background: No single reliable sepsis biomarker has been identified for risk stratification and prognostication in pediatric patients presenting to the ED. A biomarker allowing for early diagnosis would facilitate aggressive management and improve outcomes. Serum lactate (LA) is an inexpensive and readily available measure of tissue hypoxia that is predictive of mortality in adults with sepsis. Objectives: To determine if ED LA correlates with sepsis, admission, and outcome in pediatric patients with suspected infection. Hypothesis: LA is a useful predictor of pediatric sepsis, admission, and outcome in the ED. Methods: This retrospective study was performed in a community hospital ED in which a sepsis protocol involving concurrent LA draw with every blood culture obtained in the ED was initiated in 2009–2010. A total of 289 pediatric (<18 years old) patients with suspected infection had LA obtained in the ED. Pearson correlation coefficients were used to determine the relationship between LA and variables such as vital signs, basic labs, admission, length of stay (LOS), and 3-day return rate. Patients were dichotomized into those with LA >3 or £ 3. T-tests were used to compare the groups. Results: Mean LA was 2.04, SD = 1.45. Mean age was 4.5 years old, SD = 5.20. A statistically significant positive correlation was found between LA and pulse, respiratory rate (RR), WBC, platelets, and LOS, while a significant negative correlation was seen with temperature and HCO3-. When two subjects were dropped as possible outliers with LA >10, it resulted in non-significant temperature correlation, but a significant negative correlation with age and BUN was revealed. Patients in the higher LA group were more likely to be admitted (p = 0.0001) and have longer LOS. Of the discharged patients, there was no difference in mean LA level between those who returned (n = 25, mean LA of 1.88, SD = 0.88) and those who did not (n = 154, mean LA of 1.88, SD = 1.35), p = 0.99. Furthermore, mean LA levels for those with sepsis (n = 138, mean LA of 2.18, SD = 1.75) did not differ from those without sepsis (n = 147, mean LA of 1.9, SD = 1.08), p = 0.11. Conclusion: Higher LA in pediatric patients presenting to the ED with suspected infection correlated with increased pulse, RR, WBC, platelets, and decreased BUN, HCO3-, and age. LA may be predictive of hospitalization, but not of 3-day return rates or pediatric sepsis screening in the ED. S56 2012 SAEM ANNUAL MEETING ABSTRACTS Table 1 - Abstract 98: Lactate correlation to various variables Variables Correlation factor (r) p-value Sample size -0.09 -0.13 0.18 0.25 0.17 0.22 )0.18 )0.11 0.33 0.109 0.022 0.003 0.0001 0.0045 0.0004 0.004 0.78 0.0001 288 287 283 272 286 269 245 247 288 Age Temperature Pulse Respiratory rate WBC Platelet HCO3BUN LOS Table 2 - Abstract 98: Dichotomized lactate levels and correlation to various variables Variables Age (years) Temperature Pulse WBC Platelets (K) HCO3BUN LOS Respiratory Rate mean LA</=3 (n = 249) 4.7 38 135.6 11.4 290.6 23.8 10.6 0.83 29 SD mean LA>3 (n = 40) SD p-value 5.1 1.4 31.3 5.8 119.1 3.8 6.1 1.72 9.7 3.3 37.5 149.5 14.2 376.6 21.6 9.9 2.5 35.9 5.2 1.9 35 7.6 165.2 3.2 4.0 4.4 16.6 0.10 0.13 0.01 0.03 0.006 0.001 0.34 0.02 0.009 Table 3 - Abstract 98: Dichotomized lactate levels and correlation to sepsis, admission, and Return to ED Visits Categories Sepsis Admission Return to ED visit 99 if LA</=3 (n = 249) Percent if LA>3 (n = 40) Percent p-value 119 80 28 48% 33% 11% 20 28 4 51% 75% 10% 0.71 0.0001 1.0 Failure To Document The Presence Of Sepsis Decreases Adherence To Process Of Care Measures In Emergency Department Patients Stephanie Dreher, James O’Brien, Jeffrey Caterino The Ohio State University, Columbus, OH Background: As early identification and treatment of sepsis improves outcomes, physician recognition of sepsis early in its course is essential. For emergency department patients, one suggested measure of early recognition is documentation of the presence of sepsis. Objectives: To determine (1) the frequency with which admitted ED patients who receive antibiotics meet criteria for sepsis; (2) the frequency with which early documentation of sepsis occurs in ED patients; and (3) the association between documentation of sepsis and provided care. Methods: We conducted a retrospective cohort study of ED patients who received antibiotics within 24 hours of admission. We classified patients as ‘‘septic’’ based on current sepsis criteria. ‘‘Documentation of sepsis’’ was considered present if the word ‘‘sepsis’’ or ‘‘septic’’ was documented in the ED notes or admission history and physical. We determined the effect documentation of sepsis had on process outcomes: ordering of blood cultures and lactate in the ED, antibiotic ordering in <2 hours, antibiotic administration in <2 hours, and fluid administration in <2 hours. We derived a non-parsimonius propensity model to adjust for recognition of sepsis using patient vital signs, age, sex, and WBC count. We then constructed a propensity-adjusted multivariable logistic regression analysis for each outcome. The model included the propensity score, triage level, and recognition of sepsis. Results: Out of 500 admitted patients receiving antibiotics, 27.4% met criteria for sepsis and were included in the study. Of these, 25.5% had sepsis documented. In univariate analysis, documentation of sepsis was associated with ordering of blood cultures (97% vs. 71%, p = 0.001), lactate (66% vs. 29%, p = 0.001), and antibiotics within 2 hours (77% vs. 49%, p = 0.011). In the multivariate models, recognition of sepsis was associated with ordering of blood cultures (OR 8.52, 95% CI 1.08– 67.49)(p = 0.042) and ordering lactate (OR 2.60, 95% CI 1.09–6.22)(p = 0.031). There was a non-significant trend towards ordering antibiotics within 2 hours (OR 2.22, 95% CI 0.89–5.56)(p = 0.088). There was no association with 2-hour antibiotic or fluid administration. Conclusion: ED and admitting physicians failed to document the presence of sepsis 74.5% of the time. Failure to document sepsis is associated with significantly lower rates of several recommended processes of care. 100 Sonogram Measured Inferior Vena Cava Diameter Response to Intravenous Fluid Bolus Christopher T. Vogt1, Brandon Smallwood2, Michael Sanders2, Cathy Rehmeyer1 1 University of Pikeville Kentucky College of Osteopathic Medicine, Pikeville, KY; 2Pikeville Medical Center, Pikeville, KY Background: Prior research data have shown sonogram measured inferior vena cava (IVC) diameter to be a reliable non-invasive measure of a patient’s fluid status compared to invasive techniques. Objectives: The goal of this study was to determine the normal maximum/exhaled IVC diameter (MIVCD), the possible relationship to a patient’s age, sex, and BMI, and the change in the MIVCD after a one liter intravenous fluid bolus (IVF) of normal saline. Methods: This prospective observational study was performed in a rural ED and on a college campus. Two cohorts were evaluated: a convenience sample of any consenting adult (control group), and any patients receiving an IVF bolus in the ED (IVF group). Portable bedside sonogram was used to measure IVC diameter one centimeter distal to the junction of the hepatic veins with a curved abdominal probe during exhalation using a longitudinal view. All subjects in the IVF group received a scan before and after receiving a ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 one liter fluid bolus of normal saline as prescribed by the emergency medicine provider present. Subject’s sex, race, age, weight, height, time of last meal, IV gauge, bolus time, as well as pre-IVF and post-IVF heart rate and blood pressure were collected. Subjects were stratified into subgroups within heart rate, blood pressure, and BMI classifications for data analysis. The t-test for independent means was the statistical test used for statistical significance. The level of risk was set at 0.05 and the power of this pilot study was approximately 20%. Results: A total of 213 adults (94 m, 119 f) consented to the study; 153 (64 m, 89 f) in the control group, 51 (22 m, 29 f) in the IVF group. Subject’s ages ranged from 18 to 78 years of age with a mean age of 34. There were 208 Caucasians, 3 African Americans, and 2 Asian subjects. There was no statically significant relationship between MIVCD and the subject’s sex, age, race, or BMI. The mean MIVCD for our control group was found to be 16.3 mm (11 mm–21.6 mm 95% CI). The mean change in MIVCD after the one liter IVF bolus was 3 mm ()1.3 mm–7.3 mm 95% CI). Conclusion: These data support previous researchers’ findings that demonstrated the standard MIVCD in an adult is not correlated to the subject’s sex, age, race, or BMI. These data should be used as a guide for clinicians treating patients requiring intravenous fluid resuscitation where portable bedside ultrasound is available. 101 Ultrasound-Guided Vascular Access On A Phantom: A Training Model For Medical Student Education Lydia M. Sahlani, David P. Bahner, Eric J. Adkins, Diane Gorgas, Clint Allred The Ohio State University Medical Center, Columbus, OH Background: Ensuring patient safety and prevention of medical errors has become an integral part of medical education. Vascular access is an area in which medical errors can result in serious complications. Medical student training has been increasing early exposure to the clinical aspects of medicine in hopes of reducing medical errors. Objectives: We reviewed a cohort of second year medical students (MS2s) to assess their proficiency with ultrasound-guided vascular access. Methods: This study was an observational cohort study of MS2s during their Introduction to Clinical Medicine (ICM) program. Students were required to review an online training module from EMSONO.com about ultrasound-guided vascular access. Students completed a quiz testing material presented in the online module. Students participated in a didactic session on ultrasound-guided vascular access using a blue phantom block gel model. Students were divided into groups of 4–5 participants and allowed to practice the skills on the vascular access model while being proctored by an experienced provider. They had no time limitations during their practice session. After the practice session the students were graded by the same proctor using a standardized scoring sheet. The students were evaluated on • www.aemj.org S57 their ability to visualize the simulated vessel in different planes, perform vascular cannulation in both the short and long axis, the number of needle sticks attempted, and successful cannulation. Results: A total of 134 MS2s were evaluated. Twentyseven students were excluded due to incomplete data. Of the 107 students with complete data, 100% (107/107) could visualize the vessel in long axis, short axis, and visualize the needle in the near field. 103/107 (96.26%) could visualize the needle entering in the long axis while 101/107 (94.39%) could visualize the needle entering the short axis. Students were able to cannulate the vessel in two sticks or less for 99/107 (92.52%) in the long axis and 100/107 (93.46%) in the short axis. Conclusion: A structured ultrasound curriculum can help MS2s learn the psychomotor skills necessary to cannulate a vessel on a phantom using ultrasound guidance. Future studies could be developed to assess whether this skill translates into clinical practice during clerkship experiences. 102 The Tongue Blade Test: Still Useful As A Screening Tool For Mandibular Fractures? Nicholas Caputo1, Andaleeb Raja1, Christopher P. Shields1, Nathan Menke2 1 Lincoln Medical and Mental Health Center, 2 Bronx, NY; University of Pittsburgh, Pittsburgh, PA Background: Mandibular fractures are one of the most frequently seen injuries in the trauma setting. In terms of facial trauma, madibular fractures account for 40– 62% of all facial bone fractures. Prior studies have demonstrated that the use of a tongue blade to screen these patients to determine whether a mandibular fracture is present may be as sensitive as x-ray. One study showed the sensitivity and specificity of the test to be 95.7% and 63.5%, respectively. In the last ten years, high-resolution computed tomography (HCT) has replaced panoramic tomography (PT) as the gold standard for imaging of patients with suspected mandibular fractures. This study determines if the tongue blade test (TBT) remains as sensitive a screening tool when compared to the new gold standard of CT. Objectives: The purpose of the study was to determine the sensitivity and specificity of the TBT as compared to the new gold standard of radiologic imaging, HCT. The question being asked: is the TBT still useful as a screening tool for patients with suspected mandibular fractures when compared to the new gold standard of HCT? Methods: Design: Prospective cohort study. Setting: An urban tertiary care Level I trauma center. Subjects: This study took place from 8/1/10 to 8/31/11 in which any person suffering from facial trauma presented. Intervention: A TBT was performed by the resident physician and confirmed by the supervising attending physician. CT facial bones were then obtained for the ultimate diagnosis. Inter-rater reliability (kappa) was calculated, along with sensitivity, specificity, accuracy, S58 2012 SAEM ANNUAL MEETING ABSTRACTS PPV, NPV, likelihood ratio (LR) (+), and likelihood ratio (LR) (-) based on a 2 · 2 contingency tables generated. Results: Over the study period 85 patients were enrolled. Inter-rater reliability was kappa = 0.93 (SE +0.11). The table demonstrates the outcomes of both the TBT and CT facial bones for mandibular fracture. The following parameters were then calculated based on the contingency table: sensitivity 0.97 (CI 0.81–0.99), specificity 0.72 (CI 0.58–0.83), PPV 0.67 (CI 0.52–0.78), NPV 0.97 (CI 0.87–0.99), accuracy 0.81, LR(+) 3.48 (CI 2.26–5.38), LR (-) 0.04 (CI 0.01–0.31). Conclusion: The TBT is still a useful screening tool to rule out mandibular fractures in patients with facial trauma as compared to the current gold standard of HCT. history or presentation to the pediatric ED. EP read of the CXR was positive for pneumonia in 80% of patients; radiology read of the CXR was positive for pneumonia in 45% of study patients. In comparison to radiology read, the sensitivity of an EP read of pneumonia was 0.88 (95% CI 0.75–0.97) with a specificity of 0.27 (95% CI 0.15–0.34). The kappa statistic was 0.1 indicating poor agreement. Bedside US compared to radiology read had a sensitivity of 0.8 (95% CI 0.66–0.87) and a specificity of 0.93 (95% CI 0.81–0.99) with a kappa statistic of 0.74 indicating moderate agreement. Conclusion: Bedside US showed high correlation with radiology interpretation and may be a useful adjunct to radiographic evaluation in order to differentiate pneumonic infiltrates from other causes of consolidation. Table - Abstract 102: Two-by-Two Contingency Table 104 Tongue Blade Test + ) Total 103 andibular CT Fractures Facial Bones + ) Total 30 1 31 15 39 54 45 40 85 Bedside Ultrasound Evaluation Of Lung Infiltrates Stephanie G. Cohen1, Adam B. Sivitz2 1 Emory University Medical School, Atlanta, GA; 2 Newark Beth Israel Medical Center, Newark, NJ Background: CXR in pediatric patients is often obtained looking for infiltrates that may require antibiotics. A radiographic finding of consolidation may be difficult to differentiate from atelectasis leading the clinician to unnecessarily prescribe antibiotics. This is especially true for vague haziness that is commonly seen in the area of the right middle lobe (RML) of the lung. Objectives: Determine the utility of bedside US by pediatric emergency physicians (EP) to differentiate pneumonia from other causes of consolidation. Methods: This is a prospective observational study of patients aged 0–18 years presenting to an urban pediatric ED. Patients were eligible if CXR revealed a RML infiltrate noted by the pediatric EP or radiologist. Ultrasonography-trained pediatric emergency physicians performed focused lung US examination of the lower right parasternal area to specifically investigate the right middle lobe. Ultrasound findings were classified as positive if consolidation consistent with pneumonia was present, and negative when sonographic features consistent with atelectasis or lung sliding were present. Ultrasound images used for review were compared to the final pediatric radiologist read which was considered the gold standard. Images were reviewed by blinded study personnel unaware of CXR findings or clinical condition of the patient. Results: Fifty-five patients were enrolled in the study. The median age was 20 months (range 3–219 months), 56% of patients were male, and 83% had fever by Inter-rater Reliability of Emergency Physician Ultrasonography For Diagnosing Lower Extremity Deep Venous and Great Saphenous Vein Thromboses Compared To Ultrasonographic Studies Performed By Radiology Mary R. Mulcare, Tomas Borda, Debbie Hana Yi, Dana L. Sacco, Jennifer F. Kherani, David C. Riley New York Presbyterian Hospital, New York, NY Background: Several studies have been done comparing the accuracy of lower extremity ultrasonography done by emergency physicians (EPs) compared to an institution-specific criterion standard; however, none of these studies have included evaluation of the proximal great saphenous vein. A clot in the proximal great saphenous vein is usually an indication for anticoagulation. Objectives: The goal of this study is to assess the interrater reliability of EP-performed ultrasound examinations for lower extremity venous thromboembolism, including assessment of the common femoral vein (CFV), femoral vein of the thigh (FV), great saphenous vein (GSV), and popliteal vein (PV), compared to the clinical standard of ultrasound technologist performed, radiologist read examinations. We aim to demonstrate whether EPs can reliably identify venous thromboembolism in the studied veins. Methods: This clinical trial is a prospective, blinded, convenience sample, in an urban teaching hospital emergency department (ED), NewYork-Presbyterian Hospital/ Columbia University Medical Center and Allen Hospital, with an adult ED volume of approximately 120,000 patients/year. Patients with a clinical suspicion of a lower extremity venous thrombosis were enrolled. Each patient underwent an ED bedside ultrasonography exam for DVT in the four identified venous locations by an EP blinded to an official sonography exam completed by the department of radiology on the same patient. The data were analyzed using the two-rater unweighted kappa statistic to compare the two groups conducting the examination. To detect a kappa of 0.9, with a power of 0.8 and an acceptable type I error of 0.05, we need a sample size of 433 patients. This is an interim analysis. Results: There were a total of 102 enrollments included in this analysis, with 21 enrollments excluded for ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 incomplete data. The observed overall agreement and the inter-rater reliability (kappa) were: CFV, 95.1%, kappa of 0.64 (95% CI 0.35–0.93), FV, 94.9%, kappa of 0.59 (95% CI 0.26–0.92), PV, 90.9%, kappa of 0.43 (95% CI 0.12–0.73), and GSV, 91.1%, kappa of 0.42 (95% CI 0.11– 0.73). Conclusion: EPs can reliably evaluate for lower extremity venous thromboembolism, with substantial inter-rater agreement for the CFV and moderate interrater agreement for the FV, PV, and GSV. 105 A Prospective Evaluation of Emergency Department Bedside Ultrasonography for the Detection of Acute Pediatric Appendicitis David J. McLario1, Richard L. Byyny2, Michael Liao2, John L. Kendall2 1 Denver Health Medical Center, Denver, CO; 2 Denver Health Medical Center, Denver, CO Background: Appendicitis is the most common surgical emergency occurring in children. The diagnosis of pediatric appendicitis is often difficult and computerized tomography (CT) scanning is utilized frequently. CT, although accurate, is expensive, time-consuming, and exposes children to ionizing radiation. Radiologists utilize ultrasound for the diagnosis of appendicitis, but it may be less accurate than CT, and may not incorporate emergency physician (EP) clinical impression regarding degree of risk. Objectives: The current study compared EP clinical diagnosis of pediatric appendicitis pre- and post-bedside ultrasonography (BUS). Methods: Children 3–17 years of age were enrolled if their clinical attending physician planned to obtain a consultative ultrasound, CT scan, or surgical consult specific for appendicitis. Most children in the study received narcotic analgesia to facilitate BUS. Subjects were initially graded for likelihood of appendicitis based on research physician-obtained history and physical using a Visual Analogue Scale (VAS). Immediately subsequent to initial grading, research physicians performed a BUS and recorded a second VAS impression of appendicitis likelihood. Two outcome measures were combined as the gold standard for statistical analysis. The post-operative pathology report served as the gold standard for subjects who underwent appendectomy, while post 2-week telephone follow-up was used for subjects who did not undergo surgery. Various specific ultrasound measures used for the diagnosis of appendicitis were assessed as well. Results: 29/56 subjects had pathology-proven appendicitis. One subject was pathology-negative post-appendectomy. Of the 26 subjects who did not undergo surgery, none had developed appendicitis at the post 2-week telephone follow-up. Pre-BUS sensitivity was 48% (29–68%) while post-BUS sensitivity was 79% (60– 92%). Both pre- and post-BUS specificity was 96% (81– 100%). Pre-BUS LR+ was 13 (2–93), while post-BUS LR+ was 21 (3–148). Pre- and post-BUS LR- were 0.5 and 0.2, respectively. BUS changed the diagnosis for 20% of subjects (9–32%). • www.aemj.org S59 Conclusion: BUS improved the sensitivity of evaluation for pediatric appendicitis. BUS also confirmed EP clinical suspicion of appendicitis, with specificity comparable to historical norms for CT evaluation. 106 Sonographic Measurement of Glenoid to Humeral Head Distance in Normal and Dislocated Shoulders in the Emergency Department Brent Becker, Alan Chiem, Art Youssefian, Lynne Le, Michael Peyton, Negean Vandordaklou, Graciela Maldonaldo, Chris Fox UC Irvine, Orange, CA Background: There are very little data on the normal distance between the glenoid rim and the posterior aspect of the humeral head in normal and dislocated shoulders. While shoulder x-rays are commonly used to detect shoulder dislocations, they may be inadequate, exacerbate pain in the acquisition of some views, and lead to delay in treatment, compared to bedside ultrasound evaluation. Objectives: Our objective was to compare the glenoid rim to humeral head distance in normal shoulders and in anteriorly dislocated shoulders. This is the first study proposing to set normal and abnormal limits. Methods: Subjects were enrolled in this prospective observation study if they had a chief complaint of shoulder pain or injury, and received a shoulder ultrasound as well as a shoulder x-ray. The sonographers were undergraduate students given ten hours of training to perform the shoulder ultrasound. They were blinded to the x-ray interpretation, which was used as the gold standard. We used a posterior-lateral approach, capturing an image with the glenoid rim, the humeral head, as well as the infraspinatus muscle. Two parallel lines were applied to the most posterior aspect of the humeral head and the most posterior aspect of the glenoid rim. A line perpendicular to these lines was applied, and the distance measured. In anterior dislocations, a negative measurement was used to denote the fact that the glenoid rim is now posterior to the most posterior aspect of the humeral head. Descriptive analysis was applied to estimate the mean and 25th to 75th interquartile range of normal and anteriorly dislocated shoulders. Results: Eighty subjects were enrolled in this study. There were six shoulder dislocations, however only four were anterior dislocations. The average distance between the posterior glenoid rim and the posterior humeral head in normal shoulders was 8.7 mm, with a 25th to 75th inter-quartile range of 6.7 mm to 11.9 mm. The distance in our four cases of anterior dislocation was )11 mm, with a 25th to 75th interquartile range of )10 mm to )12 mm. Conclusion: The distance between the posterior humeral head to posterior glenoid rim may be 7 mm to 12 mm in patients presenting to the ED with shoulder pain but no dislocation. In contrast, this distance in anterior dislocations was greater than )10 mm. Shoulder S60 2012 SAEM ANNUAL MEETING ABSTRACTS second. Another blinded EP recorded two 6-second video clips of the intraosseous CD-US signal; one during normal crystalloid flow and another during a 10cc flush. Twenty blinded EPs later reviewed a randomized file of all video clips and rated each as (+) IO flow or (-) no IO flow. Results: Forty needles were placed and interrogated using the three confirmation methods. Eighty CD-US signal video clips were reviewed. The respective sensitivity and specificity of identifying IO placement by CD-US in the tibia was 89% (95CI 85–92) and 89% (95CI 85–92), in the humerus was 76% (95CI 71–80) and 80% (95CI 76– 84), and combined were 82% (95CI 79–85) and 84% (95CI 82–87). There was no difference between crystalloid flow and flush methods. The sensitivity and specificity of confirming placement with aspiration were 85% (95CI 61–96) and 95% (95CI 73–100). The sensitivity of confirming placement with crystalloid rate was 100% (95CI 80– 100), but the specificity was only 10% (95CI 1.8–33). Conclusion: In fresh cadavers, we found that visualization of intraosseous CD-US flow by EPs may be a reliable method of confirming IO placement. This method appears to have superior characteristics for confirmation than crystalloid rate and is comparable to aspiration. Since CD-US can be performed rapidly at any time during resuscitation, it may have the most utility of the available confirmation methods. 108 ultrasound may be a useful adjunct to x-ray for diagnosing anterior shoulder dislocations. 107 Confirmation Of Intraosseous Needle Placement With Color Doppler Ultrasound In An Adult Fresh Cadaver Model Kenton L. Anderson, Catherine Jacob, Joseph D. Novak, Daniel G. Conway, Jason D. Heiner San Antonio Military Medical Center, San Antonio, TX Background: Intraosseous (IO) needle placement during resuscitation is a valuable method of obtaining vascular access when an IV cannot promptly be established during resuscitation. Traditional methods of confirming IO placement include aspiration of blood/ marrow or the free flow of infused crystalloid. Color Doppler ultrasound (CD-US) has recently been proposed as a method to confirm IO needle placement. Objectives: We hypothesized that CD-US would have better test characteristics than traditional methods of confirming IO placement. Methods: DESIGN: A prospective observational study comparing confirmation methods for IO needles randomly placed either intraosseously or in adjacent soft tissue (ST) of the proximal tibia and humeral heads of five adult fresh unembalmed cadavers. Needle placement was verified by cutdown. OBSERVERVATIONS: Two emergency physicians (EPs) blinded to needle placement attempted to confirm IO placement either by successful aspiration or by crystalloid flow greater than one drop/ Introducing Bedside Limited Compression Ultrasound by Emergency Physicians into the Diagnostic Algorithm for Patients with Suspected DVT: A Prospective Cohort Trial Rachel Poley, Joseph Newbigging, Marco L.A. Sivilotti Queen’s University, Kingston, ON, Canada Background: Diagnosing deep venous thrombosis (DVT) relies on clinical characteristics (modified Wells score), serum D-dimer, and formal imaging, but can be inefficient. Objectives: To evaluate whether a novel diagnostic approach that incorporates bedside limited compression ultrasound (LCU) could be used to improve diagnostic efficiency for DVT. Methods: We performed a prospective cohort study of ED patients with suspected DVT. We excluded patients on anticoagulants, with a chronic DVT, leg cast or amputation, or when the results of formal imaging were already known. All patients were treated in the usual fashion based on the protocol in use at our centre: treating physicians classified patients as ‘‘DVT unlikely’’ or ‘‘DVT likely’’ using the modified Wells score, then obtained serum D-dimer (latex immunoassay) and/ or formal ultrasound imaging per protocol. Seventeen physicians were trained and performed LCU in all subjects. DVT was considered ruled out in ‘‘DVT unlikely’’ patients if the LCU was negative, and in ‘‘DVT likely’’ patients if both the LCU and D-dimer were negative. Results: We enrolled 227 patients (47% ‘‘DVT likely’’), of whom 24 had DVT. The sensitivity and specificity of the novel approach were 0.96 [95% CI 0.77, 1.00] and 0.66 [0.59, 0.72] respectively, compared with the current protocol 1.00 [0.83, 1.00] and 0.35 [0.28, 0.42]. Overall, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 the stand-alone sensitivity and specificity of LCU were 0.91 [0.70, 0.98] and 0.97 [0.92, 0.99]. Incorporating LCU into the diagnostic approach would have reduced the rate of formal imaging from 67% to 40%, the mean time to diagnostic certainty by 5.0 hours, and eliminated 24 (11%) return visits of whom 10 were empirically anticoagulated. The interobserver disagreement rate between the treating and scanning physicians for the Wells score was 19% (kappa 0.62 [0.48, 0.77]), including the one patient with DVT who would have been missed on the index visit using the new approach. Conclusion: Limited compression ultrasound holds promise as one component of the diagnostic approach to DVT, but should not be used as a stand-alone test due to imperfect sensitivity. Tradeoffs in diagnostic efficiency for the sake of perfect sensitivity remain a difficult issue collectively in emergency medicine, but need to be scrutinized carefully in light of the costs of over-investigation. 109 Point of Care Focused Cardiac Ultrasound for Pulmonary Embolism Short-Term Adverse Outcomes Jennifer M. Davis, Vishal Gupta, Rachel Liu, Christopher Moore, Andrew Taylor Yale University School of Medicine, New Haven, CT Background: Pulmonary embolism remains a potentially lethal disease, yet prior studies have demonstrated that for normotensive patients or patients without signs of right ventricular strain (RVS), adverse outcomes are less than 1%. Recent prognostic scoring symptoms have sought to identify patients who may be treated as outpatients, but have not incorporated echocardiographic measures of RVS. Objectives: To determine the sensitivity, specificity, and likelihood ratios of RVS using Point-of-Care Focused Cardiac Ultrasound (FOCUS) and hypotension for pulmonary embolism short-term adverse outcomes, and to compare them with the Pulmonary Severity Index (PESI) and Geneva predictor rule. Methods: Retrospective record review of ED patients between 1/2007–12/2010 who had both a diagnosis of pulmonary embolism by ICD9 code and a FOCUS exam. Adverse outcomes were defined as shock (SBP persistently less than 100 mmHg refractory to volume loading and requiring vasopressors), respiratory failure requiring intubation, death, recurrent venous thromboembolism, transition to higher level of care, or major bleeding within 7 days of admission. RVS on FOCUS was defined as the presence of either RV greater than or equal to LV, RV hypokinesis, or the presence of a McConnel’s sign. Results: 1318 records were identified with a diagnosis of pulmonary embolism of which 171 had a FOCUS performed. Mean age was 61 ± 18 and 47% were male. There were 27 adverse outcomes. The prevalence of RVS on FOCUS was 23%. Likelihood ratios (95%CI) for RVS, hypotension, RVS or hypotension, RVS + hypotension, and for neither RVS nor hypotension are 3.55(2.16–5.83), 2.08(1.15–3.77), 2.23(1.45–3.42), 4.33(1.43– 13), and 0.29(0.13–0.67) respectively. The table shows test characteristics for the prognostic rules. • www.aemj.org S61 Conclusion: In this retrospective study, the presence of RV strain on FOCUS significantly increases the likelihood of an adverse short term event from pulmonary embolism and its combination with hypotension performs similarly to other prognostic rules. Table - Abstract 109: Prognostic Rule RV strain or hypotension PESI Geneva Predictor Rule 110 Sensitivity (95%CI) Specificity (95%CI) 0.81(0.61–0.93) 0.78 (0.57–0.91) 0.63 (0.42–0.80) 0.69(0.60–0.77) 0.38(0.30–0.46) 0.54 (0.45–0.63) Indocyanine Green Dye Angiography Accurately Predicts Jackson Zone Survival in a Horizontal Burn Comb Model Mitchell S. Fourman, Brett T. Phillips, Laurie Crawford, Fubao Lin, Adam J. Singer, Richard A. Clark Stony Brook University Medical Center, Stony Brook, NY Background: Burns are expensive and debilitating injuries, compromising both the structural integrity and vascular supply to skin. They exhibit a substantial potential to deteriorate if left untreated. Jackson defined three ‘‘zones’’ to a burn. While the innermost coagulation zone and the outermost zone of hyperemia display generally predictable healing outcomes, the zone of stasis has been shown to be salvageable via clinical intervention. It has therefore been the focus of most acute therapies for burn injuries. While Laser Doppler Imaging (LDI) - the current gold standard for burn analysis - has been 96% effective at predicting the need for second degree burn excision, its clinical translation is problematic, and there is little information regarding its ability to analyze the salvage of the stasis zone in acute injury. Laser Assisted Indocyanine Green Dye Angiography (LAICGA) also shows potential to predict such outcomes with greater clinical utility. Objectives: To test the ability of LDI and LAICGA to predict interspace (zone of stasis) survival in a horizontal burn comb model. Methods: A prospective animal experiment was performed using four pigs. Each pig had a set of six dorsal burns created using a brass ‘‘comb’’ - creating four rectangular 10 · 20 mm full thickness burns separated by 5 · 20 mm interspaces. LAICGA and LDI scanning took place at 1 hour, 24 hours, 48 hours, and 1 week post burn using Novadaq SPY and Moor LDI respectively. Imaging was read by a blinded investigator, and perfusion trends were compared with interspace viability and contraction. Burn outcomes were read clinically, evaluated via histopathology, and interspace contraction was measured using Image J software. Results: LAICGA data showed significant predictive potential for interspace survival. It was 83.3% predictive at 24 hours post burn, 75% predictive 48 hours post burn, and 100% predictive 7 days post burn using a standardized perfusion threshold. LDI imaging failed to predict outcome or contraction trends with any degree of reliability. The pattern of perfusion also S62 2012 SAEM ANNUAL MEETING ABSTRACTS appears to be correlated with the presence of significant interspace contraction at 28 days, with an 80% adherence to a power trendline. Conclusion: Preliminary data suggest that LAICGA can potentially be used to predict burn extension, as well as to test the effectiveness of acute burn therapy. Figure – Abstract 110: LDI (left) and LAIGA (right) 48 hours post-burn 111 Ultrasound Experts Rapidly And Accurately Interpret Ultrasound Images Obtained Using Cellphone Video Cameras Transmitted By Cellular Networks Stephen Leech1, Jillian Davison1, Michelle P. Wan1, F. Eike Flach2, Linda Papa1 1 Orlando Regional Medical Center, Orlando, FL; 2Univerity of Florida Shands-Gainesville, Gainesville, FL Background: Emergency ultrasound (US) is an integral skill to the practice of EM, but is heavily user-dependent and has a steep initial learning curve. Direct supervision by an expert is not always available and may limit US use. Newer cell phones have built-in video cameras and allow video transmission via cellular networks, which would allow remote review, interpretation, and management by experts if not available on-site. Objectives: We hypothesize that experts can correctly interpret US images and guide management of patients in real time using US videos captured and transmitted via cellphone networks. Methods: This prospective observational study was a blinded image review. US images were captured using a cell phone video camera (iPhone 3GS) and transmitted to two fellowship-trained US experts via text messaging using a 3G data network. Experts interpreted images, returned interpretation, provided next step in patient management, and rated images on a five-point scale (1-poor, 2-fair, 3-adequate, 4-good, 5-excellent) via text message in real time. Experts were blinded to each other’s interpretations and ratings. Outcome measures included correct interpretation (normal/abnormal), correct next step in patient management, time from initial transmission to interpretation, and ratings of image quality. Data were analyzed using descriptive statistics, raw agreement, and Cohen’s j. Results: Two experts reviewed 50 videos from six core applications in emergency US (aorta, cardiac, DVT, FAST, GB, renal) for 100 total reviews. There were 14 normal and 36 abnormal US. Experts correctly interpreted 96/100 videos as normal or abnormal with excellent agreement, raw agreement 0.97 (95%CI 0.91–0.99), and Cohen’s j 0.93 (95%CI 0.84–1). Experts recommended the correct next step in management in 97/100 cases with excellent agreement, raw agreement 0.97 (95%CI 0.91–0.99), and Cohen’s j 0.94 (95%CI 0.87–1). Mean time from initial image transmission to interpretation was 141 seconds (95%CI 127–155) with a range from 45 seconds–504 seconds. Mean image quality rating was 4.0 (95%CI 3.9–4.1), with 98 images rated as adequate or better. Conclusion: Transmission of cell phone video camera captured US videos via text messaging allows rapid and accurate interpretation of US images by US experts. This medium can provide support to EM physicians who may not always be comfortable with US image interpretation. 112 Renal Colic: Does Urine Dip and/or Serum WBC Predict the Need For CT To Identify Kidney Stone Mimics? Raashee S. Kedia1, Kaushal Shah1, Nelson Wong1, David H. Newman1, Salah Baydoun2, Barbara Kilian2 1 Mount Sinai School of Medicine, New York, NY; 2St. Luke’s Roosevelt, New York, NY Background: n/a Objectives: Our primary objective was to identify the percentage of patients with serious alternative diagnoses when presenting with renal colic, and our secondary objective was to determine if immediately available clinical data can predict the presence or absence of dangerous alternative diagnoses. Methods: We conducted an observational study between January 2007 and June 2008 in two academic, inner city EDs with an annual census of 185,000. Inclusion criterion was ‘patient with possible renal colic’ per treating physicians. Exclusions were non-English speaking, non-literate, prisoners, and abdominal aortic aneurysm believed to be among the top three possibilities. Trained research assistants (RAs) staffed EDs on a university calendar from 8 am to midnight (62% of calendar days) and monitored the trackboard for potential cases of renal colic. If an attending physician or senior resident confirmed suspicion of renal colic, a data form was completed by the medical provider. Urine dispstick results, serum WBC, and CT scan results were recorded when obtained. ‘‘Serious alternative ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 diagnosis’’ on CT scan was defined a priori. Discharged patients were contacted after 4 weeks for follow-up. Descriptive statistics including 95% confidence intervals were calculated for all measures. Leukocytosis was defined as WBC >12 and abnormal urine dipstick was defined as presence of either leukocytes or nitrites. Results: 444 patients with suspected renal colic were enrolled. 300 (67.5%) received a CT scan of which 118 (39%) had a confirmed stone and 15 (5% [95%CI = 3– 8%]) had an alternative serious diagnosis for their flank pain. The other 124 imaged patients (41%) had no clear etiology on CT scan. Of the 144 (32.5%) that did not receive a CT scan, 42 (9.5%) were contacted and found to have no adverse outcome (hospital visit, future CT scan, or surgery within 14 days), and none appeared in the social security death index. Leukocytosis (+LR 1.3, -LR 0.9, sensitivity 29%) and abnormal urine dipstick (+LR 1.0, -LR 0.99, sensitivity 24%) either individually or combined (+LR 1.0, -LR 0.99, sensitivity 39%) yielded poor sensitivity and unhelpful likelihood ratios as a predictor for kidney stone or adverse outcome. Conclusion: Serious alternative diagnoses identified by CT were uncommon (5%) in this cohort of suspected renal colic patients, and leukocytosis and urine dipstick could not predict these individuals. 113 ‘‘Child in Hand’’ - A Prospective Cohort Study To Assess The Health Status And Psychosocial Distress Of Haitian Children One Year After 2010 Earthquake Srihari Cattamanchi1, Robert D. Macy1, Dicki Johnson Macy2, Amalia Ciottone3, Svetlana • www.aemj.org S63 Bivens3, Bader S. Al-Otaibi1, Majed Al-johani1, Shannon Straszewski1, Gregory Ciottone1 1 Harvard Medical School / Beth Israel Deaconess Medical Center, Boston, MA; 2 Boston Childrens Foundation, Boston, MA; 3 Child in Hand, Boston, MA Background: The 2010 earthquake that struck Haiti caused more than 200,000 deaths and significant damage to health care infrastructure. One year later Haitian children and youth are most at risk, with an estimated 500,000 orphans exposed to debilitating diseases. Objectives: To assess the health and psychosocial status of a sample of Haitian children, one year after the 2010 earthquake. Methods: A prospective cohort study, assessing the health and psychosocial status among Haitian children, one year after 2010 earthquake, from seven orphanages and two schools in and around Port Au Prince, Haiti. Children, ages 1–18 years, from 7 were included in the study. These children were assessed for any medical illness, which was diagnosed based on their chief complaints, history of presenting illness, vital signs, and physical examination by medical teams. Based on their findings, children were either treated on-site or sent to medical clinics for further evaluation and treatment. Some children (ages 7–18) were also screened for psychosocial distress by psychosocial teams using Child Psychosocial Distress Screener (CPDS). Results: A total of 423 Haitian children were assessed, out of whom 28 were excluded because of age >18, leaving 395 children included in the study. There were 209 (54%) males and 186 (47%) females. The mean age was 10.59 years (SD 3.84). Most common clinical findings S64 2012 SAEM ANNUAL MEETING ABSTRACTS were symptoms of abdominal pain (22.28%), joint and muscle pain (14.94%), fever (14.18%), headache (12.66%), and malnutrition (11.65%). From the above 395 children, 357 children completed CPDS screening. The seven-item CPDS rates psychosocial distress into three main public mental health treatment categories: general treatment (scores of 1–3) in 45 children, indicated treatment (scores of 4–7) in 238 children, and selected treatment (scores of 8–10) in 74 children. Mean CPDS score was 5.79 (SD 2.06), with a median score of 6.0 (range 1 to 10). Conclusion: We found only 20% of the symptoms correlated with clinical findings. Earthquake exposure, including entrapment, witnessing relatives die, displacement, and orphan status are all primary causes contributing to psychosocial distress and degradation among Haitian children. Coupled with poor hygiene, environmental and nutritional factors, our findings argue for simultaneous improvement of psychosocial well-being and child health care. 114 Systematic Review of Interventions to Mitigate the Effect of Emergency Department Crowding in the Event of a Respiratory Disease Outbreak Melinda J. Morton1, Kevin Jeng2, Raphaelle Beard1, Andrea Dugas1, Jesse M. Pines3, Richard E. Rothman1 1 Johns Hopkins School of Medicine, Baltimore, MD; 2Duke University School of Medicine, Durham, NC; 3George Washington University School of Medicine, Washington, DC Background: Seasonal influenza is a common cause of ED crowding; however, increased patient volumes associated with a true influenza pandemic will require additional planning and ED response resources. Objectives: This systematic review aimed to describe the breadth and diversity of interventions that have been reported to improve patient flow during a respiratory outbreak. Secondarily, we qualitatively assessed the effectiveness of various types of interventions to determine which interventions may be most effective in different settings to mitigate surge during an outbreak. Methods: We conducted a formal literature search including MEDLINE, EMBASE, Cochrane, PubMed, Global Health Library (WHO), ISI Web of Science, and CINAHL databases. Interventions to mitigate influenza or any known respiratory pathogen were included. Initial search results were screened by title and abstract; studies were excluded based on criteria listed in Table 1. Six intervention categories were identified a priori: Triage and Screening, Clinic-Based, Testing, Treatment, Isolation, and ‘‘Other’’ Interventions. Data on outbreak and intervention characteristics, facility characteristics, ‘‘triggers’’ for implementing interventions, and input / output measures were extracted. Results: 1761 articles were identified via the search algorithm. 1638 were excluded based on title and abstract. Of the 173 articles remaining, full text was reviewed on 136 (full text not available on 37 articles); 24 articles were selected for the final review. For full results, see Table 2. Sixteen Triage and Screening Interventions, 12 Clinic-Based, 11 Isolation, 4 Testing, 4 Treatment, and 1 ‘‘Other’’ category intervention were identified. One intervention involving school closures was associated with a 28% decrease in pediatric ED visits for respiratory illness. Conclusion: Most interventions were not tested in isolation, so the effect of individual interventions was difficult to differentiate. Interventions associated with statistically significant decreases in ED crowding were school closures, as well as interventions in all categories studied. Further study and standardization of intervention input, process, and outcome measures may assist in identifying the most effective methods of ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Table 1- Abstract 114:. Study Inclusion / Exclusion Criteria Inclusion Criteria Exclusion Criteria • Include studies with interventions designed to impact ED workflow in response to a known respiratory outbreak. Studies must include a discussion of the observed intervention effect on ED surge capacity. • Include peer-reviewed journal articles, peer-reviewed reports/ papers by nongovernmental organizations, and policy and procedure documents • Include all study types • Exclude non-English publications • Exclude studies that do not take place during a known, documented respiratory outbreak • Exclude studies that do not describe interventions that impact ED workflow • Exclude studies that detail interventions implemented indepen dently of ED surge; i.e. exclude studies with interventions that did not require a trigger for activation • Exclude interventions that were published in 2000 or earlier • Exclude animal studies • Exclude studies that have not been field-validated Table 2- Abstract 114: Output Measures, Interventions, and Results Output Measure1 Change in length of stay Change in wait time Result1 • Decreased from 241 to 212 minutes • Decreased by 2 hours • Decreased for all levels of acuity (from 0.4 to 2.1 hours) • Decreased by up to 3.5 hours • Decreased from 92.8 to 81.2 min • ‘‘Decreased’’ but not quantified in 4 studies • 44% decrease Associated Intervention • Triage and screening • Clinic, Triage and Screening, Treatment • All types of interventions • Clinic, Testing, Isolation • Triage and Screening • Various • All types implemented • Decreased from • Triage and 12% Screening, to 1% Clinic • Decreased to 4.8% • Isolation, Clinic, Triage and Screening Change in • Decreased by 3% • All types hospital implemented admissions • Decreased by 17% • Triage and Screening, Clinic, Isolation • Decreased from • Triage and 48.6% to 12.2% Screening • Decreased from • Isolation, Clinic, 21% Triage and to 18% Screening Cost • $59,000 / 30 days • Triage and Screening • $280,000 • All types implemented 1 By study, among studies reporting this data Change in left without being seen rates • www.aemj.org S65 mitigating ED crowding and improving surge capacity during an influenza or other respiratory disease outbreak. 115 Communication Practices and Planning in US Hospitals Caring for Children During the H1N1 Influenza Pandemic Marie M. Lozon1, Sarita Chung2, Daniel Fagbuyi3, Steven Krug4 1 University of Michigan, Ann Arbor, MI; 2 Children’s Hospital Boston, Boston, MA; 3 Children’s National Medical Center, Washington D.C., DC; 4Children’s Memorial Hospital, Chicago, IL Background: The H1N1 influenza pandemic produced surges of children presenting to US emergency departments (EDs) in both spring and fall of 2009. Communication strategies were critical to relay the most current information to both the public and ED and hospital staff. Objectives: To identify communication practices used by hospitals caring for children during the 2009 H1N1 influenza pandemic. Methods: A retrospective survey tool was developed and refined using a modified Delphi method with nine subject matter experts in pediatric disaster preparedness and distributed to ED medical directors and/or hospital disaster medical leaders who reported on institutional preparedness practices. The survey tool was sent to hospitals across the ten Federal Emergency Management Agency (FEMA) regions. We describe hospitals’ use and modification of communication strategies during the pandemic. Results: Seventy-eight hospitals were surveyed, with 52 responding (69%). 47% of participants reported having an external communication plan (ECP) for families, primary care physicians, and the public prior to the pandemic. 79% reported that the ECP required further modifications. Modifications include treatment and testing plans for primary care providers (76%), hotline/ central call center for community health care providers (58%), creation of a central website (53%), direct telephone messages to families (39%), and use of social media (26%). 84% of participants reported having an established internal communication plan (ICP) for staff prior to the pandemic. 74% reported that the ICP required further modifications. Modifications include creation of algorithms for treatment and testing (89%), flexibility to make institutional changes for testing and treatment based on CDC guidelines (69%), algorithms for patient placement to inpatient units (61%), e-mail messaging to all staff members (53%), and central website (36%). Conclusion: During H1N1 pandemic, more institutions report having an established internal communication plan than external plans. Most plans required further modifications predominantly around testing and treating policies, with some participants reporting modifications of the external communication plan to include the use of social media. S66 116 2012 SAEM ANNUAL MEETING ABSTRACTS An Investigation of the Association between Extended Shift Lengths, Sleepiness, and Occupational Injury and Illness among Nationally Certified EMS Professionals Antonio Ramon Fernandez1, J. Mac Crawford2, Jonathan R. Studnek3, Michael L. Pennell2, Timothy J. Buckley2, Melissa A. Bentley4, John R. Wilkins III2 1 EMS Performance Improvement Center University of North Carolina - Chapel Hill, Chapel Hill, NC; 2The Ohio State University, Columbus, OH; 3The Center for Prehospital Medicine, Carolinas Medical Center and the Mecklenburg EMS Agency, Mecklenburg, NC; 4The National Registry of EMTs, Columbus, OH Background: The link between extended shift lengths, sleepiness, and occupational injury or illness has been shown, in other health care populations, to be an important and preventable public health concern but heretofore has not been fully described in emergency medical services (EMS). Objectives: Evaluate the influence of extended shift lengths and sleepiness on occupational injury or illness among nationally certified EMS professionals. Methods: In 2009, previous respondents to the Longitudinal EMT Attributes and Demographics Study were mailed a survey. This survey included the Epworth Sleepiness Scale (ESS), items inquiring about health and safety, and work-life characteristics. Occupational injury or illness items (involved in ambulance crash while driving, missed work due to occupational injury or illness, needle-stick while performing EMS duties) were combined generating one dichotomous outcome variable. Multiple logistic regression modeling was performed, forcing ESS and shift length (<24 hours, ‡24 hours) into the model. Results: The response rate was 67.2% (n = 1,078). Ambulance crashes, missed work due to occupational injury or illness, and needle sticks were reported in 1.8%, 13.1%, and 4.0% of respondents, respectively. Combining these variables revealed 17.5% (186/1,060) experienced occupational injury or illness in the past 12 months. The relationship between shift length and odds of occupational injury or illness differed according to overtime work (p = 0.01): among individuals who did not work mandatory overtime, the odds of occupational injury or illness for those who worked ‡24-hours was 1.72 (95% CI 1.01–2.95) times that of individuals who worked <24-hours. There was no statistically significant difference when comparing shift lengths for those who worked mandatory overtime (OR 1.13, 95% CI 0.59– 2.18). For every ESS point increase, the odds of reporting an occupational injury or illness increased by 7% adjusting for shift-length and overtime work (OR 1.07; 95%CI 1.03–1.12). Conclusion: This study revealed significant associations between occupational injury or illness and shift length, working mandatory overtime, and sleepiness. Results suggest that EMS professionals’ health and safety can be protected by preventing mandatory overtime and extended shift lengths. 117 Does An Intimate Partner Violence Kiosk Intervention in the ED Impact Subsequent Safety Behaviors? Justin Schrager, Debra Houry, Shakiyla Smith Emory University School of Medicine, Atlanta, GA Background: Computer kiosks in the ED have not previously been employed to screen for intimate partner violence (IPV) or disseminate health information. Objectives: To assess the effect of an ED-based computer screening and referral intervention for IPV victims and to determine what characteristics resulted in a positive change in their safety. We hypothesized that women who were experiencing severe IPV and/or were in contemplation or action stages would be more likely to endorse safety behaviors. Methods: We conducted the intervention for female IPV victims at three urban EDs using a computer kiosk to deliver targeted education about IPV and violence prevention as well as referrals to local resources. All adult English-speaking non-critically ill women triaged to the ED waiting room were eligible to participate. The validated Universal Violence Prevention Screening Protocol was used for IPV screening. Any who disclosed IPV further responded to validated questionnaires for alcohol and drug abuse, depression, and IPV severity. The women were assigned a baseline stage of change (precontemplation, contemplation, action, or maintenance) based on the URICA scale for readiness to change behavior surrounding IPV. Participants were contacted at 1 week and 3 months to assess a variety of pre-determined actions such as moving out, to prevent IPV during that period. Statistical analysis (chi-square testing) was performed to compare participant characteristics to the stage of change and whether or not they took protective action. Results: A total of 1,474 people were screened and 154 disclosed IPV and participated in the full survey. 53.3% of the IPV victims were in the precontemplative stage of change, and 40.3% were in the contemplation stage. 110 women returned at 1 week of follow-up (71.4%), and 63 (40.9%) women returned at 3 months of followup. 55.5% of those who returned at 1 week, and 73% of those who returned at 3 months took protective action against further IPV. There was no association between the various demographic characteristics and whether or not a woman took protective action. Conclusion: ED-based kiosk screening and health information delivery is both a feasible and effective method of health information dissemination for women experiencing IPV. Stage of change was not associated with actual IPV protective measures. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 118 Prime Time Television Programing Fails to Lead Safely By Example: No Seat Belts No Helmets David Milzman1, Han Huang1, Kyle Pasternac1, Michael Phillipone2, Jenica Ferretti-Gallon2, Anna Ruff2 1 Georgetown University School of Medicine, Washington, DC; 2Georgetown University, Washington, DC Background: A 1998 Michigan State University study recorded prime time TV portrayal of 25% seat belt usage when actual national usage was 65% that year. In 13 years since, the US reported usage approaches 90%. Other safety precautions for personal protection, such as helmet use for motor and pedal cycles, remain at 50%. Objectives: Compare prime time TV traffic/safety exposures and seat belt and helmet use with USDOT NHTSA figures. Methods: Researchers watched non-news, non realityTV totaling 53 programs across 10 weeks of spring 2011 prime time (8–11 PM EST) from the following networks: ABC, CBS, NBC, FOX. Commercials were excluded. All instances of seat belt usage (driver and passenger), helmets (bikes and motorcycle), and miscellaneous pedestrian and vehicular traffic infractions were also recorded. Results: a total of 273 hours of prime time TV was viewed with an overall rate for proper seat belt usage in 37.6% (95% CI 32.4–42.9) of drivers, and 22.3% (95% CI18.5–26.0) of passengers. Proper seating and childseat usage, not noted in original 1998 study, was only 14%. Helmets were used by 15.9% of bicyclists, and 70.3% of motorcyclists. There was 17% rate of pedestrian and 22% vehicular traffic violations, also. Overall proper 2011 restraint use was 30.1% (95%CI 25.4–34.6). This figure represents only a 4.2% rise and NS increase since the prior study. Portrayal of prime time TV seatbelt usage rose a 4.8% (p £ 0.11) from 1998 to 2011 while actual US seat belt use increased significantly. Conclusion: Recent studies have found traffic safety behaviors continue to increase in US population; however, major TV network programs have not incorporated such simple safety changes into current programming despite prior study into these deficiencies. A poor example continues to be set. 119 The Effect Of Young Unlicensed Drivers On Passenger Safety Restraint Use In U.S. Fatal Crashes: Concern For Risk Spillover Effect? Jonathan Fu1, Michael J. Crowley1, Jim Dziura1, Craig L. Anderson2, Federico E. Vaca1 1 Yale University School of Medicine, New Haven, CT; 2University of California, Irvine School of Medicine, Irvine, CA Background: Despite recent prevention gains, motor vehicle crashes continue to top the list of causes of death for US adolescents and young adults. Many of these deaths involve young unlicensed drivers who are more likely to be in fatal crashes and to engage in • www.aemj.org S67 high-risk driving behaviors like impaired driving, speeding, and driving unrestrained. In a crash context, the influence of these high-risk behaviors may spill over to adversely affect passengers safety restraint use. Objectives: To examine the effect of young unlicensed drivers on safety restraint use and mortality of their front seat passengers. Methods: A retrospective analysis of the National Highway Traffic Safety Administration’s Fatality Analysis Reporting System from years 1996–2008 was conducted. Fatal crashes involving unlicensed drivers (15– 24 yrs) and their front seat passengers (15–24 yrs) were included. Contingency tables, univariate, and multivariate logistic regression were undertaken to assess the relationship between unlicensed driving and passenger restraint use, controlling for established predictors of restraint use, including sex, time of day, alcohol use, number of occupants, crash year, and crash location (rural vs. urban). Results: 85,563 15–24 year-old front seat passenger crash fatalities occurred from 1996–2008. 14,447 (19%) of their drivers were unlicensed or inadequately licensed. Rates of unlicensed driving ranged from 17% to 21% and trended upwards. Compared to passengers of licensed drivers, passengers of unlicensed drivers had decreased odds of wearing a safety restraint (OR 0.65, 95% CI 0.63–0.67). Other significant factors were male passenger (0.75, 0.73–0.77), driver drinking (0.37, 0.36–0.39), rural location (0.62, 0.60–0.64), and crash year (1.06, 1.06–1.07). Conclusion: We found a strong negative correlation between unlicensed driving and front seat passenger restraint use, suggesting a significant risk spillover effect. Unlicensed driving is involved in a disproportionate number of fatal crashes and plays an important role in the safety of not only the drivers but also their passengers. Our findings highlight an alarming trend that has considerable implications for US highway safety and the public’s health. Further in-depth study in this area can guide future countermeasures and traffic safety programs. 120 Emerging Conducted Electrical Weapon Technology: Is it Effective at Stopping Further Violence? Jeffrey D. Ho1, Donald M. Dawes2, James D. Sweeney3, Paul C. Nystrom1, James R. Miner1 1 Hennepin County Medical Center, Minneapolis, MN; 2Lompoc Valley Medical Center, Lompoc, CA; 3Florida Gulf Coast University, Ft. Myers, FL Background: Conducted electrical weapons (CEWs) are effective law enforcement tools used to control violent persons, thus preventing further violence and injury. Older generation TASER X26 CEWs are most widely in use. New generation TASER X2 CEWs will begin to replace them. X2 technology differences are a multi-shot feature and redesigned electrical waveform/output characteristics. It is not known if the X2 will be as effective in preventing further violence and injury. S68 2012 SAEM ANNUAL MEETING ABSTRACTS Objectives: We present a pilot, head-to-head comparison of X26 and X2 effectiveness in stopping a motivated person. The objective is to determine comparative injury prevention effectiveness of the newer CEW. Methods: Four humans had metal CEW probe pairs placed. Each volunteer had two probe pairs placed (one pair each on the right and left of the abdomen/inguinal region). Superior probes were at the costal margin, 5 inches lateral of midline. Inferior probes were vertically inferior at predetermined distances of 6, 9, 12, and 16 inches apart. Each volunteer was given the goal of slashing a target 10 feet away with a rubber knife during CEW exposure. As a means of motivation, they believed the exposure would continue until they reached the goal (in reality, the exposure was terminated once no further progress was made). Each volunteer received one exposure from a X26 and a X2 CEW. The exposure order was randomized with a 2-minute rest between them. Exposures were recorded on a hi-speed, hi-resolution video. Videos were reviewed and scored by six physician, kinesiology, and law officer experts using standardized criteria for effectiveness including degree of upper and lower extremity, and total body incapacitation, and degree of goal achievement. Reviews were descriptively compared independently for probe spread distances and between devices. Results: There were 8 exposures (4 pairs) for evaluation and no discernible, descriptive reviewer differences in effectiveness between the X26 and the X2 CEWs when compared. Conclusion: New generation CEWs have improved safety technology while exhibiting similar performance when compared to older generation CEWs. There is no discernible effectiveness difference between them. New generation CEWs appear to be as effective in stopping a motivated person when compared to older generation CEWs. 121 Barriers to Colorectal Cancer Screening as Preventive Health Measure among Adult Patients Presenting to the Emergency Department Nidhi Garg, Sanjey Gupta New York Hospital Queens, Flushing, NY Background: Colorectal cancer is the second leading cause of cancer death in the United States. Health promotion and disease prevention are increasingly recognized activities that fall within the scope of emergency medicine. Objectives: To identify barriers in colorectal cancer screening in adult patients (‡50 yrs) of age presenting to an ED that services a large immigrant and non-English speaking population. Methods: A prospective, cross-sectional, survey based study was conducted at urban Level I trauma center with annual ED visits of 120,000/year. Trained research assistants interviewed a convenience sample of patients over 36 months with a three-page survey recording demographics, knowledge of occult blood in stool testing (OCBST), and colonoscopy based on current preventive health recommendations from the Agency for Healthcare Research and Quality (AHRQ). Chi-square test was used for categorical data as appropriate. Logistic regression was performed for the significant factors. Results: A total of 904 males ‡50 yrs were interviewed over the study period with a median age of 68 yrs (IQR 59–79), 384 (34%) immigrants and 825 (91%) insured. Overall, 647 (71%) subjects and 714 (79%) subjects had the knowledge about OCBST and colonoscopy, respectively and 24 (2.6%) and 27 (3%) did not answer the respective questions. Total 581 (64%) and 606 (67%) subjects had OCBST and colonoscopy in the past, and 28 (3.1%), 29 (3.2%) did not answer the respective questions. There was no significant difference in OCBST and colonoscopy stratified by immigrant status. Separate logistic regression models were compared with OCBST and colonoscopy as outcomes, adjusted with knowledge of obtaining these tests by AHRQ recommendations, age, immigration status, insurance status, and smoking status. OCBST and colonoscopy were associated with having knowledge about OCBST and colonoscopy, OR 36 (CI 23–56), p < 0.001 and OR 26 (CI 15–43), p < 0.001 and insurance status OR 2.2 (CI 1.2–4.2), p = 0.02 and OR 3.5 (CI 1.9–6.2), p < 0.001. Immigration status was associated with OCBST OR 1.6 (CI 1.1–2.4), p = 0.02 but was not associated with colonoscopy. Conclusion: Males are more likely to have prostate cancer screening if they have insurance and are educated about it. These interventions and preventive health practice reinforcements can be easily accomplished in ED while interviewing a patient. 122 Impact of Rising Gasoline Prices on Bicycle Injuries in the United States, 1997–2009 Mairin Smith, M. Kathryn Mutter, Jae Lee, Jing Dai, Mark Sochor, Matthew J. Trowbridge University of Virginia, Charlottesville, VA Background: The trend towards higher gasoline prices over the past decade in the U.S. has been associated with higher rates of bicycle use for utilitarian trips. This shift towards non-motorized transportation should be encouraged from a physical activity promotion and sustainability perspective. However, gas price induced changes in travel behavior may be associated with higher rates of bicycle-related injury. Increased consideration of injury prevention will be a critical component of developing healthy communities that help safely support more active lifestyles. Objectives: The purpose of this analysis was to a) describe bicycle-related injuries treated in U.S. emergency departments between 1997 and 2009 and b) investigate the association between gas prices and both the incidence and severity of adult bicycle injuries. We hypothesized that as gas prices increase, adults are more likely to shift away from driving for utilitarian travel toward more economical non-motorized modes of transportation, resulting in increased risk exposure for bicycle injuries. Methods: Bicycle injury data for adults (16–65 years) were obtained from the National Electronic Injury Surveillance System (NEISS) database for emergency ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 department visits between 1997–2009. The relationship between national seasonally adjusted monthly rates of bicycle injuries, obtained by a seasonal decomposition of time series, and average national gasoline prices, reported by the Energy Information Administration, was examined using a linear regression analysis. Results: Monthly rates of bicycle injuries requiring emergency care among adults increase significantly as gas prices rise (p < 0.0001, see figure). An additional 1,149 adult injuries (95% CI 963–1,336) can be predicted to occur each month in the U.S. (>13,700 injuries annually) for each $1 rise in average gasoline price. Injury severity also increases during periods of high gas prices, with a higher percentage of injuries requiring admission. Conclusion: Increases in adult bicycle use in response to higher gas prices are accompanied by higher rates of significant bicycle-related injuries. Supporting the use of non-motorized transportation will be imperative to address public health concerns such as obesity and climate change; however, resources must also be dedicated to improve bicycle-related injury care and prevention. 123 Obesity and Seatbelt Use: A Fatal Relationship Dietrich Jehle, Joseph Consiglio, Jenna Karagianis, Gabrielle Jehle SUNY@Buffalo, Williamsville, NY Background: Motor vehicle crashes are a leading cause of mortality in the United States. Although seatbelts significantly reduce the risk of death, a number of subgroups of individuals tend not to wear their seatbelts. A third of the population is now considered to be obese and obese drivers may find it more difficult to buckle up a standard seatbelt. Objectives: In this study, we hypothesized that obese drivers were less likely to wear seatbelts than their normal weight counterparts. Methods: A retrospective study was conducted on the drivers in severe motor vehicle crashes entered into the FARS (Fatality Analysis Reporting System) database • www.aemj.org S69 between 2003 and 2009. This database includes all motor vehicle crashes in United States that resulted in a death within 30 days. The study was limited to drivers (336,913) of passenger vehicles in severe crashes. A number of pre-crash variables were found to be significantly associated with seatbelt use. These were entered into a multivariate logistic regression model using stepwise selection. Drivers were grouped into weight categories based on the World Health Organization definitions of obesity by BMI. Seatbelt use was then examined by BMI, adjusted for 12 pre-crash variables that were significantly associated with seatbelt use. Results: The odds of seatbelt use for normal weight individuals were found to be 67% higher than the odds of seatbelt use in the morbidly obese. The table below displays the relationship of seatbelt use between the different weight groups and the morbidly obese. Odds ratios (OR) for each comparison are displayed with the lower and upper 95% confidence limits. Conclusion: Seatbelt use is significantly less likely in obese individuals. Automobile manufacturers need to investigate methods of making seatbelt use easier for the obese driver in order to save lives in this population. Table - Abstract 123: Underweight vs. Morbidly Obese Normal Wt. vs. Morbidly Obese Overweight vs. Morbidly Obese Slightly Obese vs. Morbidly Obese Moderately Obese vs. Morbidly Obese 124 OR L CL U CL 1.616 1.666 1.596 1.397 1.233 1.464 1.538 1.472 1.284 1.120 1.784 1.805 1.730 1.520 1.358 Days Out of Work Do Not Correlate with Emergency Department Pain Scores for Patients with Musculoskeletal Back Pain Barnet Eskin, John R. Allegra Morristown Memorial Hospital, Morristown, NJ Background: This is a secondary analysis of data collected for a randomized trial of oral steroids in S70 2012 SAEM ANNUAL MEETING ABSTRACTS emergency department (ED) musculoskeletal back pain patients. We hypothesized that higher pain scores in the ED would be associated with more days out of work. Objectives: To determine the degree to which days out of work for ED back pain patients are correlated with ED pain scores. Methods: Design: Prospective cohort. Setting: Suburban ED with 80,000 annual visits. Participants: Patients aged 18–55 years with moderately severe musculoskeletal back pain from a bending or twisting injury £ 2 days before presentation. Exclusion criteria included nonmusculoskeletal etiology, direct trauma, motor deficits, and employer-initiated visits. Observations: We captured initial and discharge ED visual analog pain scores (VAS) on a 0–10 scale. Patients were contacted approximately 5 days after discharge and queried about the days out of work. We plotted days out of work versus initial VAS, discharge VAS, and change in VAS and calculated correlation coefficients. Using the Bonferroni correction because of multiple comparisons, alpha was set at 0.02. Results: We analyzed 67 patients for whom complete data were available. The mean age was 40 ± 9 years and 30% were female. The average initial and discharge ED pain scales were 8.0 ± 1.5 and 5.7 ± 2.2, respectively. On follow-up, 88% of patients were back to work and 36% did not lose any days of work. For the plots of the days out of work versus the initial and discharge VAS and the change in the VAS, the correlation coefficients (R2) were 0.03 (p = 0.17), 0.08 (p = 0.04), and 0.001 (p = 0.87), respectively. Conclusion: For ED patients with musculoskeletal back pain, we found no statistically significant correlation between days out of work and ED pain scores. 125 Prevalence of Cardiovascular Disease Risk Factors Among a Population of Volunteer Firefighters David Jaslow, Melissa Kohn, Molly Furin Albert Einstein Medical Center, Philadelphia, PA Background: Cardiovascular disease (CVD) is cited annually by NIOSH as the most common cause of line of duty death among U.S. firefighters (FF). There is scant scientific literature and no state or national databases to document CVD risk factors among the volunteer FF workforce, which represents 71% of all FF nationwide. Objectives: To describe CVD risk factors among a population of volunteer FF. Methods: 24 (100%) FF age 18 and older of a rural Pennsylvania volunteer fire department completed an NFPA 1582 fitness for duty exam in 2011 that was funded by a FEMA Assistance to Firefighters Grant. The exams consisted of OSHA 1910.134 respiratory protection questionnaire, complete physical exam, blood work and urinalysis, vision and hearing tests, 12lead ECG, hepatitis titers, colon cancer screening, chest x-ray and PFTs if indicated, and a personal wellness profile. We performed a cross-sectional observational study of this population to determine the prevalence of eight ‘‘major’’ CVD risk factors cited by the AHA: age, sex, blood pressure or history of hypertension, BMI, elevated blood sugar or history of diabetes, cholesterol levels, smoking history, and level of physical activity. Descriptive statistics, effect size, and 95% confidence intervals are reported. Results: All (100%) FF in this volunteer department are male and their median age is 44 (range 19–87). 23/24 (96%, 95% CI: 78, 99%) have BMIs in the overweight or obese category. In their personal wellness profile, 17/24 (71%, 95% CI: 50, 85%) report physical inactivity, 12/24 (50%, 95% CI: 31, 69%) have a family history of CVD, 8/ 24 (33%, 95% CI:18, 53%) are smokers, and only one FF was diabetic. 14/24 (58%, 95% CI: 39, 76%) FF had either a history of hypertension or were hypertensive upon exam. 14/24 (58%, 95% CI: 39, 76%) FF also had either a history of hyperlipidemia or elevated total cholesterol, triglycerides, or LDL on their blood panel. Conclusion: 22/24 (92%, 95% CI: 74, 98%) firefighters in a small rural volunteer fire department have four or more CVD risk factors and 15/24 (63%, 95% CI:43, 91%) have five or more CVD risk factors. Fire department physicians and EMS medical directors who serve as consultants to the volunteer fire service should promote annual or biannual fitness for duty examinations as a tool to evaluate both personal and agency-wide CVD risk. The FEMA AFG program is one option to fund such an endeavor. 126 Prevalence Rates Of Intimate Partner Violence Perpetration Among Male Emergency Department Patients Daniel S. Bell, Shakiyla Smith, Debra E. Houry Emory University, Atlanta, GA Background: The Institute of Medicine recently recommend screening women for IPV in health care settings, but little research has been conducted on screening men for perpetration. Objectives: To assess the feasibility of ED computerized screening for IPV perpetration (IPVP) and the prevalence of IPVP among urban ED male patients. We hypothesized there would be an 80% acceptance rate in screening and that we would identify an IPVP rate of 20% in male patients. Methods: We conducted a cross-sectional study over a 6-week period in an urban ED at a Level I trauma center using computer kiosks loaded with validated scales to identify male perpetrators of IPVP. All male adult patients over 18 years of age who presented to the ED during study hours triaged to the waiting room were eligible to participate. Patients were excluded if they are non-English speaking, acutely intoxicated, critically ill, or either medically or psychiatrically and unable to complete a 20-minute questionnaire. At a private computer kiosk patients answered questions on general health, substance abuse, mental health, and IPVP using an eight-item IPVP screen developed by KV Rhodes, three substance abuse screens (HONC, TWEAK, and DAST), and two mental health surveys (PC-PTSD and PC-BDI). Patients received target health and resource information after the survey. IPVP prevalence was evaluated with descriptive statistics. Chi-square tests analyzed differences in among perpetrators and non-perpetrators. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Results: 113 men were approached for survey, and 94 were eligible for the study, of whom 67 (71%) completed questionnaires; 19 were not eligible and 27 were not interested. Of the men who had been in a relationship in the past year (n = 25), 16% screened positive for intimate partner violence perpetration and 44% endorsed at least one IPV perpetration behavior. Conclusion: We found that men accepted a screening protocol for perpetration in the ED and self-disclosed perpetration at relatively high rates. 127 Human Physiologic Effects of a New Generation Conducted Electrical Weapon Jeffrey D. Ho1, Donald M. Dawes2, Paul C. Nystrom1, James R. Miner1 1 Hennepin County Medical Center, Minneapolis, MN; 2Lompoc Valley Medical Center, Lompoc, CA Background: Conducted Electrical Weapons (CEWs) are common law enforcement tools used to subdue and repel violent subjects and, therefore, prevent further injury or violence from occurring in certain situations. The TASER X2 is a new generation of CEW that has the capability of firing two cartridges in a ‘‘semi-automatic’’ mode, and has a different electrical waveform and different output characteristics than older generation technology. There have been no data presented on the human physiologic effects of this new generation CEW. Objectives: The objective of this study was to evaluate the human physiologic effects of this new CEW. Methods: This was a prospective, observational study of human subjects. An instructor shot subjects in the abdomen and upper thigh with one cartridge, and subjects received a 10-second exposure from the device. Measured variables included: vital signs, continuous spirometry, pre- and post-exposure ECG, intra-exposure echocardiography, venous pH, lactate, potassium, CK, and troponin. Results: Ten subjects completed the study (median age 31.5, median BMI 29.4, 80% male). There were no important changes in vital signs or in potassium. The • www.aemj.org S71 median increase in lactate during the exposure was 1.2, range 0.6 to 2.8. The median change in pH was )0.031, range )0.011 to 0.067. No subject had a clinically relevant ECG change, evidence of cardiac capture, or positive troponin up to 24 hours after exposure. The median change in creatine kinase (CK) at 24 hours was 313, range )40 to 3418. There was no evidence of impairment of breathing by spirometry. Baseline median minute ventilation was 14.2, which increased to 21.6 during the exposure (p = 0.05), and remained elevated at 21.6 post-exposure (p = 0.01). Conclusion: We detected a small increase in lactate and decrease in pH during the exposure, and an increase in CK 24 hours after the exposure. The physiologic effects of the X2 device appear similar to previous reports for ECD devices. 128 Use of Urinary Catheters in U.S. EDs 1995– 2009: A Potentially Modifiable Cause of Catheter-Associated Urinary Tract Infection? Jennifer Gibson Chambers1, Jeremiah Schuur2 1 University of New England, Biddeford, ME; 2 Brigham and Women’s Hospital, Boston, MA Background: Catheter-associated urinary tract infection (CAUTI) is the most prevalent hospital-acquired infection. In 2007, the Centers for Disease Control (CDC) published guidelines for reducing CAUTI, including appropriateness criteria for urinary catheters (UCs). Objectives: To calculate frequency and trends of UC placement and potentially avoidable UC (PAUC) placement in US EDs. We hypothesized that ED use of UCs and PAUCs in admitted patients did not decrease after the CDC’s guideline publication. Methods: We analyzed the National Hospital Ambulatory Medical Care Survey (NHAMCS), a weighted probability sample of US ED visits, from 1995–2009 for use of UCs in adults. UCs were classified as PAUC if the primary diagnosis did not meet CDC appropriateness criteria. Use of UCs and PAUCs before (2005–7) and after (2008–9) the CDC guideline were compared with a S72 2012 SAEM ANNUAL MEETING ABSTRACTS chi-square test. Predictors of ED placement of UC for admitted patients were assessed with multivariate logistic regression, results shown as odds ratio (OR) and 95% CI. Statistics controlled for the survey sampling design. Results: UC placement varied from 22 to 33 per 1000 adult ED visits, peaking in 2003. Overall, 1.6% (CI 1.5– 1.7%) of discharged patients and 8.5% (CI 7.9–9.1%) of admitted patients received UCs. More than half of EDplaced UCs were for potentially avoidable diagnoses. There was not a significant change in UC placement (8.7% vs. 8.0%, p = 0.4) or PAUC placement (4.8% vs. 4.7%, p = 0.4) in admitted ED patients after CDC guideline publication. Predictors of UCs in admitted patients included increasing age (‡80 y vs. 18–59 y, OR 3.1, CI 2.7–3.6), female sex (OR 1.3, CI 1.2–1.4), race (Hispanic vs. white, 0.8, CI 0.6–0.9), arrival by ambulance (OR 2.5, CI 2.3–2.8), increasing urgency (‡2 h vs. immediate, OR 0.8, CI 0.6–1.1), and longer ED visits (‡4 h vs. <2 h, OR 1.4, CI 1.2–1.7); facility characteristics included region (South vs. Northeast, OR 2.2, CI 1.9–2.5), teaching ED (OR 1.3, CI 1.2–1.5), and urban location (OR 1.3, CI 1.1– 1.5). The most common reasons for visit and diagnosis categories among patients receiving UCs are shown in the table. Conclusion: The high rates of PAUC suggest a potential for reduction of UCs in admitted ED patients - a proven strategy to reduce CAUTI. Publication of the CDC CAUTI guideline in 2007 did not affect ED use of UCs. Table - Abstract 128: Most Common Reasons for Visit and Discharge Diagnoses Resulting in UC Placement In U.S. EDs Reason for visit Stomach and abdominal pain, cramps and spasms Other urinary dysfunctions Shortness of breath Chest pain and related symptoms (not referable to a specific body system) Discharge diagnosis (Clinical Classification Software groupings) Genitourinary symptoms and ill-defined conditions Urinary tract infections Abdominal pain Congestive heart failure; non-hypertensive Average # receiving UC per year % with RFV who received UC 256,800 4.2% 177,600 60.1% 144,500 78,000 5.8% 1.4% Average # receiving UC per year % with diagnosis who received UC 236,600 44.6% 167,900 119,400 96,700 8.8% 3.5% 14.0% 1 University of Illinois College of Medicine at Peoria, Peoria, IL; 2University of Arkansas for Medical Sciences, Little Rock, AR; 3University of Alabama Medical Center and Children’s Health Center, Birmingham, AL Background: Motor vehicle crashes (MVC) are the leading cause of childhood fatality, making child passenger safety restraint (CPS) usage a public health priority. While MVCs in rural environments are associated with increased injuries and fatalities, no literature currently separates and focuses CPS misuse by geographic location. Objectives: We hypothesize that proper CPS usage will be lower in a rural population as compared to a similar matched urban population. Methods: A multisite (Alabama, Arkansas, Illinois), observational, case-control study was performed using rural (economically and population controlled) CPS unscheduled check data collected during the Strike Out Child Passenger Injury Trial and unscheduled urban CPS check data matched by age, site, and year. All CPS checks were performed using nationally certified CPS technicians who utilized the best practice standards of the American Academy of Pediatrics and collected subject demographics, misuse patterns, and interventions using identical definitions. Misuse patterns were defined using National Highway Traffic Safety Administration (NHTSA) standardized criteria and examined by state, location, age, and type. Pearson chi-square and Fisher’s test were conducted using SAS 9.2. The two-tailed p values were calculated and p = 0.05 was considered for statistical significance. Results: Four-hundred eighty-four CPS checks (242 rural and 242 urban) involving 603 total children (<1 years 46 (8%), 1–3 years 215 (36%), 4–8 years 321 (53%), ‡9 years 21 (3%)) from three states (AL 43 (7%), AR 442 (73%), IL 118 (20%)) were examined; of which 86% had at least one documented CPS misuse (arrived unrestrained 6%, improper direction 6%, harness incorrect 66%, LATCH incorrect 52%, airbag location 11%, seatbelt loose 49%, incorrect tether 63%). CPS misuse did not vary by age category (p = 0.31) but did by state (p = 0.001). Rural CPS misuse was more common than urban CPS misuse (91.5% vs. 81.2%; p = 0.0002, OR = 2.5, 95% CI = 1.5–4.1). Conclusion: In this multisite study, rural location was associated with higher CPS misuse. CPS education and resources that target rural populations specifically appear to be justified. 130 129 Child Passenger Restraint Misuse in Rural vs. Urban Children: A Multisite CaseControl Study John W. Hafner1, Stephanie Kok1, Huaping Wang1, Dale Wren1, Kathy Baker1, Mary E. Aitken2, Byron L. Anderson2, Beverly K. Miller2, Kathy W. Monroe3 A National Estimate of Injuries to Elders from Falls in the Home that Presented to U.S. Emergency Departments in 2009 Uwe Stolz, Alejandro Gonzalez, Sief Naser, Katy Orr University of Arizona, Tucson, AZ Background: An estimated 40% of community dwelling adults 65 or older fall at least once each year in the US, and 1 in 40 falls result in a hospitalization. In fact, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 25% of all elderly hospital admissions are directly related to falls. Falls are an important preventable cause of morbidity and mortality in the elderly and they pose a tremendous financial burden on our health care system and society in general. Objectives: To investigate the basic epidemiology of home falls that occur in those 65 years of age or older with injuries that present to US emergency departments (EDs). We hypothesized that the severity of injuries would be significantly related to age and the consumer product associated with each injury. Methods: We used the National Electronic Injury Surveillance System (NEISS), a national probability sample of US EDs that collects patient data for all consumer product-related injuries. Inclusion criteria were all injuries that occurred in a residential setting in 2009 to adults 65 or older. Exclusion criteria were all injuries not related to or caused by a home fall. Fall-related injuries were identified by key word searches. National estimates and 95% confidence intervals (CIs) for frequencies and proportions were calculated by accounting for the complex survey data. Results: There were 22,134 observations corresponding to a national estimate of 938,577 (95% CI 769,981– 1,107,173) residential fall-related injuries in adults 65 or older that presented to US EDs in 2009. Females accounted for 65% (95% CI 63–66) of injuries. A total of 30% (95% CI 27–33) were severe injuries requiring hospitalization or further observation. Frequency of severe injuries was 36.3% in those 85 or older, 31.2% for those 75–84, and 22.3% for those 65–74 (p < 0.001). A total of 92% (95% CI 89–94) of falls were mechanical falls while 8% (95% CI 6–11) had a physiological cause (syncope, alcohol, etc.). The table shows the top eight homerelated consumer products associated with home falls. After excluding floors and walls, stairs were the most common injury-related home product (10.6%, see table). Toilets had the highest proportion of severe injuries (40%). Conclusion: Nearly 1 million Americans 65 and older sustained a home fall-related injury in 2009 requiring treatment at a US ED. Further efforts to study and prevent home falls in the elderly are needed. Table - Abstract 130: Top Eight Products Related to Home Falls (Excluding Floors & Walls) Consumer Products Total Stairs Bed Chair Bathtubs/ Showers Tables/Night Stands Toilet Rugs Ladders Actual Observations (n) National Estimate (N) Weighted Percentage (95% CI) 22,134 2,983 2,270 1,072 1,045 938,577 120,444 93,928 44,677 44,503 527 22,845 2.1 (1.9–2.4) 558 539 435 22,531 21467 17,604 2.1 (1.9–2.4) 2.0 (1.6–2.4) 1.6 (1.3–1.8) 100.0 10.6 9.0 4.5 4.0 (9.3–12.0) (8.2–9.8) (4.1–4.9) (3.5–4.6) • 131 www.aemj.org S73 Prevalence of Bicycle Helmet Use By Users of Public Bicycle Sharing Programs Christopher M. Fischer1, Czarina E. Sanchez1, Mark Pittman2, David Milzman2, Kathryn A. Volz1, Han Huang2, Shiva Gautam1, Leon D. Sanchez1 1 Beth Israel Deaconess Medical Center/Harvard Medical School, Boston, MA; 2Georgetown University/Washington Hospital Center/ MedStar, Washington, DC Background: Public bicycle sharing (bikeshare) programs are becoming increasingly common in the US and around the world. These programs make bicycles easily accessible for hourly rental to the public. There are currently 15 active bikeshare programs in cities in the US, and more than 30 programs are being developed in cities including New York and Chicago. Despite the importance of helmet use, bikeshare programs do not provide the opportunity to purchase or rent helmets. While the programs encourage helmet use, no helmets are provided at the rental kiosks. Objectives: We sought to describe the prevalence of helmet use among adult users of bikeshare programs and users of personal bicycles in two cities with recently introduced bicycle sharing programs (Boston, MA and Washington, DC). Methods: We performed a prospective observational study of bicyclists in Boston, MA and Washington, DC. Trained observers collected data during various times of the day and days of the week. Observers recorded the sex of the bicycle operator, type of bicycle, and helmet use. All bicycles that passed a single stationary location in any direction for a period of between 30 and 90 minutes were recorded. Data are presented as frequencies of helmet use by sex, type of bicycle (bikeshare or personal), time of the week (weekday or weekend), and city. Logistic regression was used to estimate the odds ratio for helmet use controlling for type of bicycle, sex, day of week, and city. Results: There were 43 observation periods in two cities at 36 locations. 3,073 bicyclists were observed. There were 562 (18.2%) bicylists riding bikeshare bicycles. Overall helmet use was 45.5%, although helmet use varied significantly with sex, day of use, and type of bicycle (see figure). Bikeshare users were helmeted at a lower rate compared to users of personal bicycles (19.2% vs 51.4%). Logistic regression, controlling for type of bicycle, sex, day of week, and city demonstrate that bikeshare users had higher odds of riding unhelmeted (OR 4.34, 95% CI 3.47–5.50). Women had lower odds of riding unhelmeted (OR 0.62, 0.52–0.73), while weekend riders were more likely to ride unhelmeted (OR 1.32, 1.12–1.55). Conclusion: Use of bicycle helmets by users of public bikeshare programs is low. As these programs become more popular and prevalent, efforts to increase helmet use among users should increase. S74 132 2012 SAEM ANNUAL MEETING ABSTRACTS Keeping Infants Safe and Secure for Schools (KISS): A School-Based Abusive Head Trauma Prevention Program Faisal Mawri1, Elaine Pomeranz1, Brain Nolan2, Rachel Stanley1 1 University of Michigan Health System, Ann Arbor, MI; 2Hurley Medical Center, Flint, MI Background: Abusive head trauma (AHT) represents one of the most severe forms of traumatic brain injury (TBI) among abused infants with 30% mortality. Young adult males account for 75% of the perpetrators. Most AHT prevention programs are hospital-based and reach a predominantly female audience. There are no published reports of school-based AHT prevention programs to date. Objectives: 1. To determine whether a high schoolbased AHT educational program will improve students’ knowledge of AHT and parenting skills. 2. To evaluate the feasibility and acceptability of a school-based AHT prevention program. Methods: This program was based on an inexpensive commercially available program developed by the National Center on Shaken Baby Syndrome. The program was modified to include a 60-minute interactive presentation that teaches teenagers about AHT, parenting skills, and caring for inconsolable crying infants. The program was administered in three high schools in Flint, Michigan during spring 2011. Student’s knowledge was evaluated with a 17-item written test administered pre-intervention, post-intervention, and two months after program completion. Program feasibility and acceptability were evaluated through interviews and surveys with Flint area school social workers, parent educators, teachers, and administrators. Results: In all, 342 high school students (40% male) participated. Of these, 317 (92.7%) completed the pretest and post-test with 171 (50%) completing the twomonth follow-up test. The mean pre-intervention, postintervention, and two-month follow-up scores were 53%, 87%, and 90% respectively. From pre-test to posttest, mean score improved 34%, p < 0.001. This improvement was even more profound in young males, whose mean post-test score improved by 38%, p < 0.001. Of the 69 participating social workers, parent educators, teachers, and administrators, 97% ranked the program as feasible and acceptable. Conclusion: Students participating in our program showed an improvement in knowledge of AHT and parenting skills which was retained after two months. Teachers, social workers, parent educators, and school administrators supported the program. This local pilot program has the potential to be implemented on a larger scale in Michigan with the ultimate goal of reducing AHT amongst infants. 133 Will Patients Exaggerate Their Symptoms To Increase The Likelihood Of A Cash Settlement? Thomas Gilmore1, J. Matthew Fields1, Ian Storch2, Wayne Bond Lau1, Gerald O’Malley1 1 Thomas Jefferson University Hospital, Philadelphia, PA; 2Thomas Jefferson Medical College, Philadelphia, PA Background: Fear of litigation has been shown to affect physician practice patterns, and subsequently influence patient care. The likelihood of medical malpractice litigation has previously been linked with patient and provider characteristics. One common concern is that a patient may exaggerate symptoms in order to obtain monetary payouts; however, this has never been studied. Objectives: We hypothesize that patients are willing to exaggerate injuries for cash settlements and that there are predictive patient characteristics including age, sex, income, education level, and previous litigation. Methods: This prospective cross-sectional study spanned June 1 to December 1, 2011 in a Philadelphian urban tertiary care center. Any patient medically stable enough to fill out a survey during study investigator availability was included. Two closed-ended paper surveys were administered over the research period. Standard descriptive statistics were utilized to report incidence of: patients who desired to file a lawsuit, patients previously having filed lawsuits, and patients willing to exaggerate the truth in a lawsuit for a cash ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 settlement. Chi-square analysis was performed to determine the relationship between patient characteristics and willingness to exaggerate injuries for a cash settlement. Results: Of 126 surveys, 11 were excluded due to incomplete data, leaving 115 for analysis. The mean age was 39 with a standard deviation of 16, and 40% were male. The incidence of patients who had the desire to sue at the time of treatment was 9%. The incidence of patients who had filed a lawsuit in the past was 35%. Of those patients, 26% had filed multiple lawsuits. Fifteen percent [95% CI 9–23%] of all patients were willing to exaggerate injuries for cash settlement. Sex and income were found to be statistically significant predictors of willingness to exaggerate symptoms: 22% of females vs. 4% of males were willing to exaggerate (p = 0.01), and 20% of people with income less than $100,000/yr vs. 0% of those with income over $100,000/ yr were willing to exaggerate (p = 0.03). Conclusion: Patients at a Philadelphian urban tertiary center admit to willingness to exaggerate symptoms for a cash settlement. Willingness to exaggerate symptoms is associated with female sex and lower income. 134 Expert Consensus Meeting Recommendations on Community Consultation for Emergency Research Conducted with an Exception from Informed Consent Lynne D. Richardson1, Ilene Wilets1, Meg Smirnoff1, Rosamond Rhodes1, Cindy Clesca1, Patria Gerardo1, Katherine Lamond2, Robert Lowe3, Jill Baren2 1 Mount Sinai School of Medicine, New York, NY; 2University of Pennsylvania, Philadelphia, PA; 3Oregon Health & Science University, Portland, OR Background: Federal regulations allow an exception from informed consent (EFIC) for emergency research on life-threatening conditions but require additional protections, such as community consultation, prior to granting such an exception. Objectives: To develop consensus on effective methods of conducting and evaluating community consultation for EFIC research. Methods: An expert meeting was convened to develop recommendations about best practices for community consultation. An invited panel of experienced EFIC researchers including representation from federally funded emergency care research networks, IRB members, and community representatives met to review the experiences of emergency care networks in conducting EFIC trials and the findings of the Community VOICES study in order to identify areas of consensus. Results: Twenty experts participated and a total of eleven recommendations were developed. All participants agreed community consultation efforts should consist of two-way open-ended communication between the PI or senior study staff and community members utilizing several effective modalities such as focus groups, meetings (group or individual) with community leaders/representatives, or in-person interviews conducted by a member of the research team. Expert consensus did not endorse • www.aemj.org S75 some frequently used modalities such as random-digit dialing telephone surveys, written questionnaires, webbased surveys, and social networks (e.g. Facebook, Twitter) as meeting the federal requirements and recommended these not be used in isolation to conduct community consultation. Participants endorsed methodology developed by the Community VOICES study showing that five domains (feasibility, composition of participants, participant perception, investigator perception, and quality of communication) are essential in evaluating the effectiveness of community consultation efforts. Conclusion: This expert meeting promulgated recommendations regarding best practices for conducting and evaluating community consultation for EFIC research. 135 Is A History Of Psychiatric Disease Or Substance Abuse Associated With An Increased Incidence Of Syncope Of Unknown Etiology? Zev Wiener1, Nathan I. Shapiro2, Shamai A. Grossman2 1 Harvard Medical School, Boston, MA; 2 Harvard Medical School, Beth Israel Deaconess Medical Center, Boston, MA Background: Current data suggest that as many as 50% of patients presenting to the ED with syncope leave the hospital without a defined etiology. Prior studies have suggested a prevalence of psychiatric disease as high as 26% in patients with syncope of unknown etiology. Objectives: To determine whether psychiatric disease and substance abuse are associated with an increased incidence of syncope of unknown etiology. Methods: Prospective, observational, cohort study of consecutive ED patients ‡18 presenting with syncope was conducted between 6/03 and7/06. Patients were queried in the ED and charts reviewed about a history of psychiatric disease, use of psychiatric medication, substance abuse, and duration. Data were analyzed using SAS with chi-square and Fisher’s exact tests. Results: We enrolled 519 patients who presented to the ED after syncope, 159 of whom did not have an identifiable etiology for their syncopal event. 36.5% of those without an identifiable etiology were male. 166 (32%) patients had a history of or current psychiatric disease (42% male), and 55 patients (11%) had a history of or current substance abuse (60% male). Among males with psychiatric disease, 39% had an unknown etiology of their syncopal event, compared to 22% of males without psychiatric disease (p = 0.009). Similarly, among all males with a history of substance abuse, 45% had an unknown etiology, as compared to 24% of males without a history of substance abuse (p = 0.01). A similar trend was not identified in elderly females with psychiatric disease (p = 0.96) or substance abuse (p = 0.19). However, syncope of unknown etiology was more common among both men and women under age 65 with a history of substance abuse (47%) compared to those without a history of substance abuse (27%; p = 0.01). Conclusion: Our results suggest that psychiatric disease and substance abuse are associated with increased S76 2012 SAEM ANNUAL MEETING ABSTRACTS incidence of syncope of unknown etiology. Patients evaluated in the ED or even hospitalized with syncope of unknown etiology may benefit from psychiatric screening and possibly detoxification referral. This is particularly true in men. (Originally submitted as a ‘‘late-breaker.’’) 136 Scope of Practice and Autonomy of Physician Assistants in Rural vs. Urban Emergency Departments Brandon T. Sawyer, Adit A. Ginde Department of Emergency Medicine, University of Colorado School of Medicine, Aurora, CO Background: Physician assistants (PAs) are being utilized in greater capacity in both rural and urban EDs to supplement the EM workforce shortages, improve efficiency of emergency physicians, and reduce the cost of emergency care. Objectives: We sought to compare the scope of practice and autonomy of EM PAs practicing in rural vs. urban EDs, and hypothesized that rural PAs would have a broader scope of practice and higher reported autonomy, while receiving less direct supervision. Methods: Using the American Academy of Physician Assistants Masterfile, we surveyed a random sample of 400 U.S. PAs who self-identified EM as their specialty. We classified location as rural or urban by zip code-based Rural-Urban Commuting Area codes, and oversampled 200 rural PAs to ensure adequate rural representation. We asked all PAs about conditions managed, procedures performed, and physician supervision, comparing groups using chi-square test. Results: To date, 223 (56%) of 400 PAs in 44 states responded, of whom 112 rural and 85 urban PAs currently practice in EDs. In the past year, rural PAs more frequently managed acute coronary syndrome (94% vs 84%); cardiac arrest (64% vs 43%); stroke (85% vs 71%); anaphylaxis (80% vs 67%); multi-system trauma (82% vs 66%); active labor (40% vs 25%); and critically ill child (83% vs 61%; all p < 0.05). While rural PAs were less likely to have performed bedside ultrasound (49% vs 61%) and lumbar puncture (44% vs 60%), they were more likely to have performed bag-valve-mask ventilation (74% vs 52%); intubation (64% vs 42%); needle thoracostomy (20% vs 6%); tube thoracostomy (42% vs 27%); and thoracotomy (5% vs 0%) in the past year (all p < 0.05). Rural PAs more often reported never having a physician present in the ED compared to urban PAs (33% vs 2%), and less often reported always having a physician present (57% vs 93%; p < 0.001). When no physician was present in the ED, rural PAs were also less likely to have a physician onsite in less than 10 minutes (33% vs 71%; p = 0.047). Additionally, rural PAs were less likely to have at least one supervising physician be EM-board certified (67% vs 99%; p < 0.001). Conclusion: Rural PAs reported a broader scope of practice, more autonomy, and less access to physician supervision than urban PAs. Adequate training and supervision of all EM PAs, particularly those in rural EDs, should be considered a high priority to optimize the quality of emergency care. 137 Video Education Intervention in the Emergency Department Nancy Stevens, Amy L. Drendel, Steven J. Weisman Medical College of Wisconsin, Milwaukee, WI Background: After discharge from an emergency department (ED), pain management often challenges parents, who significantly under-treat their children’s pain. Rapid patient turnover and anxiety make education about home pain treatment difficult in the ED. Video education standardizes information and circumvents insufficient time and literacy. Objectives: To evaluate the effectiveness of a 6-minute instructional video for parents that targets common misconceptions about home pain management. Methods: We conducted a randomized, double-blinded clinical trial of parents of children ages 1–18 years who presented with a painful condition, were evaluated, and discharged home in June and July 2011. Parents were randomized to a pain management video or an injury prevention control video. Primary outcome was the proportion of parents who gave pain medication at home. These data were recorded in a home pain diary and analyzed using a chi-square test. Parents’ knowledge about pain treatment was tested before, immediately following, and 2 days after intervention. McNemar’s test statistic determined odds that knowledge correlated with the intervention group. Results: 100 parents were enrolled: 59 watched the pain education video, and 41 the control video. 72.9% completed follow up, providing information about home pain education use. Significantly more parents provided at least one dose of pain medication to their children after watching the educational video: 96% vs. 80% (difference 16%, 95% CI 7.8%, 31.3%). The odds the parent had correct knowledge about pain treatment significantly improved immediately following the educational video for knowledge about pain scores (p = 0.04), the effect of pain on function (p < 0.01), and pain medication misconceptions (p < 0.01). These significant differences in knowledge remained 3 days after the video intervention. Conclusion: The educational video about home pain treatment viewed by parents significantly increased the proportion of children receiving pain medication at home and significantly improved knowledge about at-home pain management. Videos are an efficient tool to provide medical advice to parents that improves outcomes for children. 138 Secondary Shockable Rhythms: Prognosis in Out-of-Hospital Cardiac Arrests with Initial Asystole or Pulseless Electrical Activity and Subsequent Shockable Rhythms Andrew J. Thomas, Mohamud R. Daya, Craig D. Newgard, Dana M. Zive, Rongwei Fu Oregon Health & Science University, Portland, OR Background: Non-shockable cardiac arrest rhythms (pulseless electrical activity and asystole) represent an ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 increasing proportion of reported cases of out-of-hospital cardiac arrest (OHCA). The prognostic significance of conversion from non-shockable to shockable rhythms during the course of resuscitation has been debated in published literature. Objectives: To evaluate whether OHCA survival with an initially non-shockable cardiac arrest rhythm is improved with subsequent conversion to a shockable rhythm. Methods: This study is a secondary analysis of the prospectively collected Epistry: Cardiac Arrest, an epidemiologic registry organized by the Resuscitation Outcomes Consortium (ROC). Epistry collects data from all out-of-hospital cardiac arrests at ten North American sites followed through hospital discharge. The sample for this analysis includes OHCA from six US and two Canadian sites from December 1, 2005 through May 31, 2007. The investigational cohort includes all EMS-treated adult (18 and older) cardiac arrest patients who presented with non-shockable cardiac arrest rhythms and were treated by EMS personnel. We compared survival to hospital discharge between patients who did versus those who did not develop a subsequent shockable rhythm which we defined as receiving subsequent defibrillation (presumed conversion to VF/ VT). Missing data were handled using multiple imputation. Multivariable logistic regression was used to adjust for potentially confounding variables: age, sex, public location, witnessed status, bystander resuscitation, EMS response interval, and ROC site. Results: A total of 6,593 adult cardiac arrest cases presented in non-shockable rhythms, were treated by EMS, and had known survival status. Survival to discharge in patients who converted to a shockable rhythm during out-of-hospital resuscitation was 2.7% compared to 2.7% for those who did not convert to a shockable rhythm, a statistically insignificant difference (chi-square, p = 0.90). After accounting for known confounders, the adjusted odds of survival for conversion to a shockable rhythm was not associated with improved survival (OR 1.03, 95% CI: 0.71–1.50). Conclusion: Out-of-hospital cardiac arrest patients presenting in PEA/asystole had no better or worse survival to hospital discharge with conversion to a shockable rhythm during EMS resuscitation efforts. 139 Racial Disparities in Stress Test Utilization in a Chest Pain Unit Anthony Napoli1, Esther Choo1, Bethany Desroches2, Jessica Dai1 1 Warren Alpert Medical School of Brown University, Providence, RI; 2Downstate College of Medicine, New York, NY Background: Epidemiologic studies have demonstrated racial disparities in the workup of emergency department (ED) patients with chest pain and the referral of admitted patients for subsequent catheterization and bypass surgery. Objectives: To determine if similar disparities occur in the stress test utilization of low-risk chest pain patients admitted to ED chest pain units (CPU). • www.aemj.org S77 Methods: This was a prospective, observational study of consecutive admitted CPU patients in a large-volume academic urban ED. Cardiology attendings round on all patients and stress test utilization is driven by their recommendation. Eligibility criteria include: age>18, AHA low/intermediate risk, nondynamic ECGs, and normal initial Troponin I. Patients >75 and with a history of CAD or co-existing active medical problem were excluded. Based on prior studies and our estimated CPU census and demographic distribution, we estimated a sample size of 2,242 patients in order to detect a difference in stress utilization of 7% (2-tailed, a = 0.05, b = 0.8). We calculated a TIMI risk prediction score and a Diamond & Forrester (D&F) CAD likelihood score on each patient. T-tests were used for univariate comparisons of demographics, cardiac comorbidities, and risk scores. Logistic regression was used to estimate odds ratios (ORs) for receiving testing based on race, controlling for insurance and either TIMI or D&F score. Results: Over 18 months, 2,451 patients were enrolled. Mean age was 53 ± 12, and 54% (95% CI 52–56) were female. Sixty percent (95% CI 58–62) were Caucasian, 12% (95% CI 10–13) African American, and 24% (95% CI 23–26) Hispanic. Mean TIMI and D&F scores were 0.5 (95% CI 0.5–0.6) and 38% (95% CI 37–39). The overall stress testing rate was 52% (95% CI 50–54). After controlling for insurance status and TIMI or D&F scores, African American patients had significantly decreased odds of stress testing (ORTIMI 0.67 (95% CI 0.52–0.88), ORD&F 0.68 (95% CI 0.51–0.89)). Hispanics had significantly decreased odds of stress testing in the model controlling for D&F (ORD&F 0.78 (95% CI 0.63–0.98)). Conclusion: This study confirms that disparities in the workup of African American patients in the CPU are similar to those found in the general ED and the outpatient setting. Further investigation into the specific provider or patient level factors contributing to this bias is necessary. 140 The Usefulness of the 3-Minute Walk Test in Predicting Adverse Outcomes in ED Patients with Heart Failure and COPD Ian G. Stiell1, Catherine M. Clement2, Lisa A. Calder1, Brian H. Rowe3, Jeffrey J. Perry1, Robert J. Brison4, Bjug Borgundvaag5, Shawn D. Aaron1, Eddy Lang6, Alan J. Forster1, George A. Wells7 1 University of Ottawa, Ottawa, ON, Canada; 2 Ottawa Hospital Research Institute, Ottawa, ON, Canada; 3University of Alberta, Edmonton, AB, Canada; 4Queen’s University, Kingston, ON, Canada; 5University of Toronto, Toronto, ON, Canada; 6University of Calgary, Calgary, AB, Canada; 7University of Ottawa Heart Institute, Ottawa, ON, Canada Background: ED physicians frequently treat and make disposition decisions for patients with acute exacerbations of heart failure (HF) and COPD. Objectives: We sought to evaluate the usefulness of a unique, structured 3-minute walk test in predicting the S78 2012 SAEM ANNUAL MEETING ABSTRACTS risk for serious adverse events (SAE) amongst HF and COPD patients. Methods: We conducted a prospective cohort study in six large, academic EDs and enrolled 1,504 adult patients who presented with exacerbations of HF or COPD. After treatment, each patient underwent a 3-minute walk test, supervised by registered nurses or respiratory therapists, who monitored heart rate and oxygen saturation and evaluated the Borg score. Patients walked at their own pace around the ED on room air or home oxygen levels, without physical support from staff. We evaluated patients for multiple clinical and routine laboratory findings. Both admitted and discharged patients were followed for SAEs, defined as death, intubation, admission to a monitored unit, myocardial infarction, or relapse back to the ED requiring admission within 14 days. We evaluated both univariate and multivariate associations of the walk test components with SAE. Results: The characteristics, respectively, of the 559 HF and 945 COPD patients were: mean age 76.0, 72.6; male 56.4%, 51.6%; too ill to start walk test 13.2%, 15.3%; unable to complete walk test 15.4%, 21.5%. Outcomes for HF and COPD were SAE 11.6%, 7.8%; death 2.3%, 1.0%. We found univariate associations with SAE for these walk test components: too ill to walk (both HF, COPD P < 0.0001); highest heart rate ‡110 (HF P = 0.02, COPD P = 0.10); lowest SaO2 < 88% (HF P = 0.42, COPD P = 0.63); Borg score ‡5 (HF P = 0.47, COPD P = 0.52); walk test duration £ 1 minute (HF P = 0.07. COPD P = 0.22). After adjustment for multiple clinical covariates with logistic regression analyses, we found ‘‘walk test heart rate ‡110’’ had an odds ratio of 1.9 for HF patients and ‘‘too ill to start the walk test’’ had an odds ratio of 3.5 for COPD patients. Conclusion: We found the 3-minute walk test to be easy to administer in the ED and that maximum heart rate and inability to start the test were highly associated with adverse events in patients with exacerbations of HF and COPD, respectively. We suggest that the 3-minute walk test be routinely incorporated into the assessment of HF and COPD patients in order to estimate risk of poor outcomes. 141 The Effect of Prone Maximal Restraint (PMR, aka ‘‘Hog-Tie’’) Position on Cardiac Output and Other Hemodynamic Measurements Davut J. Savaser, Colleen Campbell, Theodore C. Chan, Virag Shah, Christian Sloane, Allan V. Hansen, Edward M. Castillo, Gary M. Vilke UCSD, San Diego, CA Background: The prone maximal restraint (PMR), hogtie or hobble position has been used by law enforcement and emergency care personnel to restrain the acutely combative or agitated patient. The position places the subject prone with wrists handcuffed behind the back and secured to the ankles. Some have argued that PMR, as well as the weight force required to place an individual into the position, can negatively affect cardiac function. Objectives: To measure the effect of PMR with and without weight force on measures of cardiac function including vital signs, oxygenation, stroke volume (SV), cardiac output (CO), and left ventricular outflow tract diameter (LVOTD). Methods: We conducted a randomized prospective cross-over study of healthy volunteers placed in five different body positions: supine, prone, PMR, PMR with 50 lbs added to the subject’s back (PMR50), and PMR with 100 lbs added to the subject’s back (PMR100) for 3 minutes. Data were collected on subject vital signs and echocardiographic measurement of SV, CO, and LVOTD, measured by credentialed ED faculty sonographers. Anthropomorphic measurements of height, weight, arm span, chest circumference, and BMI were also collected. Data were analyzed using repeated measures ANOVA to evaluate changes with each variable with respective positioning. Results: 25 male subjects were enrolled in the study, ages ranging from 22 to 43 years. Cardiac output did change from the supine to prone position, decreasing on average by 0.61 L/min (p = 0.013, 95% CI [0.142, 1.086]). However, there was no significant change in CO when placing the patient in the PMR position ()0.11 L/min, p = 0.489, 95% CI [)0.43, 0.21]), PMR50 position (+0.19 L/ min, p = 0.148, 95% CI [)0.07, 0.46]), or the PMR100 position (+0.14 L/min, p = 0.956, 95% CI [)0.29, 0.27]) as compared with the prone position. Also, mean HR (70.7, 68.4, 72.1, 71.8, 75.8 bpm) and mean arterial pressure (88.1, 86.0, 88.9, 84.7, 95.1 mmHg) for each respective position did not demonstrate any hemodynamic compromise. Conclusion: While there was a difference from supine position, there was no difference in CO between prone and any PMR position with and without weight force. Moreover, there was no evidence of hemodynamic compromise in any of these positions. 142 The Associations Between Numeracy and Health Literacy and 30-Day Recidivism in Adult Emergency Department Patients Suspected of Having Acute Heart Failure Candace McNaughton, Sean Colllins, Cathy Jenkins, Patrick Arbogast, Karen Miller, Sunil Kripalani, Russell Rothman, Allen Naftilan, Robert Dittus, Alan Storrow Vanderbilt University, Nashville, TN Background: Heart failure is a common, costly condition. While prior work suggests a relationship between literacy and clinical outcomes, no large prospective studies have evaluated the relationships between literacy or numeracy and 30-day recidivism. Objectives: Evaluate the association between numeracy and health literacy and 30-day recidivism in ED patients with signs and symptoms of acute heart failure (AHF). Methods: Convenience sample of a prospective cohort of adult ED patients at three hospitals who presented with signs and symptoms of AHF between 1/2008 and 12/2011. Patients meeting modified Framingham Criteria were consented for enrollment within three hours of ED evaluation. Research assistants administered subjective measures of numeracy and health literacy. Thirty-day ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 follow-up information was obtained by phone interview and confirmed with chart review. The primary outcome was 30-day recidivism to the ED or hospital; secondary outcomes evaluated components of the primary outcome individually, and subgroup analysis evaluated the relationship between numeracy and literacy and 30-day recidivism in patients with and without a clinical diagnosis of AHF. Adjustment was made for age, sex, race, insurance status, diabetes, renal disease, abnormal hemoglobin, low ejection fraction, number of days at risk for recidivism, and hospital site. Results: Of the 1823 patients enrolled in the cohort, 988 took the subjective numeracy and/or the subjective health literacy test. Thirty patients were excluded because they died during the index hospitalization or were admitted for the entire follow-up period, leaving 958 patients for analysis. Mean age was 61 years, and 48% were female; additional clinical characteristics are in Table 1. Results of adjusted analysis are found in Table 2. Lower numeracy was associated with increased 30-day recidivism (OR 1.02, 95%CI 1.00–1.03 per point change on a 43-point scale); the relationship for literacy was not significant (OR 1.04, 95% CI 1.00–1.08 per point change on the 13-point health literacy scale). For patients with AHF, lower health literacy and numeracy were associated with increased odds of 30-day recidivism. Conclusion: Lower numeracy and health literacy are associated with higher risk of 30-day ED and hospital recidivism for ED patients with AHF. Table 1 - Abstract 142: Clinical Characteristics Diabetes, No. (%) Chronic kidney disease, No. (%) Hgb <13 or ‡17, No. (%) EF<30%, No. (%) £High school (9–12), No. (%) Subjective literacy, median (IQR) Low subjective literacy, No. (%) Subjective numeracy, median (IQR) Low subjective numeracy, No. (%) Numeracy/ Literacy Cohort (n = 958) Total Cohort (n = 1823) 391 205 571 193 625 13 369 31 339 745 380 1095 345 1200 n/a n/a n/a n/a (41) (21) (60) (20) (66) (9–15) (39) (23–38) (37) (41) (21) (60) (20) (68) • www.aemj.org 143 S79 The Influence of Clinical Context on Patients’ Code Status Preferences John E. Jesus1, Matthew B. Allen2, Glen E. Michael3, Michael W. Donnino4, Shamai A. Grossman4, Caleb P. Hale4, Anthony C. Breu4, Alexander Bracey4, Jennifer L. O’Connor4, Jonathan Fisher4 1 Christiana Care Health Systems, Newark, DE; 2 Brigham and Women’s Hospital, Boston, MA; 3 University of Virginia University Hospital, Charlottesville, VA; 4Beth Israel Deaconess Medical Center, Boston, MA Background: Many patients have preferences regarding the use of cardiopulmonary resuscitation (CPR) and intubation that go undocumented and incompletely understood, despite the presence of do-not-resuscitate (DNR)/do-not-intubate (DNI) orders. Objectives: To assess patients’ awareness and understanding of their code status as it applies to the hypothetical scenarios. Methods: A prospective survey of patients with documented DNR/DNI code status was conducted from October 2010 to October 2011. Patients were surveyed by a research assistant starting with a validated cognitive assessment. The researcher then administered the survey consisting of four scenarios of varying degrees of severity and reversibility (angioedema, pneumonia, severe stroke, and cardiac arrest). Each patient was asked whether he or she would agree to specific treatments in specific situations, and about whom he or she would want to make health care decisions (previous declaration, family, or health care provider). Descriptive statistics including SD and 95% CI were calculated using Microsoft Excel and SPSS 17. Results: 110 patients were identified and screened; 3 patients failed the cognitive screen, 5 patients had code statuses inconsistent with DNR/DNI, and 2 patients were unable to complete the survey. Patients had a mean age of 78 (SD 14). 2% (CI 0–5) of patients were not aware of their documented code status. While 98% of patients knew they had a documented code status of DNR/DNI, 58% (CI 48–68) would want to be intubated in the face of a life-threatening but potentially reversible situation, and 20% (CI 12–28) of patients would want a trial of resuscitation including intubation and Table 2 - Abstract 142: Outcomes, Adjusted Logistic Regression Outcomes, OR (95% CI) Any 30-day recidivism, adjusted Subgroup: patients with clinical diagnosis of heart failure Adjusted Secondary Outcomes Any ED visit for AHF Any ED visit for complaint other than AHF Any unscheduled hospitalization for AHF Any unscheduled hospitalization for complaint other than AHF Literacy, Continuous P Low Literacy (SLS<12) 1.04 (1.00–1.08) 1.05 (1.00–1.10) 0.06 0.05 1.19 (0.89–1.6) 1.05 (1.00–1.10) 1.03 (0.96–1.11) 1.02 (0.98–1.07) 0.38 0.31 0.98 (0.91–1.07) 1.05 (1.00–1.11) P Low Numeracy (SNS<28) P 0.03 0.05 1.28 (0.95–1.75) 1.29 (0.90–1.85) 0.11 0.17 1.02 (0.99–1.05) 1.01 (0.99–1.02) 0.17 0.38 1.44 (0.82–2.53) 1.12 (0.80–1.56) 0.2 0.5 0.7 1.01 (0.98–1.04) 0.67 1.28 (0.72–2.28) 0.4 0.03 1.02 (1.00–1.04) 0.02 1.51 (1.04–2.18) 0.03 P Numeracy, Continuous 0.24 0.05 1.02 (1.00–1.03) 1.02 (1.00–1.04) 1.03 (0.96–1.11) 1.02 (0.98–1.07) 0.38 0.31 0.7 0.98 (0.91–1.07) 0.03 1.05 (1.00–1.11) S80 2012 SAEM ANNUAL MEETING ABSTRACTS Table - Abstract 143: Pt. desires intubation Family decides Doctor decides One week trial of resuscitation Long-term family care Opportunity for family to say goodbye Angioedema Pneumonia Severe Stroke Cardiac Arrest 58% 28% 5% 3% 52% 37% 30% 28% 31% 20% 13% 12% 39% 47% 24% 20% 5% 7% 2% 1% 35% 33% 33% 28% CPR in the setting of cardiac arrest. Across all scenarios, 32% (CI 28–37) would want to be kept alive long enough for family members to say good-bye. The willingness to be intubated decreased with the potential reversibility of the disease process (P < 0.001). See table. No demographic factors including education or religion predicted such discrepancies. Conclusion: Important discrepancies exist between patients’ code status and their actual end of life preferences. These discrepancies may lead to the denial of life-saving or life-prolonging care from patients. A better way of elucidating patient end of life preferences is needed. 144 End of Life Decision-Making for Patients Admitted Through the ED: Patient Demographics, Hospital Attributes, and Changes over Time Derek K. Richardson, Dana Zive, Craig D. Newgard Oregon Health & Science University, Portland, OR Background: Early studies suggested that racial, economic, and hospital-based factors influence the do not resuscitate (DNR) status of admitted patients, though it remains unknown how these factors apply to patients admitted through the ED and whether use is increasing over time. Objectives: To examine patient and hospital attributes associated with DNR orders placed within the first 24 hours among patients admitted through the ED and changes over time. Methods: This was a population-based, retrospective cross-sectional study of patients over 65 years old admitted to acute care hospitals in California between 2002 and 2010; the subset of patients admitted through the ED formed the primary sample. The primary variable (outcome) was placement of a DNR order within 24 hours of admission. We tested associations between early DNR orders, hospital characteristics, patient demographics, and year. For data analysis, we used descriptive statistics and multivariable logistic regression models with generalized estimating equations to account for clustering within hospitals. Results: 9,507,921 patients older than 65 were admitted to California hospitals over the 9-year period, of whom 10.8% had an early DNR order; 83% of early DNR orders were placed on patients admitted through the ED. Among patients over 65 admitted through the ED (n = 6,396,910), early DNR orders were less frequent at teaching hospitals (9.5% vs. 13.7%), for-profit hospitals (8.6% vs. 14.6% nonprofit), non-rural hospitals (12.0% vs. 26.2%), and large hospitals (15.0% vs. 11.1% smallest quartile hospitals) (all p < 0.0001). In regression modeling adjusted for clustering, the prior trends were reproduced. Additionally, decreased DNR frequency was associated with race (black OR 0.59, 95% CI 0.51–0.67; Asian OR 0.70, 95% CI 0.59– 0.82), ethnicity (Hispanic OR 0.61, 95% CI 0.55–0.68), sex (male OR 0.90, 95% CI 0.88–0.92), and MediCal insurance (OR 0.70, 95% CI 0.57–0.85). Over the last ten years, rates of early DNR use have steadily increased (figure, p < .0005). Conclusion: While statewide rates of early DNR use have increased over time among patients admitted through the ED, there is variable penetrance of this practice between hospital types and use is less common among some patient groups. These patterns may suggest barriers to end-of-life discussions or differing cultural or institutional beliefs. 145 Disparities in the Successful Enrollment of Study Subjects Involving Informed Consent in the Emergency Department Lea H. Becker, Marcus L. Martin, Chris A. Ghaemmaghami, Pamela A. Ross, Robert E. O’Connor University of Virginia Health System, Charlottesville, VA Background: It has been established that certain populations are often under-represented in clinical trials for many reasons. Objectives: The objective of this study was to investigate differences in consent rates between patients of different demographic groups who were invited to participate in minimal-risk clinical trials conducted in an academic emergency department. Methods: This descriptive study analyzed prospectively collected data of all adult patients who were identified as qualified participants in ongoing minimal risk clinical trials. These trials were selected for this review because they presented minimal factors known to be associated ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 with patient refusal (e.g. additional length of stay, travel, adverse drug reactions). Consented patients underwent one to three blood draws, telephone follow-up, no modification of treatment, and received no payment. Prior to being invited to participate, patients self-identified their race according to the 2010 US Census definitions. Age, sex, and race were recorded without any patient identifiers. The primary endpoint was whether approached patients consented or declined participation. Statistical analysis was performed using Fisher’s exact test. Results: Of the 627 patients approached for enrollment, 291 (42%) were women and 336 (58%) were men, and 488 (78%) were self-described as white. A total of 496 (79%) consented and 131 (21%) declined. The consent rate was 80% for women and 78% for men (p = 0.49 ns). The consent rate was 83% for whites and 67% for non-whites (p < 0.0001). The average ages of consenting vs. declining patients were 57.8 and 60.7, respectively (p = 0.92 ns). Within each race classification, there were no differences in rates of consent by age or sex. Conclusion: Significant demographic differences continue to exist in the rate of consent in emergency department based studies involving minimal risk and time commitment for study subjects. These enrollment disparities may severely limit study external validity. Although this review may show some improvement compared to previous studies, enrollment and consent strategies aimed at reducing demographic disparities are needed for the development and application of evidence-based emergency care. 146 Informed Consent for Computerized Tomography via Video Educational Module in the Emergency Department Lisa H. Merck, Michael Holdsworth, Joshua Keithley, Debra Houry, Laura A. Ward, Kimberly E. Applegate, Douglas W. LoweryNorth, Kate Heilpern Emory University, Atlanta, GA Background: Increasing rates of patient exposure to computerized tomography (CT) raise questions about appropriateness of utilization, as well as patient awareness of radiation exposure. Despite rapid increases in CT utilization and published risks, there is no national standard to employ informed consent prior to radiation exposure from diagnostic CT. Use of written informed consent for CT (ICCT) in our ED has increased patient understanding of the risks, benefits, and alternatives to CT imaging. Our team has developed an adjunct video educational module (VEM) to further educate ED patients about the CT procedure. Objectives: To assess patient knowledge and preferences regarding diagnostic radiation before and after viewing VEM. Methods: The VEM was based on ICCT currently utilized at our tertiary care ED (census 37,000 patients/ year). ICCT is written at an 8th grade reading level. This fall, VEM/ICCT materials were presented to a convenience sample of patients in the ED waiting room • www.aemj.org S81 9 AM–7 PM, Monday-Sunday. Patients who were <18 years of age, critically ill, or with language barrier were excluded. To quantify the educational value of the VEM, a six-question pretest was administered to assess baseline understanding of CT imaging. The patients then watched the VEM via iPad (Macintosh) and reviewed the consent form. An eight-question post-test was then completed by each subject. No PHI were collected. Pre- and post-test results were analyzed using McNemar’s test for individual questions and a paired t-test for the summed score (SAS version 9.2). Results: 100 patients consented and completed the survey. The average pre-test score for subjects was poor, 66% correct. Review of VEM/ICCT materials increased patient understanding of medical radiation as evidenced by improved post-test score to 79%. Mean improvement between tests was 13% (p < 0.0001). 78% of subjects responded that they found the materials helpful, and that they would like to receive ICCT. Conclusion: The addition of a video educational module improved patient knowledge regarding CT imaging and medical radiation as quantified by pre- and posttesting. Patients in our study sample reported that they prefer to receive ICCT. By educating patients about the risks associated with CT imaging, we increase informed, shared decision making – an essential component of patient-centered care. 147 Does Pain Intensity Reduce Willingness to Participate in Emergency Medicine Research? Alexander T. Limkakeng, Caroline Freiermuth, Weiying Drake, Giselle Mani, Debra Freeman, Abhinav Chandra, Paula Tanabe, Ricardo Pietrobon Duke University, Durham, NC Background: Multiple factors affect patients’ decisions to participate in ED research. Prior studies have shown that altruism and convenience to the patient may be important factors. Patients’ pain levels have not been explored as a reason for refusal. Objectives: We sought to determine the relationship between patients’ pain scores and their rate of consent to ED research. We hypothesized that patients with higher pain scores would be less likely to consent to ED research. Methods: Retrospective observational cohort study of potential research subjects in an urban academic hospital ED with an average annual census of approximately 70,000 visits. Subjects were adults older than 18 years with chief complaint of chest pain within the last 12 hours, making them eligible for one of two cardiac biomarker research studies. The studies required only blood draws and did not offer compensation. Two reviewers extracted data from research screening logs. Patients were grouped according to pain score at triage, pain score at the time of approach, and improvement in pain score (triage score - approach score). The main outcome was consent to research. Simple proportions for consent rates by pain score tertiles were calculated. Two multivariate logistic regression analyses S82 2012 SAEM ANNUAL MEETING ABSTRACTS were performed with consent as outcome and age, race, sex, and triage or approach pain score as predictors. Results: Overall, 396 potential subjects were approached for consent. Patients were 58% Caucasian, 49% female, and with an average age of 57 years. Six patients did not have pain scores recorded at all and 48 did not have scores documented within 2 hours of approach and were excluded from relevant analyses. Overall, 80.1% of patients consented. Consent rates by tertiles at triage, at time of approach, and by pain score improvement are shown in Tables 1 and 2. After adjusting for age, race, and sex, neither triage (p = 0.75) nor approach (p = 0.65) pain scores predicted consent. Conclusion: Research enrollment is feasible even in ED patients reporting high levels of pain. Patients with modest improvements in pain levels may be more likely to consent. Future research should investigate which factors influence patients’ decisions to participate in ED research. Table 1 - Abstract 147: Consent Rates by Pain Scores. Pain Score Triage Score (n) Consent Rates by Triage Score 0 1–3 4–6 7–10 105 86 104 95 77.1% 84.9% 79.8% 80.0% Approach Score (n) Consent Rates by Approach Score 139 61 76 69 76.3% 85.2% 80.3% 78.3% Table 2 - Abstract 147: Consent Rates by Pain Improvement Pain Improvement Pain Worsened (-1 to -10) 0 1 2 3–10 n Consent rate 36 231 26 29 68 77.8% 80.1% 84.6% 89.7% 79.4% Methods: We performed a 16-center, prospective cohort study of hospitalized children age <2 years with a physician diagnosis of bronchiolitis. For three consecutive years from November 1 until March 31, beginning in 2007, researchers collected clinical data and a nasopharyngeal aspirate. Intensive care unit visits were oversampled. Polymerase chain reaction was conducted for 15 viruses and viral load testing for the five most common viruses. Analysis used chi-square and Kruskal Wallis tests and multivariable logistic regression. Model validation was performed using bootstrapping. Results are reported as odds ratios (OR) with bias-corrected and accelerated 95% confidence intervals (95%CI). Results: Of the 2,207 enrolled children, 161 (7%) required CPAP or intubation. Overall, the median age was 4 months; 59% were male; 61% were white, 24% black and 36% Hispanic; 43% had respiratory syncytial virus (RSV)-A, 30% had RSV-B, and 26% had rhinovirus. In the multivariable model predicting CPAP/intubation, the significant factors were: age <2 months (OR 4.3; 95%CI 1.7–11.5), mother smoked during pregnancy (OR 1.4; 95%CI 1.1–1.9), birth weight <5 pounds (OR 1.7; 95%CI 1.0–2.6), breathing difficulty began <24 hours prior to admission (OR 1.6; 95%CI 1.2–2.1), presence of apnea (OR 4.8; 95%CI 2.5–8.5), inadequate oral intake (OR 2.5; 95%CI 1.3–4.3), severe retractions (OR 11.1; 95%CI 2.4–33.0), and room air oxygen saturation <85% (OR 3.3; 95%CI 2.0–4.8). Sex, race/ethnicity, viral etiology, and viral load were not predictive. Conclusion: In this multicenter study of children hospitalized with bronchiolitis neither specific viruses nor their viral load predicted the need for CPAP or intubation, but young age, low birth weight, presence of apnea, severe retractions, and oxygen saturation <85% did. We also identified that children requiring CPAP or intubation were more likely to have mothers who smoked during pregnancy and a rapid respiratory worsening. Mechanistic research in these high-risk children may yield important insights for the management of severe bronchiolitis. 149 148 A Multicenter Study To Predict Continuous Positive Airway Pressure And Intubation For Children Hospitalized With Bronchiolitis Patricio De Hoyos1, Jonathan M. Mansbach2, Pedro A. Piedra3, Ashley F. Sullivan1, Sunday Clark4, Janice A. Espinola1, Tate F. Forgey1, Carlos A. Camargo Jr.1 1 Massachusetts General Hospital, Boston, MA; 2 Children’s Hospital Boston, Boston, MA; 3 Baylor College of Medicine, Houston, TX; 4 University of Pittsburgh, Pittsburgh, MA Background: It is unclear which children hospitalized with bronchiolitis will require continuous positive airway pressure (CPAP) or intubation. Objectives: To examine the historical, clinical, and infectious pathogen factors associated with CPAP or intubation. Prevalence of Abusive Injuries in Siblings and Contacts of Abused Children Daniel Lindberg Brigham & Women’s Hospital, Boston, MA Background: Siblings and children who share a home with a physically abused child are thought to be at high risk for abuse. However, rates of injury in these children are unknown. Disagreements between medical and Child Protective Services professionals are common and screening is highly variable. Objectives: Our objective was to measure the rates of occult abusive injuries detected in contacts of abused children using a common screening protocol. Methods: This was a multi-center, observational cohort study of 20 child abuse teams who shared a common screening protocol. Data were collected between Jan 15, 2010 and April 30, 2011 for all children <10 years undergoing evaluation for physical abuse and their contacts. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 For contacts of abused children, the protocol recommended physical examination for all children <5 years, skeletal survey and physical exam for children <24 months, and physical exam, skeletal survey, and neuroimaging for children <6 months old. Results: Among 2,825 children evaluated for abuse, 618 met criteria as ‘‘physically abused’’ and these had 477 contacts. For each screening modality, screening was completed as recommended by the protocol in approximately 75% of cases. Of 134 contacts who met criteria for skeletal survey, new injuries were identified in 16 (12.0%). None of these fractures had associated findings on physical examination. Physical examination identified new injuries in 6.2% of eligible contacts. Neuroimaging failed to identify new injuries among 25 eligible contacts less than 6 months old. Twins were at significantly increased risk of fracture relative to other nontwin contacts (OR 20.1). Conclusion: These results support routine skeletal survey for contacts of physically abused children <24 months old, regardless of physical examination findings. Even for children where no injuries are identified, these results demonstrate that abuse is common among children who share a home with an abused child, and support including contacts in interventions (foster care, safety planning, social support) designed to protect physically abused children. Table - Abstract 149: Results of Screening Tests Screening Test Skeletal Survey Neuroimaging (CT or MR) Physical Examination 150 # Eligible # Tested # With Injury % Injured 95%CI 134 101 16 11.9 7.5–18.5 25 19 0 0.0 0.0–13.7 355 259 22 6.2 4.1–9.3 Validity of the Canadian Triage and Acuity Scale for Children: A Multi-centre, Database Study Jocelyn Gravel1, Eleanor Fitzpatrick2, Serge Gouin3, Kelly Millar4, Sarah Curtis5, Gary Joubert6, Kathy Boutis7, Chantal Guimont8, Ran D. Goldman9, Sasha Alexander Dubrovsky10, Robert Porter11, Darcy Beer12, Martin H. Osmond13 1 Sainte-Justine UHC, Université de Montréal, Montreal, QC, Canada; 2IWK Health Centre, Halifax, NS, Canada; 3CHU Sainte-Justine, Montreal, QC, Canada; 4Alberta Children’s Hospital, Calgary, AB, Canada; 5Stollery Children’s Hospital, Edmonton, AB, Canada; 6 Children’s Hospital of Western Ontario, London, ON, Canada; 7The Hospital for Sick Children, Toronto, ON, Canada; 8Centre Hospitalier Université Laval, Quebec, QC, Canada; 9BC Children’s Hospital, Vancouver, BC, Canada; 10Montreal Children’s Hospital, • www.aemj.org S83 Montreal, QC, Canada; 11Janeway Children’s Health and Rehabilitation Centre, St-John’s, NL, Canada; 12Children’s Hospital of Winnipeg, Winnipeg, MB, Canada; 13Children’s Hospital of Eastern Ontario, Ottawa, ON, Canada Background: The Canadian Triage and Acuity Scale (CTAS) is a five-level triage tool constructed from a consensus of experts. Objectives: To evaluate the validity of the Canadian Triage and Acuity Scale (CTAS) for children visiting multiple pediatric Emergency Departments (ED) in Canada. Methods: This was a retrospective study evaluating all children presenting to eight paediatric, universityaffiliated EDs during one year in 2010–2011. In each setting, information regarding triage and disposition were prospectively registered by clerks in the ED database. Anonymized data were retrieved from the ED computerized database of each participating centre. In the absence of a gold standard for triage, hospitalisation, admission to intensive care unit (ICU), length of stay in the ED, and proportion of patients who left without being seen by a physician (LWBS) were used as surrogate markers of severity. The primary outcome measure was the association between triage level (from 1 to 5) and hospitalisation. The association between triage level and dichotomous outcomes was evaluated by a chi-square test, while a Student’s t-test was used to evaluate the association between triage level and length of stay. It was estimated that the evaluation of all children visiting these EDs for a one year period would provide a minimum of 1,000 patients in each triage level and at least 10 events for outcomes having a proportion of 1% or more. Results: A total of 404,841 children visited the eight EDs during the study period. Pooled data demonstrated hospitalisation proportions of 59%, 30%, 10%, 2%, and 0.5% for patients triaged at level 1, 2,3, 4, and 5 respectively (p < 0.001). There was also a strong association between triage levels and admission to ICU (p < 0.001), the proportion of children who LWBS (p < 0.001), and length of stay (p < 0.001). Conclusion: The CTAS is a valid triage tool for children as demonstrated by its good correlation with markers of severity in multiple pediatric emergency departments. 151 Diagnosing Intussusception by Bedside Ultrasonography in the Pediatric Emergency Department Jessica Zerzan, Alexander Arroyo, Eitan Dickman, Hector Vazquez Maimonides Medical Center, Brooklyn, NY Background: Intussusception (INT) is the most common cause of bowel obstruction in children ages 3 months to 6 years. If not recognized and treated expeditiously, edema and hypoperfusion can lead to decreased rates of nonsurgical reduction, bowel perforation, necrosis, and even death. Many hospitals do not S84 2012 SAEM ANNUAL MEETING ABSTRACTS have 24-hour availability of ultrasonography through the radiology department (RADS). Bedside ultrasonography (BUS) performed by pediatric emergency medicine (PEM) physicians for the diagnosis of INT has not previously been studied. This technique could decrease time to diagnosis and definitive treatment. Objectives: The primary objective of this study is to compare the BUS results obtained by PEM physicians with ultrasounds performed by staff in RADS when evaluating patients for INT. Our hypothesis is that with focused training, PEM physicians can perform and interpret BUS for INT with results similar to that performed in RADS. Methods: We conducted a prospective observational diagnostic study in a community-based academic urban pediatric ED with an annual census of 36,000 patients. The principal investigator provided a brief in-service consisting of a didactic and hands-on BUS training session for all PEM attendings and fellows. We enrolled patients aged 3 months through 6 years in whom there was clinical suspicion of INT. After obtaining informed consent, the treating PEM physician performed a focused BUS, recorded specific images, and documented an interpretation of the study. Then the patient had an additional abdominal ultrasound in RADS. Results between PEM BUS and RADS sonography were compared using a kappa analysis. Results: 99 patients were enrolled. Kappa = 0.825 (0.63– 1.0). +LR = 40 (10–160), -LR = 0.11 (0.02–0.72). Sensitivity = 89% (0.05–0.99). Specificity = 98% (0.91–0.99). Conclusion: This study demonstrates that PEM physicians can accurately utilize BUS to evaluate patients for INT after a brief training period. This has the potential to expeditie definitive diagnosis and treatment of a true pediatric abdominal emergency and offers further evidence of the utility of BUS in the pediatric emergency department. Limitations of the study are those inherent in convenience sampling and its small sample size. Further large studies are warranted to determine whether these results can be replicated across other settings. Objectives: To determine the test performance characteristics for point-of-care (POC) US performed by pediatric emergency medicine (PEM) physicians compared to radiographic diagnosis of elbow fractures. Methods: This was a prospective study of children <21 years presenting to the ED with elbow injuries requiring x-rays. Patients were excluded if they arrived at the ED with elbow x-rays or a diagnosis of fracture. Prior to the start of the study, PEM physicians received a one-hour didactic and hands-on training session on US examination of the elbow. Before obtaining x-rays, the PEM physician performed a brief elbow US using a linear 10–5 MHz transducer probe and recorded images and clips in longitudinal and transverse views. A positive US for fracture at the elbow was defined as the PEM physician’s determination of an elevated posterior fat pad (PFP) and/or lipohemarthrosis (LH) of the PFP. All study patients received a standard care elbow x-ray in the ED and clinical telephone follow-up. The gold standard for fracture in this study was defined as fracture on initial or follow-up x-rays as determined by a radiologist. Results: 122 patients were enrolled with a mean age of 7.6 (± 5.4) years. 42/122 (34%) had a positive x-ray for fracture. Of the 65/122 (53%) patients with a positive US, 58/65 (89%) had an elevated PFP, 56/65 (86%) had LH, and 49/65 (75%) had both an elevated PFP and LH. A positive elbow ultrasound with an elevated PFP or LH had a sensitivity of 0.98 (95% CI 0.88–1.00), specificity of 0.72 (95% CI 0.61–0.81), positive predictive value of 0.67 (95% CI 0.55–0.77), negative predictive value of 0.98 (95% CI 0.91–1.00), positive likelihood ratio of 3.54 (95% CI 2.45–5.10), and negative likelihood ratio of 0.03 (95% CI 0.01–0.22) for a fracture. The use of POC elbow US would reduce the need for x-rays in 56/122 (46%) patients with elbow injuries but would miss one fracture. Conclusion: POC US was found to be highly sensitive for elbow fractures when performed by PEM physicians. A negative POC US may reduce the need for x-rays in children with elbow injuries. Table - Abtsract 151: Results 153 +BUS )BUS Total 152 +RADS )RADS Total 8 1 9 2 88 90 10 89 99 Accuracy of Point-of-Care Ultrasound for Diagnosis of Elbow Fractures in Children Joni E. Rabiner1, Hnin Khine1, Jeffrey R. Avner1, James W. Tsung2 1 Children’s Hospital at Montefiore, Bronx, NY; 2 Mount Sinai Medical Center, New York, NY Background: Ultrasound (US) has been shown to be useful in the diagnosis of pediatric skeletal injuries. It can be performed accurately and reliably by emergency department (ED) physicians with focused US training. Persistent Failure to Understand Key Elements of Medication Safety after Pediatric Emergency Department Visit Margaret E. Samuels-Kalow1, Anne M. Stack2, Stephen C. Porter3 1 BWH/MGH Harvard Affiliated Emergency Medicine Residency, Boston, MA; 2Children’s Hospital Boston, Boston, MA; 3The Hospital for Sick Children, Toronto, ON, Canada Background: Parents frequently leave the emergency department (ED) with incomplete understanding of the diagnosis and plan, but the relationship between comprehension and post-care outcomes has not been well described. Objectives: To explore the relationship between comprehension and post-discharge medication safety. Methods: We completed a planned secondary analysis of a prospective observational study of the ED discharge process for children aged 2–24 months. After discharge, parents completed a structured interview to assess ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 comprehension of the child’s condition, the medical team’s advice, and the risk of medication error. Limited understanding was defined as a score of 3–5 from 1 (excellent) to 5 (poor). Risk of medication error was defined as a plan to use over-the-counter cough/cold medication and/or an incorrect dose of acetaminophen (measured by direct observation at discharge or reported dose at follow-up call). Parents identified as at risk received further instructions from their provider. The primary outcome was persistent risk of medication error assessed at phone interview 5–10 days post-discharge. Results: The original cohort screened 270 patients, of whom 204 (76%) were eligible and consented. 150/204 (74%) completed the assessment at discharge. 130/150 (87%) completed the follow-up interview. At discharge, 38 parents (25%) reported limited understanding and 48 (31%) had a risk of medication error. At follow-up, 42 parents (32%) were at risk of medication error, of whom 22 (52%) had also been at risk at discharge. Limited parental understanding at discharge was significantly associated with risk of medication error at follow-up [odds ratio (OR) 3.9, 95% CI 1.7, 8.9], and remained so after adjustment for parental language, health literacy, and error risk at discharge (aOR 3.1, 95% CI 1.1, 8.8). Risk of medication error at discharge was significantly associated with risk at follow-up, despite provider intervention after the risk was identified, (OR 4.6, 95% CI 2.0, 10.3), with a trend toward significance after adjustment for language and literacy (aOR 2.4, 95% CI 0.96, 6.0). Conclusion: Parental report of limited understanding predicted risk of medication errors after discharge from the pediatric ED. Despite provider intervention, families at risk of medication error at discharge appear to experience persistent risk of error after leaving the ED. 154 Use Of Oral Disintegrating Ondansetron To Treat Nausea In Prehospital Patients Steven Weiss1, Lynne Fullerton1, Phillip Froman2 1 University of New Mexico, Albuquerque, NM; 2 Albuquerque Ambulance, Albuquerque, NM Background: Use of various antiemetics or use of crystalloid alone has been shown to improve nausea in the ED and hospital. Their prehospital use is not yet clear. Objectives: Our hypotheses were that during EMS transport: (1) the amount of saline given is not related to a change in nausea and, (2) that the addition of ondansetron Oral disintegrating tablet (ODT) decreases the degree of nausea. Methods: During the first phase of the study, EMS providers completed a form whenever a patient with nausea and/or vomiting was assessed and transported to one of the area hospitals. Patients were asked to rate their nausea on a visual analogue scale (VAS) and a Likert scale. Saline administration and active vomiting were documented. During the second phase, ODT ondansetron was added for all patients with moderate or severe nausea and providers continued to complete the same forms. • www.aemj.org S85 The forms completed during phase 1 (saline-only) and phase 2 (ODT ondansetron) were evaluated and compared. For both phases, the primary outcome measures were change in VAS nausea rating (0 = no nausea, 100 = most nausea ever) from beginning to end of the transport, and the results on the Likert scale completed at the end of the transport. Negative changes in VAS scores indicated improvement in nausea. Spearman correlations and Wilcoxon tests were used for analysis. Results: Data were collected from 274 transports in phase 1, and 372 transports in phase 2. The average age was 50 ± 12 yrs. In phase 1, 135/228 (59%) received normal saline (mean volume/sd = 265 ± 192 ml).There was no significant correlation between either the VAS change or the Likert scale results and amount of fluid administration in the saline-only phase of the study. Conversely, during phase 2, patients receiving ondansetron ODT showed significant improvement in both measures of nausea. (difference in VAS change: )24.6, 95% CI 20.9, 28.3). See figure. Conclusion: There was no relationship between nausea and quantity of saline during an EMS transport. However, ODT ondansetron resulted in a significant improvement in nausea. 155 Does the Addition of the Mucosal Atomizer Device Increase Fentanyl Administration in Prehospital Pediatric Patients? Daniel O’Donnell, Luke Schafer, Andrew Stevens Indiana University School of Medicine, Indianapolis, IN Background: Prehospital research demonstrates that providers inadequately treat pain in pediatric patients. A major barrier to administering analgesics to children is the perceived discomfort of intravenous access. The delivery of intranasal analgesia may be a novel solution to this problem. Objectives: We investigated whether the addition of the Mucosal Atomizer Device (MAD) as an alternative for fentanyl delivery would improve overall fentanyl administration rates in pediatric patients transported by a large urban EMS system. S86 2012 SAEM ANNUAL MEETING ABSTRACTS Methods: We performed a historical control trial comparing the rate of pediatric fentanyl administration 6 months before and 6 months after the introduction of the MAD. Study subjects were pediatric trauma patients (age <16 years) transported by a large urban EMS agency. The control group was composed of patients treated in the 6 months before introduction of the MAD. The experimental group included patients treated in the 6 months after the addition of the MAD. Two physicians reviewed each chart and determined whether the patient met predetermined criteria for the administration of pain medication. A third reviewer resolved any discrepancies. Fentanyl administration rates were measured and compared between the two groups. We used two-sample t-tests and chi-square tests to analyze our data. Results: 228 patients were included in the study: 137 patients in the pre-MAD group and 91 in the post-MAD group. There were no significant differences in the demographic and clinical characteristics of the two groups. 42 (30.4%) patients in the control arm received fentanyl. 34 (37.8%) of patients in the experimental arm received fentanyl with 36% of the patients receiving fentanyl via the intranasal route. The addition of the MAD was not associated with a statistically significant increase in analgesic administration. Age and mechanism of injury were statistically more predictive of analgesia administration. Conclusion: While the addition of the Mucosal Atomizer Device as an alternative delivery method for fentanyl shows a trend towards increased analgesic administration in a prehospital pediatric population, age and mechanism of injury are more predictive in who receives analgesia. Further research is necessary to investigate the effect of the MAD on pediatric analgesic delivery. Table - Abstract 155: Calculated Odds Ratio of Receiving Fentanyl Post MAD Female Age Initial RR Initial GCS Initial Pulse Burn vs. Other Fall/Musc. vs. Other MVC vs. Other Odds Ratio 95% CI P Value 1.27 0.84 1.17 0.99 1.32 1.02 7.66 2.05 0.36 (0.66–2.44) (0.44–1.58) (1.06–1.30) 0.92–1.06) (0.58–3.01) (0.99–1.04) (1.52–38.57) (0.99–4.25) (0.13–1.03) 0.468 0.580 0.002 0.738 0.512 0.083 0.014 0.055 0.053 156 EMS Provider Self-efficacy Retention after Pediatric Pain Protocol Implementation April Jaeger1, Maija Holsti2, Nanette Dudley2, Xiaoming Sheng2, Kristin Gurley3, Kathleen Adelgais4 1 Sacred Heart Medical Center, Spokane, WA; 2 University of Utah, Salt Lake City, UT; 3 University of Pittsburgh Institute of Aging, Pittsburgh, PA; 4University of Colorado, Denver, CO Background: Prehospital providers (PHPs) commonly treat pain in pediatric patients. Self-efficacy (SE) is a person’s judgment of his or her capability to perform certain actions and is congruent with clinical performance. An initial pilot study showed that a pediatric pain protocol (PPP) increases PHP-SE. The retention of PHP-SE after PPP implementation is unknown. Objectives: To evaluate the retention of PHP-SE after PPP implementation. Methods: This was a prospective study evaluating PHPSE before (pre) and after (post) a PPP introduction and 13 months later (13-mo). PHP groups received either PPP review and education or PPP review alone. The PPP included a pain assessment tool. The SE tool, developed and piloted by pediatric EMS experts, uses a ranked ordinal scale ranging from ‘certain I cannot do it’ (0) to ‘completely certain I can do it’ (100) for 10 items: pain assessment (3 items), medication administration (4) and dosing (1), and reassessment (2). All 10 items and an averaged composite were evaluated for three age groups (adult, child, toddler). Paired sample t-tests compared post- and 13-mo scores to pre-PPP scores. Results: Of 264 PHPs who completed initial surveys, 146 PHPs completed 13-mo surveys. 106 (73%) received education and PPP review and 40 (27%) review only. PPP education did not affect PHP-SE (adult P = 0.87, child P = 0.69, toddler P = 0.84). The largest SE increase was in pain assessment. This increase persisted for child and toddler groups at 13 months. The immediate increase in composite SE scores for all age groups persisted for the toddler group at 13 months. Conclusion: Increases in composite and pain assessment PHP-SE occur for all age groups immediately after PPP introduction. The increase in pain assessment SE persisted at 13 months for pediatric age groups. Composite SE increase persisted for the toddler age group alone. Table - Abstract 156: Self-efficacy Scores Adult Averaged Composite 0–100 (95% CI) Pain Assessment 0–100 (95% CI) *P < 0.05 Pre Post 13-mo Pre Post 13-mo 82.40 84.99 84.28 89.42 94.64 92.57 (78.55–86.35) (81.37–88.62)* (80.80–87.76) (85.42–93.67) (91.55–97.74)* (89.72–95.42) Child 75.37 80.91 77.62 77.86 89.14 84.14 (71.14–79.59) (77.09–84.73)* (73.70–81.55) (72.93–82.80) (85.99–92.30)* (80.68–87.61)* Toddler 64.64 76.89 71.39 50.89 83.86 72.50 (60.19–69.10) (73.02–80.76)* (67.03–75.74)* (45.21–56.57) (80.24–87.47)* (67.87–77.13)* ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 157 Hemodynamic Changes In Patients Receiving Ketamine Sedation By Emergency Medical Services Jay L. Kovar, Guy R. Gleisberg, Eric R. Ardeel, Abdullah Basnawi, Mark E.A. Escott Baylor College Of Medicine / EMS Collaborative Research Group Background: Ketamine is widely used to achieve sedation and analgesia. In the prehospital setting, ketamine facilitates endotracheal intubation, pain control, and initial management of acutely agitated patients. Knowledge of the hemodynamic effects of ketamine in the prehospital setting is limited and may prevent acceptance of the practice. Objectives: Describe the hemodynamic effects of prehospital ketamine administration in patients treated by paramedics while undergoing continuous monitoring. Methods: Retrospective review of electronic prehospital care records for 98 consecutive patients receiving prehospital ketamine from paramedics in Montgomery County, Texas between August 1, 2010 and October 25, 2011. Ketamine administration indications were: control of violent/agitated patients requiring treatment and transport; sedation and analgesia after trauma; facilitation of intubation and mechanical ventilation. All patients receiving ketamine were treated under unified prehospital protocols and were eligible for study inclusion. Exclusion criteria included those waveforms: where hemodynamic monitoring data did not contain at least three data points within the duration of action; distorted by explainable events; or from subsequent doses in a multi-dose administration. Linear regression trend analysis modeled each waveform and the percentage difference algorithm was used to calculate each hemodynamic parameter effect. Results: 219 of 490 (45%) waveforms met inclusion criteria. Observed hemodynamic percentage difference medians: heart rate decreased by 0.2% (range )99 to +87%); SpO2 increased by 0.6% (range )16 to +33%); etCO2 increased by 1.3% (range )75 to +123%); respiratory rate decreased by 14% (range )85 to +156%); and mean arterial pressure 0.0% (range )78 to +80%). Mean age was 41 years (range 3–94 years), 50 subjects (51%) were male, and the mean ketamine dose was 150 mg (range 25–500 mg). Conclusion: Our findings demonstrate the feasibility of prehospital hemodynamic monitoring in patients receiving ketamine. Minimal hemodynamic changes were observed during the dose response window. Further studies are necessary to discern the clinical implications of the findings in these specific parameters. 158 Do Paramedics Give Epinephrine When Indicated For Anaphylaxis? Herbert G. Hern1, Harrison Alter1, Joseph Barger2 1 Alameda County - Highland, Oakland, CA; 2 Contra Costa County EMS Agency, Martinez, CA Background: Allergic and anaphylactic reactions in the pediatric population are a serious and potentially • www.aemj.org S87 deadly disease. Instances of allergic and anaphylactic reactions have been increasing and the need for lifesaving intervention with epinephrine must remain an important part of EMS provider training. Objectives: To characterize dosing of and timing of epinephrine, diphenhydramine, and albuterol in the pediatric patient with severe allergy and anaphylaxis. Methods: We studied medication administration in pediatric patients with severe allergy or anaphylaxis during our study period of 19 months, from January 1, 2010 to July 10, 2011. We compared rates of epinephrine, diphenhydramine, and albuterol given to patients with allergic conditions (including anaphylaxis). In addition, we calculated the rate of epinephrine administration in cases of anaphylaxis and determined what percentage of time the epinephrine was given by EMS or prior to EMS arrival. Results: Out of 239,320 total patient contacts, 12,898 were pediatric patients. Of the pediatric patient contacts, 199 were transported for allergic complaints. Of those with allergic complaints, 97 of 199 (49%; 95% CI 42%, 56%) had symptoms consistent with anaphylaxis and indications for epinephrine. Of these 97, 49 (51%, 95% CI 41%, 60%) were given epinephrine. Among patients in whom epinephrine was indicated and given, 42 (86%; 95% CI 73%, 94%) were given epinephrine prior to EMS arrival (by parent, school, or clinic). Of the 55 patients who might have benefited from epinephrine from EMS, 7 (14%; 95% CI 6%, 24%) received epinephrine, 28 (51%; 95% CI 38%, 64%) received diphenhydramine, and 28 (51%; 95% CI 38%, 64%) received albuterol. Conclusion: In anaphylaxis patients who met criteria for epinephrine, only 49% received epinephrine and the overwhelming majority received it prior to EMS arrival. EMS providers were far more likely to give diphenhydramine or albuterol than epinephrine in patients who met criteria for epinephrine. EMS personnel are not treating anaphylaxis appropriately with epinephrine. 159 Weight-Based Pediatric Dosing Errors Are Common Among EMS Providers Joseph Barger1, Herbert G. Hern2, Patrick Lickiss3, Harrison Alter2, Maria Fairbanks1, Micheal Taigman3, Monica Teeves4, Karen Hamilton4, Leslie Mueller4 1 Contra Costa County EMS Agency, Martinez, 2 CA; ACMCHighland, Oakland, CA; 3 American Medical Response, Oakland, CA; 4 American Medical Response, Concord, CA Background: Pediatric medications administered in the prehospital setting are given infrequently and dosage may be prone to error. Calculation of dose based on known weight or with use of length-based tapes occurs even less frequently and may present a challenge in terms of proper dosing. Objectives: To characterize dosing errors based on weight-based calculations in pediatric patients in two similar emergency medical service (EMS) systems. Methods: We studied the five most commonly administered medications given to pediatric patients weighing S88 2012 SAEM ANNUAL MEETING ABSTRACTS 36 kg or less. Drugs studied were morphine, midazolam, epinephrine 1:10,000, epinephrine 1:1000, and diphenhydramine. Cases from the electronic record were studied for a total of 19 months, from January 2010 to July 2011. Each drug was administered via intravenous, intramuscular, or intranasal routes. Drugs that were permitted to be titrated were excluded. An error was defined as greater than 25% above or below the recommended mg/kg dosage. Results: Out of 248,596 total patients, 13,321 were pediatric patients. 7885 had documented weights of <36 kg and 241 patients were given these medications. We excluded 72 patients for weight above the 97%ile or below the 3%ile, or if the weight documentation was missing. Of the 169 patients and 187 doses, errors were noted in 53 (28%; 95% CI 22%, 35%). Midazolam was the most common drug in errors (29 of 53 doses or 55%; 95% CI 40%, 68%), followed by diphenhydramine (11/53 or 21%; 95% CI 11%, 34%), epinephrine (7/53 or 13%; 95% CI 5%, 25%), and morphine sulfate (6/53 or 11%; 95% CI, 4%, 23%). Underdosing was noted in 34 of 53 (64%; 95% CI 50%, 77%) of errors, while excessive dosing was noted in 19 of 53 (36%; 95% CI 23%, 50%). Conclusion: Weight-based dosing errors in pediatric patients are common. While the clinical consequences of drug dosing errors in these patients are unknown, a considerable amount of inaccuracy occurs. Strategies beyond provision of reference materials are needed to prevent pediatric medication errors and reduce the potential for adverse outcomes. 160 Drivers Of Satisfaction: What Components Of Patient Care Experience Have The Greatest Influence On The Overall Perceptions Of Care? Timothy Cooke1, Alaina Aguanno2, John Cowell2, Richard Schorn1, Markus Lahtinen1, Andrew McRae2, Brian Rowe3, Eddy Lang2 1 Health Quality Council of Alberta, Calgary, AB, Canada; 2University of Calgary, Calgary, AB, Canada; 3University of Alberta, Edmonton, AB, Canada Background: Improving patient satisfaction is a core mandate of all health care institutions and EDs. Which specific components of patient care have the greatest influence on the overall patient experience is poorly understood. Objectives: This study examines the magnitude of influence of specific modifiable factors (e.g. wait time) on the patient ED experience. Methods: This was a cross-sectional survey of a random patient sample from 12 urban/regional EDs in winter 2007 and 2009. A previously validated questionnaire, based on the British Healthcare Commission Survey, was distributed according to a modified Dillman protocol. Exclusion criteria: age 0–15 years, left prior to being seen/treated, died during the ED visit, no contact information, presented with a ‘‘privacy’’ sensitive case. Nine previously identified and validated composite variables (staff care/communication, respect, pain management, wait time/crowding, facility cleanliness, discharge communication, wait time communication, medication communication, and privacy) were evaluated for their influence on patients’ overall rating of care. Composite variables and rating of care were rated on a scale from from 0 (lowest) to 100 (highest). Calculated influence is represented as a standardized composite coefficient. Results: 21,639 surveys were distributed with a response rate of 46%. Composite variable coefficients were: staff care/communication (0.38), respect (0.17), pain management (constituent variables not comparable so decomposed for regression), wait time/crowding (0.09), facility cleanliness (0.13), discharge communication (0.10), wait time communication (0.03), medication communication (no significant influence), and privacy (0.02). Thus, if a patient’s staff care/communication composite score increases from 50/100 to 70/100, an initial global rating of care of 70/100 is predicted to increase to 78/100. Path analysis showed cascading effects of pain management and wait time on other variables. Variable coefficients affect patients according to situational relevance. Conclusion: Global patient ED experience is most strongly influenced by staff care/communication. Pain management and wait time synergistically effect downstream variables. Efforts to improve patient satisfaction should focus on these factors. 161 Street Outreach Rapid Response Team for the Homeless Lindsay M. Harmon1, Aaron Kalinowski1, Dorian Herceg2, Anthony J. Perkins1 1 Indiana University School of Medicine, Indianapolis, IN; 2Wishard Hospital, Indianapolis, IN Background: Homelessness affects up to 3.5 million people a year. The homeless present more frequently to EDs, their ED visits are four times more likely to occur within 3 days of a prior ED evaluation, and they are admitted up to five times more frequently than others. We evaluated the effect of a Street Outreach Rapid Response Team (SORRT) on the health care utilization of a homeless population. A nonmedical outreach staff responds to the ED and intensely case manages the patient: arranges primary care follow-up, social services, temporary housing opportunities, and drug/ alcohol rehabilitation services. Objectives: We hypothesized that this program would decrease the ED visits and hospital admissions of this cohort of patients. Methods: Before and after study at an urban teaching hospital from June, 2010–December, 2011 in Indianapolis, Indiana. Upon identification of homeless status, SORRT was immediately notified. Eligibility for SORRT enrollment is determined by Housing and Urban Development homeless criteria and the outreach staff attempted to enter all such identified patients into the program. The patients’ health care utilization was evaluated in the 6 months prior to program entry as compared to the 6 months after enrollment by prospectively ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 collecting data and a retrospective medical record query for any unreported visits. Since the data were highly skewed, we used the nonparametric signed rank test to test for paired differences between periods. Results: 22 patients met criteria but two refused participation. The 20-patient cohort had 388 total ED visits (175 pre and 213 post) with a mean of 8.8 (SD 10.1) and median of 6.5 (range 1–44) ED visits in 6 months preSORRT as compared to a mean of 10.7 (SD 19.5) and median of 5.0 (0–90) in 6 months post-SORRT (p = 0.815). There were 28 total inpatient admissions pre-intervention and 27 post-intervention, with a mean of 1.4 (SD 2.0) and median of 0.5 (0.7) per patient in the pre-intervention period as compared to 1.4 (SD 1.9) and 1.0 (0–6) in the post-intervention period (p = 0.654). In the pre-SORRT period 50.0% had at least one inpatient admission as compared to 55.0% post-SORRT (p = 1.00). There were no differences in ICU days or overall length of stay between the two periods. Conclusion: An aggressive case management program beginning immediately with homeless status recognition in the ED has not demonstrated success in decreasing utilization in our population. 162 Communicating with Patients with Limited English Proficiency: Analysis of Interpreter Use and Comfort with a Second Language David T. Chiu, Jonathan Fisher, Alden Landry Beth Israel Deaconess Medical Center, Boston, MA Background: The number of Limited English Proficiency (LEP) patients presenting to emergency departments is increasing. Professional interpreters decrease errors in communication, improve patient satisfaction, equalize utilization of health care resources, and improve outcomes. Working with interpreters is often times inconvenient, forcing providers to self-translate or use family members despite not having any training. Objectives: To determine interpreter usage and selftranslation patterns by ED physicians. Methods: This is a cross-sectional survey study that was conducted at an urban, tertiary referral, academic hospital with 55,000 ED visits. All ED physicians at the institution were included; 45 attending and 37 postgraduate. The authors excluded themselves from the study. The anonymous survey consisted of questions regarding interpreter utilization, comfort with interpreters, self-translation, and use of family or nontrained staff for translation. Proportions and confidence intervals were calculated using SPSS 17. Results: 77 physicians completed the survey yielding a 96% response rate. 97% reported working with an interpreter at least once a month, with 47% working with an interpreter daily. Only 5% (CI 0–10) reported always working with an interpreter with 95% (CI 90–100) working with an interpreter sometimes or often. 77% (CI 67– 83) use ED staff to interpret, while 81% (CI 72–90) use family to interpret. 99% responded that they were comfortable taking a history, performing a physical exam, and discussing treatment plans or reassuring results with an interpreter. 10% reported being uncomfortable • www.aemj.org S89 giving bad news with an interpreter. 48% reported selftranslating with LEP patients. 47% of self-translators reported comfort with history taking while 33% are comfortable giving results and discharging. Conclusion: While ED physicians work with interpreters for LEP patients a majority of the time, only a small minority of ED physicians work with them all of the time. A large number use non-trained staff or family for translation. Nearly half of physicians self-translate; however, only half of those are comfortable taking a history and only a third comfortable with giving results or discharging. Future studies include barriers to interpreter use in the ED and testing those who self-translate for proficiency. 163 Why Did You Go to the Emergency Department? Findings from the Health Quality Council of Alberta Urban and Regional Emergency Department Patient Experience Report Eddy Lang1, Timothy Cooke2, Alaina Aguanno1, John Cowell2, Brian Rowe3, Andrew McRae1, Fareen Zaver4 1 University of Calgary, Calgary, AB, Canada; 2 Health Quality Council of Alberta, Calgary, AB, Canada; 3University of Alberta, Edmonton, AB, Canada; 4Mayo Clinic, Rochester, MN Background: Understanding why patients seek ED care is relevant to policy development. Large-scale population-based determinations of what motivates an ED visit are lacking. Objectives: To characterize and obtain a populationbased and in-depth understanding of what drives patients to seek ED care. Methods: Cross-sectional survey of a random patient sample from 12 urban/regional EDs in winter 2007 and 2009. A previously validated questionnaire, based on the British Healthcare Commission Survey, was distributed according to a modified Dillman protocol. Exclusion criteria: age 0–15 years, left prior to being seen/ treated, died during the ED visit, no contact information, presented with a privacy sensitive case. Aggregate responses are reported unless statistically significant differences were identified. Sample weights were applied adjusting for inter-facility sample proportion differences. Results: 21,639 surveys were distributed. Response rate was 46%. Patients sought ED care on the advice of a health care professional (35%), after consulting with a friend/family member (34%), and/or based on own assessment of need (34%). The ED was perceived as the best place for my problem (46% in 2007 and 48% in 2009; p = 0.03), the only choice available at the time (43%), and/or the most convenient place to go (12%). Most patients consulted the ED for a new illness (32%) or new injury (27%). The remainder of presentations were associated with pre-existing conditions: worsened chronic condition (23%), complication of recent medical care (13%), routine care (2%), follow-up care (2%), or other (2%). 89% of patients had a personal family doctor/specialist whom they saw for most of their S90 2012 SAEM ANNUAL MEETING ABSTRACTS health care needs and 96% of those patients had visited this physician within the past 12 months. Conclusion: Most patients consult with others prior to presenting for ED care. The decision to visit an ED is based on perceived need rather than convenience. The majority of presentations are related to new conditions, exacerbation of chronic conditions, or complications of prior care. 164 Variation of Patient Preferences for Written and Cell Phone Instructional Modality of Discharge Instructions by Patient Health Literacy Level Travis D. Olives, Roma G. Patel, Rebecca S. Nelson, Aileen Yew, Scott Joing, James R. Miner Hennepin County Medical Center, Minneapolis, MN Background: Outpatient antibiotics are frequently prescribed from the ED, and limited health literacy may affect compliance with recommended treatments. It is unknown whether the preference for discharge instructional modality varies by health literacy level. Objectives: We sought to determine if patient preference for multimodality discharge instructions for outpatient antibiotic therapy varies by health literacy level. Methods: This was a secondary analysis of a prospective randomized trial that included consenting patients discharged with outpatient antibiotics from an urban county ED with an annual census of 100,000. Patients unable to receive text messages or voice-mails were excluded. Health literacy was assessed using a validated health literacy assessment, the Newest Vital Sign (NVS). Patients were randomized to a discharge instruction modality: 1) standard care, typed and verbal medication and case-specific instructions; 2) standard care plus text-messaged instructions sent to the patient’s cell phone; or 3) standard care plus voice-mailed instructions sent to the patient’s cell. Patients were called at 30 days to determine preference for instruction delivery modality. Preference for discharge instruction modality was analyzed using z-tests for proportions. Results: 758 patients were included (55% female, median age 30, range 5 months to 71 years); 98 were excluded. 23% had an NVS score of 0–1, 31% 2–3, and 46% 4–6. Among the 51.1% of participants reached at 30 days, 26% preferred a modality other than written. There was a difference in the proportion of patients who preferred discharge instructions in written plus another modality (see table). With the exception of written plus another modality, patient preference was similar across all NVS score groups. Conclusion: In this sample of urban ED patients, more than one in four patients prefer non-traditional (text message, voice-mail) modalities of discharge instruction delivery to standard care (written) modality alone. Additional research is needed to evaluate the effect of instructional modality on accessibility and patient compliance. All Health Care is Not Local: An Evaluation of the Distribution of Emergency Department Care Delivered in Indiana John T. Finnell, J Marc Overhage, Shaun Grannis Indiana University, Indianapolis, IN 165 Background: The emergency department (ED) delivers a major portion of health care - often with incomplete knowledge about the patient. As such, EDs are particularly likely to benefit from a health information exchange (HIE). Objectives: To describe patient crossover rates throughout the entire state of Indiana over a three-year period. This information provides one estimate of the opportunities for cross-institutional data to influence medical care. Methods: The Indiana Public Health Emergency Surveillance System (PHESS) sends real-time registration information for emergency department encounters. We validated these transactions, determined which were unique visits, then matched the patients across the state. Table - Abstract 164: Patient-reported Preferences for Discharge Instruction Modality NVS Health Literacy Score: Written % (n) Written plus another modality Texted Voice-mailed Text and voice-mail Text, voice-mail, and written No preference Limited (0–1) 40 (32) 13.75 (11) 18.75 (15) 7.50 (6) 3.75 (3) 0.00 (0) 16.25 (13) Possibly limited (2–3) 31.82 42.21 10.39 3.25 4.55 1.30 6.49 (49) (65) (16) (5) (7) (2) (10) Adequate (4–6) 44.08 20.39 12.50 5.92 5.26 0.66 11.18 (67) (31) (19) (9) (8) (1) (17) p-val, NVS score 0–1 compared to 2–6 0.732 0.0017 0.083 0.294 0.664 0.374 0.052 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S91 Results: Over the three-year study period, we found 2.8 million patients generated 7.4 million ED visits. The average number of visits was 2.6 visits/patient (range 1–385). We found more than 40% of ED visits during the study period were for patients having data at multiple institutions. When examining the network density, we found nearly all EDs share patients with more than 80 other EDs. (image) Conclusion: Our results help clarify future health care policy decisions regarding optimal NHIN architecture and discount the notion that ‘all health care is local’. 166 Cumulative SAPS II Score Fails To Predict Mortality In Out-of-Hospital Cardiac Arrest Justin D. Salciccioli, Cristal Cristia, Andre Dejam, Tyler Giberson, Tara Melillo, Amanda Graver, Michael N. Cocchi, Michael W. Donnino BIDMC Center for Resuscitation Science, Boston, MA Background: Severity of illness scores can predict outcomes in critically ill patients. However, the calibration of existing scoring systems in post-cardiac arrest patients is poorly established. Objectives: To determine if the Simplified Acute Physiology Score (SAPS II) will predict mortality in out-ofhospital cardiac arrest (OHCA). Methods: We performed an observational study of adult cardiac arrest at an urban tertiary care hospital during the period from 12/2007 to 12/2010. Data were collected prospectively and recorded in the Utstein style. Inclusion criteria: 1. Adult (>18 years); 2. OHCA; 3. Return of spontaneous circulation. Traumatic cardiac arrests were excluded. Patient demographics, co-morbid conditions, vital signs, laboratory data, and in-hospital mortality were recorded (Table). SAPS II scores were calculated. We used simple descriptive statistics to describe the study population and logistic regression to predict mortality with SAPS II as a continuous predictor variable. Forward stepwise logistic regression selection was used to identify individual SAPS II variables that contribute to the sensitivity of the score. Discrimination was assessed using area under the curve (AUC) of the receiver operating characteristic (ROC) curve. Results: 115 patients were analyzed. The median age was 68 years (IQR 55–79) and 28% were female. Median SAPS II score was 67 (IQR 53–77) and 61% of patients died. Cumulative SAPS II score was a poor predictor of mortality (p = 0.19, OR 1.012, 95% CI 0.99–1.03) and demonstrated poor discrimination (AUC 0.61). Stepwise selection identified the following individual SAPS II variables to predict mortality: HCO3 (p = 0.006, OR 1.46, 95%CI 1.12–1.90), GCS (p = 0.03, OR 1.12, 95%CI 1.01– 1.23), and age (p = 0.02, OR 1.12, 95%CI 1.02–1.23); together these are strong predictors of mortality: p = 0.02, AUC: 0.75 (Figure). Conclusion: Cumulative SAPS II scoring fails to predict mortality in OHCA. The risk scores assigned to age, GCS, and HCO3 independently predict mortality and combined are good mortality predictors. These findings suggest that an alternative severity of illness score should be used in post-cardiac arrest patients. Future studies should determine optimal risk scores of SAPS II variables in a larger cohort of OHCA. Table - Abstract 166: Baseline Characteristics of Out-of-Hospital Cardiac Arrest Patients Characteristics Total number (n) Age yr(+ ⁄ )SD) Female- no. (%) Initial arrest rhythm Ventricular fibrillation/tachycardia Pulseless electrical activity Asystole Unknown Therapeutic hypothermia- no. (%) 167 OHCA 115 66 (17) 32 (28) 52 34 21 7 54 (44) (30) (18) (6) (47) Restoring Coronary Perfusion Pressure Before Defibrillation After Chest Compression Interruptions Ryan A. Coute, Timothy J. Mader, Adam R. Kellogg, Scot A. Millay, Lennard C. Jensen Baystate Medical Center, Springfield, MA Background: Generation of a threshold coronary perfusion pressure (CPP) is required for defibrillation success during VF cardiac arrest resuscitation. Rescue shock (RS) outcomes are directly related to the CPP achieved. Chest compression interruptions (for ECG rhythm analysis) cause a precipitous drop in CPP. Objectives: To determine the extent to which CPP recovers to pre-pause levels with 20 seconds of CPR after a 10-second interruption in chest compressions for ECG rhythm analysis. Methods: This was a secondary analysis of prospectively collected data from an IACUC-approved protocol. Fortytwo Yorkshire swine (weighing 25–30 kg) were instrumented under anesthesia. VF was electrically induced. After S92 2012 SAEM ANNUAL MEETING ABSTRACTS 12 minutes of untreated VF, CPR was initiated and a standard dose of epinephrine (SDE) (0.01 mg/kg) was given. After 2.5 minutes of CPR to circulate the vasopressor, compressions were interrupted for 10 seconds to analyze the ECG rhythm. This was immediately followed by 20 seconds of CPR to restore CPP before the first RS was delivered. If the RS failed, CPR resumed and additional vasopressors (SDE, and vasopressin 0.57 mg/kg) were given and the sequence repeated. The CPP was defined as aortic diastolic pressure minus right atrial diastolic pressure. The CPP values were extracted at three time points: immediately after the 2.5 minutes of CPR, following the 10-second pause, and immediately before defibrillation for the first two RS attempts in each animal. Eighty-three sets of measurements were logged from 42 animals. Descriptive statistics were used to analyze the data. Results: The data were not normally distributed. Our findings are presented in the table. Interrupting chest compressions to analyze the ECG VF rhythm caused a significant drop in CPP. Resuming CPR for 20 seconds prior to delivery of a RS restored CPP to 94.0% (95%CI 89.7–97.9) of the pre-pause values. Conclusion: Our data suggest that delaying defibrillation for 20 seconds after a pause in chest compressions and resuming CPR while the defibrillator is charging can restore CPP to levels more favorable to RS success. Whether or not this actually translates into greater RS success, higher rates of ROSC, and improved survival remains to be determined. 168 Rescue Shock Timing And Outcomes Following 12 Minutes Of Untreated Ventricular Fibrillation Ryan A. Coute, Timothy J. Mader, Adam R. Kellogg, Scot A. Millay, Lennard C. Jensen Baystate Medical Center, Springfield, MA Background: According to the three-phase time-sensitive model of VF, the metabolic phase begins after 10 minutes of untreated cardiac arrest. Optimal CPR duration prior to first rescue shock (RS) to maximize the probability of successful defibrillation during this phase remains unknown. Objectives: The purpose of this study was to determine if 3 minutes of CPR prior to first RS is sufficient to achieve ROSC after 12 minutes of untreated VF. Methods: This is a secondary analysis of prospectively collected data from an IACUC-approved protocol. Forty-eight Yorkshire swine (weighing 25–30 kg) were instrumented under anesthesia. VF was electrically induced. After 12 minutes of untreated VF, CPR was initiated (and continued prn) and a standard dose of epinephrine (SDE) (0.01 mg/kg) was given (and repeated every 3 minutes prn). The first RS was delivered after 3 minutes of CPR (and every 3 minutes thereafter prn). A failed RS was followed (in series) by vasopressin (VASO, 0.57 mg/kg), amiodarone (AMIO, 4.3 mg/kg), and sodium bicarbonate (BICARB, 1 mEq/ kg) prn. Resuscitation attempts continued until ROSC was achieved or 20 minutes elapsed without ROSC. The primary outcome measures were ROSC (SBP>80 mmHg for >60s) and survival (SBP>60 mmHg for 20 minutes). Data were analyzed using descriptive statistics. Vasopressor support was available to maintain SBP>60 mmHg following ROSC. Results: ROSC was achieved in 25 of the 48 (52%) animals. Survival occurred in 23 of the 48 (48%) animals. Our findings are summarized in the table. Conclusion: Our data suggest that during the metabolic phase of VF, 3 minutes of CPR and 1 SDE may be insufficient to achieve ROSC on first RS attempt. A longer duration of CPR and/or additional vasopressors may increase the likelihood of successful defibrillation on first attempt. Table - Abstract 168: 169 Effect of Time of Day on Prehospital Care During Out-of-Hospital Cardiac Arrest Sarah K. Wallace, Benjamin S. Abella, Frances S. Shofer, Marion Leary, Robert W. Neumar, C. C. Mechem, David F. Gaieski, Lance B. Becker, Roger A. Band Hospital of the University of Pennsylvania, Philadelphia, PA Background: There are over 300,000 out-of-hospital cardiac arrests (OHCAs) each year in the United States. In most cities, the proportion of patients who achieve prehospital return of spontaneous circulation (ROSC) is less than 10%. The association between time of day and OHCA outcomes in the prehospital setting is unknown. Objectives: We sought to determine whether rates of prehospital ROSC varied by time of day. We hypothesized that night OHCAs would exhibit lower rates of ROSC. Methods: We performed a retrospective review of cardiac arrest data from a large, urban EMS system. Included were all OHCAs occurring in individuals >18 years of age from 1/1/2008 to 12/31/2010. Excluded were traumatic arrests and cases where resuscitation measures were not performed. Day was defined as 7:00 am–6:59 pm, while night was 7:00 pm–6:59 am. We examined the association between time of day and paramedic-perceived prehospital ROSC in unadjusted and adjusted analyses. Variables included age, sex, race, presenting rhythm, AED application by a bystander or first responder, defibrillation, and bystander CPR performance. Analyses were performed using chisquare tests and logistic regression. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S93 Results: Our study population comprised 3742 arrest cases (42.6% at night). Mean age was 65.4 (SD 16.8) years. Males comprised 55.3% of the cohort; 53.0% were black. The unadjusted rate of ROSC was 3.1% at night vs. 4.4% during the day (p = 0.034). Night OHCAs were significantly less likely to receive bystander-initiated CPR than day OHCAs (4.6% vs. 7.5%, p < 0.001). AED application did not vary significantly by time of day. Patients with night OHCAs were significantly younger (64.2 vs. 66.4 years, p < 0.001) and less likely to present in a shockable rhythm (16.8% vs. 20.1%, p = 0.012). After adjusting for significant prehospital and patient-level risk factors, the association between time of day and ROSC was no longer significant (odds ratio [OR] 1.17, 95% CI 0.79–1.73, p = 0.43). Time of day remained significantly associated with bystander CPR (OR 2.59, 95% CI 1.66–4.03, p < 0.001). Conclusion: In adjusted analysis, the significant difference observed between time of day and ROSC no longer persisted. Night arrests remained significantly less likely to receive important prehospital care measures, such as bystander CPR, suggesting they may exhibit more variable prehospital management by lay and first responders. 170 Lung Protective Ventilation is Uncommon among ED Patients Brian M. Fuller1, Nicholas M. Mohr2, Craig A. McCammon3, Rebecca Bavolek1, Kevin Cullison4, Matthew Dettmer1, Jacob Gadbaw5, Sarah Kennedy1, Nicholas Rathert1, Christine Taylor5 1 Washington University School of Medicine, St. Louis, MO; 2University of Iowa Carver College of Medicine, Iowa City, IA; 3Barnes-Jewish Hospital, St. Louis, MO; 4St. Louis University School of Medicine, St. Louis, MO; 5Washington University in St. Louis, St. Louis, MO Background: Endotracheal intubation is frequently performed in the emergency department (ED), but no evidence exists to guide ED mechanical ventilation. Clinical data suggest that ventilator-induced lung injury can occur quickly, and may be preventable with the use of protective ventilatory strategies. Objectives: (1) To describe current ED mechanical ventilation practice, and (2) to determine whether low tidal volume ventilation as the initial mechanical ventilation strategy in ED patients with severe sepsis is associated with lower 28-day in-hospital mortality and more ventilator-free days. Methods: Single-center, retrospective observational cohort study of 250 adult patients with severe sepsis who were intubated in a 90,000-visit urban academic ED between June 2005 and June 2010. All patients with suspected infection and either lactate ‡ 4 mmol/L or SBP £ 90 after a 30 mL/kg fluid bolus were included. Results: Two-hundred forty (96.0%) patients were ventilated initially with volume-targeted ventilation. The median ED length of stay was 5.5 hours (IQR 4.2–7.5). One-hundred twenty-one (48.4%) patients ventilated in the ED with severe sepsis died. Corrected for ideal body weight (IBW), tidal volumes greater than 8 mL/kg IBW were used in 51.2% of patients. Ventilator peak pressures >35 cmH2O were observed in 22% of patients. Obesity was not predictive of high tidal volumes (p = 0.60). Low tidal volume ventilation (<8 mL/kg IBW) was not associated with 28-day in-hospital mortality (42% vs. 51%, p = 0.17) or ventilator-free days (22.2 vs. 20.8 days, p = 0.34). Conclusion: Low tidal volume ventilation was not associated with 28-day in-hospital mortality, but high tidal volume and high peak pressure ventilation are common among intubated ED patients. These findings are hypothesis-generating for future clinical trials, as the role of lung protective ventilation has been largely unexplored in the ED (Grant UL1 RR024992, NIHNCRR). S94 171 2012 SAEM ANNUAL MEETING ABSTRACTS Life-Threatening Etiologies of Altered Mental Status in Children Antonio Muniz Dallas Regional Medical Center, Mesquite, TX Background: The epidemiology of children who present to an emergency department with altered mental status is not well known. The diagnostic evaluation of these children is controversial and is quite variable between physicians. Objectives: The study’s objective was to determine the factors that may place a child at risk for life-threatening causes (LTC) of altered mentation. Methods: Prospective observational evaluation of all children <17 years-old who presented as their primary chief complaint with altered mental status. No children were excluded. Data were analyzed using Stata 11 with continuous variables expressed as means, while categorical variables were summarized as frequencies of occurrence and assessed for statistical significance using chi-square test or Fisher’s exact. Results: There were 203 children, 119 (58.6%) males, 140 (68.9%) African Americans, and 62 (30.5%) Caucasians. Mean age was 8.2 ± 5.9 years-old (95% CI 7.2– 9.1). Hypotension occurred in 6 (2.9%) and hypertension in 22 (10.8%). Tachycardia occurred in 38 (18.7%) and tachypnea in 27 (13.3%). There were 31 different diagnoses. Most common diagnoses were: overdoses 27 (13.3%), ethanol intoxication 22 (10.8%), seizures 20 (9.8%), dehydration 19 (9.3%), medication side effects 13 (6.4%), and psychiatric disorders 12 (5.9%). There were 24 (11.8%) with life-threatening causes (LTC). Hypotension occurred in 4 (1.9%) and tachycardia in 17 (8.3%) in those with life-threatening causes. WBC count was elevated in 20 (9.8%); 15 of these had lifethreatening causes. Tachycardia had a sensitivity 70.5%, specificity 89%, positive predictive value (PPV) 48%, and negative predictive value (NPV) 95.8% for LTC. Tachypnea had a sensitivity 47%, specificity 89%, PPV 50%, and NPV 93% for LTC. Leukocytosis had a sensitivity 66.6%, specificity 97.6%, PPV 80%, and NPV 95.3% for LTC. Combining tachycardia, tachypnea, and leukocytosis predicted those children with LTC with a sensitivity 100%, specificity 83.5%, PPV 44%, and NPV 100%. Conclusion: Altered mentation in children is uncommon and in some children may have life-threatening etiologies. In those with life-threatening causes of altered mentation, abnormal vital signs and leukocytosis were able to detect all these children. 172 The Impact of Early DNR Orders on Patient Care and Outcomes Following Resuscitation from Out of Hospital Cardiac Arrest Derek K. Richardson, Dana Zive, Mohamud Daya, Craig D. Newgard Oregon Health & Science University, Portland, OR Background: Among patients successfully resuscitated from out-of-hospital cardiac arrest (OHCA) and surviving to hospital admission, prior research has suggested that decisions regarding withdrawal of care are best made 48 to 72 hours after admission. Objectives: To evaluate use of early (<24 hours) DNR (do not resuscitate) orders and their association with therapeutic procedure counts and outcomes among patients successfully resuscitated from OHCA surviving to hospital admission. Methods: This was a population-based, retrospective cross-sectional study of adult patients admitted to 332 acute care hospitals in California through the ED with a primary diagnosis of cardiac arrest from 2002 to 2010. Our primary exposure variable was a DNR order recorded within 24 hours of admission. We evaluated in-hospital mortality, ventilation, and procedures based on ICD-9 codes. We assessed associations between early DNR orders and hospital characteristics (size, rural vs. urban, teaching hospital, OHCA volume, ownership, hospital identity). We used descriptive statistics and multivariable models to analyze the sample. Results: 5,212 patients were admitted to California hospitals after OHCA over the 9 year period, of whom 1,692 (32.5%) had a DNR order documented within 24 hours of admission. Compared to post-arrest patients without an early DNR order, these patients were less likely to undergo cardiac catheterization or stenting (1.1% vs. 4.3%), ICD/pacemaker placement (0.1% vs. 1.1%), transfusion (7.6% vs. 11.2%) or ventilation over 96 hours (8.7% vs. 18.6%) (all p < 0.0001). Patients with early DNR orders were less likely to survive to hospital discharge (5.2% vs. 21.6%) with a short length of stay (median 1 day). Variability in hospital rates of early DNR placement was substantial, even after restricting to the top quartile of OHCA volume (10.5%–64.0%); in multivariate models, specific hospital factors were not associated with early DNR order placement. Conclusion: Successfully resuscitated OHCA patients with early DNR orders had fewer therapeutic interventions and lower survival to hospital discharge. There was substantial variability in early DNR use between hospitals. Providers, patients, and surrogates should be aware of the outcomes associated with early DNR placement when making this crucial decision. 173 Does a Simulation Module Educational Intervention Improve Physician Compliance and Reduce Patient Mortality in Severe Sepsis and Septic Shock? Michelle Sergel1, Erik Nordquist1, Brian Krieger1, Alan Senh1, Conal Roche1, Nicole Lunceford1, Tamara Espinoza1, Rashid Kysia1, Bharat Kumar2, Renaud Gueret1, John Bailitz1 1 Cook County (Stroger), Chicago, IL; 2Rush University Research Mentoring Program, Chicago, IL Background: The Institute for Healthcare Improvement (IHI) recommends a ‘‘bundle’’ comprised of a checklist of key clinical tasks to improve physician compliance and reduce patient mortality in severe sepsis and septic shock (S&S). In order to educate and improve compliance with the six-hour goals of the IHI S&S bundle, we conducted a simulation module educational intervention (SMEI) for all emergency department (ED) physician staff. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Objectives: Determine whether a SMEI helps to improve physician compliance with IHI bundle and reduce patient mortality in ED patients with S&S. Methods: We conducted a pre-SMEI retrospective review of four months of ED patients with S&S to determine baseline pre-SMEI physician compliance and patient mortality. We designed and completed a SMEI attended by 25 of 28 ED attending physicians and 28 of 30 ED resuscitation residents. Finally, we conducted a twenty-month post-SMEI prospective study of ongoing physician compliance and patient mortality in ED patients with S&S. Results: In the four month pre-SMEI retrospective review, we identified 23 patients with S&S, with a 61% physician overall compliance and mortality rate of 30%. The average ED physician SMEI multiple-choice pre-test score was 74%, and showed a significant improvement in the post-test score of 94% (p = 0.0003). Additionally, 87% of ED physicians were able to describe three new clinical pearls learned and 85% agreed that the SMEI would improve compliance. In the twenty months of the post-SMEI prospective study, we identified 144 patients with S&S, with a 75% physician overall compliance, and mortality rate of 21%. Relative physician compliance improved 23% (p = 0.0001) and relative patient mortality was reduced by 32% (p < 0.0001) when comparing pre- and post-SMEI data. Conclusion: Our data suggest that a SMEI improves overall physician compliance with the six hour goals of the IHI bundle and reduces patient mortality in ED patients with S&S. 174 Direct Linkage of Low-Acuity Emergency Department Patients with Primary Care: A Pseudo-Randomized Controlled Trial Kelly M. Doran1, Ashley C. Colucci2, Cherry Huang2, Calvin K. Ngai2, Robert A. Hessler2, Andrew B. Wallach3, Michael Tanner3, Lewis R. Goldfrank2, Stephen P. Wall2 1 Robert Wood Johnson Foundation Clinical Scholars Program, Yale University School of Medicine and U.S. Department of Veteran’s Affairs, New Haven, CT; 2Department of Emergency Medicine, New York University, New York, NY; 3General Internal Medicine, Bellevue Hospital Center, New York, NY Background: Having a usual source of primary care is known to improve health. Currently only two-thirds of ED patients have a usual source of care outside the ED, far short of Healthy People 2020’s target of 84%. Prior attempts to link ED patients with primary care have had mixed results. Objectives: To determine if an intervention directly linking low-acuity patients with a primary care clinic at the time of an ED visit could lead to future primary care linkage. Methods: DESIGN: Pseudo-randomized controlled trial. SETTING: Urban safety-net hospital. SUBJECTS: Adults presenting to the ED 1/07–1/08 for select problems a layperson would identify as low-acuity. Patients were excluded if they arrived by EMS, had a PCP outside our • www.aemj.org S95 hospital, were febrile, or the triage nurse felt they needed ED care. Consecutive patients were enrolled weekday business hours when the primary care clinic was open. Patients were assigned to usual care in the ED if a provider was ready to see them before they had completed the baseline study survey. Otherwise they were offered the intervention if a clinic slot was available. INTERVENTION: Patients agreeing to the intervention were escorted to a primary care clinic in the same hospital building. They were assigned a personal physician and given an overview of clinic services. A patient navigator ensured patients received timely same-day care. Intervention group patients could refuse the intervention and instead remain in the ED for care. Both clinic and ED patients were given follow-up clinic appointments, or a phone number to call for one, as per usual provider practice. ANALYSIS: The main outcome measure was primary care linkage, defined as having one or more primary care clinic visits within a year of the index ED visit for patients with no prior PCP. Results: 1,292 patients were potentially eligible and 853 were enrolled (662 intervention and 191 controls). Groups had similar baseline characteristics. Nearly 75% in both groups had no prior PCP. Using an intention to treat analysis, 50.3% of intervention group patients with no prior PCP achieved successful linkage (95%CI 45.7–54.9%) vs. 36.9% of the control group (95%CI 28.9–45.4%). Conclusion: A point-of-care program offering low-acuity ED patients the opportunity to instead be seen at the hospital’s primary care clinic resulted in increased future primary care linkage compared to standard ED referral practices. 175 Ambulatory Care Sensitive Conditions And The Likelihood Of 30-Day Hospital Readmissions Through The ED Joseph A. Tyndall, Wei Hou, Doug Dame, Donna Carden University of Florida, Gainesville, FL Background: Emergency department (ED) visits have increased dramatically over the last decade along with growing concerns about available resources to support care delivery within this safety net. Recent evidence suggests that 30-day hospital readmissions through the ED are also rising. Hospital admission for one of 14 ambulatory care sensitive conditions (ACSC) is thought to be preventable with timely and effective outpatient care. However, the risk of 30-day hospital readmission in patients with an ACSC is unexplored. Objectives: The objective of this study was to examine 30-day readmission rates through the ED for patients discharged from an inpatient setting with an ACSC compared to patients without a preventable hospitalization. Methods: Adult ED admissions between Jan 1, 2006 and August 1, 2010 were evaluated. Administrative data were processed using AHRQ’s QI Windows Application Version 4.1a and SAS to flag specific diagnosis codes that qualified as an ACSC. Statistical analysis was performed using SAS v9.2. Chi-square testing was used to test for significant differences in 30-day readmission through the ED in patients previously admitted for an S96 2012 SAEM ANNUAL MEETING ABSTRACTS ACSC compared to those admitted for a non-preventable condition. Charlson co-morbidity index was applied to discharge data to control for coexisting illness. Cumulative hazard curves for readmission within 30 days were generated based on Nelson-Aalon estimators and comparison was done by a log-rank test. A p value £ 0.05 was considered significant. Results: 78,982 index admissions were analyzed with 5475 (6.9%) representing admissions for an ACSC. Of 12,574 readmissions within 30 days of index hospitalization, 990 represented admission for an ACSC. Patients with ACSC-associated admissions were more likely to be readmitted within 30 days than patients with nonpreventable admissions (p < 0.0001). Patients with an ACSC-associated admission and a Charlson comorbidity index of >1 had the highest rate of readmission within 30 days of a hospital discharge. Conclusion: Patients with ACSC-associated hospital admissions are more likely to present to the ED and be readmitted within 30 days. A Charlson comorbidity index of >1 exacerbates the risk of a 30-day readmission. 176 Visit Urgency Between Frequent Emergency Department Users In A Large Metropolitan Region Network Edward M. Castillo, Gary M. Vilke, James P. Killeen, Jesse J. Brennan, Chan T. Chan University of California, San Diego, San Diego, CA Background: There is growing focus on so-called ‘‘Hot spotter’’ or frequent flier (FF) patients who are high utilizers of health care resources, particularly acute care services, from both care quality and cost standpoints. Objectives: We sought to evaluate FF use of emergency services in a large, metropolitan area served by multiple hospital emergency departments (EDs). Methods: We conducted a region-wide retrospective cohort study of all visits to 16 hospital EDs in a metropolitan region of 3.2 million from 1/1/08–12/31/10 using data from the California Office of Statewide Health Planning and Development (OSHPD) inpatient and ED dataset. Data included demographics and visit specific information. FF patients were defined as those having at least 6 ED visits and Super Users (SU) were defined as having at least 24 ED visits within any 12-month span of the study period. Visits were compared between groups as to primary reason for the ED visit and visit urgency using the ‘‘Billings Algorithm’’ to determine ED visit necessity defined as non-emergent (NE), emergent: primary care treatable (EPCT), emergent: preventable (EP), and emergent: not preventable (ENP). The probability of each urgency level is reported and compared for differences. Results: During the study period, 925,719 patients with a valid patient identifier were seen in the ED resulting in 2,016,537 total visits. Of these, 28,969 patients (3%) were classified as FF and responsible for 347,004 of all visits (17.2%), and 1,261 patients (0.1%) were SU and responsible for 77,080 (3.8%) of all visits. FF and SU were more likely than all other ED patients to present with mental health (5.2% vs 5.6% vs 2.3%, respectively) and drug/alcohol-related problems (3.2% vs 5.7% vs 1.2%, respectively). With respect to visit necessity, visits by FF and SU were more likely than other ED patients to be classified as NE (21.4%, 23.9%, 18.9%, respectively), whereas there were no other differences in the other necessity classifications. All comparisons were statistically significant. Conclusion: In this 3-year study conducted in a large metropolitan area, FF and SU patients who are seen across the EDs in the region present with mental health, drug/alcohol, and non-emergent problems more frequently than other ED patients. 177 Return Visits to the Emergency Department and Costs: A Multi-State Analysis Reena Duseja, Ted Clay, Mitzi Dean, R. Adams Dudley UCSF, San Francisco, CA Background: Return visits to the emergency department (ED) or hospital after an index ED visit affect patient care and place a strain on the U.S. health system. However, little is known at a multi-state level about the rates at which ED patients return for care, the characteristics of patients who return, or the costs associated with return visits. Objectives: The primary objectives were to determine the (1) rates of return for ED care or for hospitalization after an index ED visit, (2) characteristics of patients who return, and (3) costs associated with those return visits. Methods: Data were used from six geographically dispersed states (Arizona, California, Florida, Nebraska, Utah, and Hawaii) that had available linkage files, for the years 2006–2008, in the Healthcare Cost and Utilization Project (HCUP) State Emergency Department Databases (SEDD) and State Inpatient Databases (SID) databases. Patients 17 or younger were excluded. Costs were calculated using publicly available Medicare rates for the year 2008, using CPT codes for ED visits, and DRG codes for inpatient admissions. Results: Among 37,363,104 records from 2006–2008, 34,168,691 (91.5%) had encrypted identifiers to allow linkage. The largest number of return visits occurred between days 0 and 3; the cumulative return visit rate was 8.2%, 5.8% of those resulting in a patient being discharged back to the community. The largest percentage of return visits within 3 days and discharged were between the ages 18–44 (60%), while 26.5% return visits that resulted in admission were over 65 years of age. Patients with Medicaid had the highest rate of return to the ED within 3 days (10.5%), while those with Medicare had the highest rate of admission if they returned (3.9%). The Clinical Classification System (CCS) groups with the highest 3-day revisit rates were skin and subcutaneous tissue infections (20.7%), and administrative/social visits accounted for 10.1% of ED returns. Within 3 days, ED cumulative costs were $550 million, with return visits leading to an inpatient visit totaling $4 billion. Conclusion: Using a population-level, longitudinal, and multi-state analysis, the rate of return visits within 3 days is higher than previously reported, with nearly 1 in 12 returning back to the ED. We also provide the first estimation of health care costs for ED revisits. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 178 Can Patients Accurately Assess Their Own Acuity? Findings From A Large Scale Emergency Department Patient Experience Survey Alaina Aguanno1, Timothy Cooke2, John Cowell2, Andrew McRae1, Brian Rowe3, Eddy Lang1 1 University of Calgary, Calgary, AB, Canada; 2 Health Quality Council of Alberta, Calgary, AB, Canada; 3University of Alberta, Edmonton, AB, Canada Background: The ability of patients to accurately determine their level of urgency is important in planning strategies that divert away from EDs. In fact, an understanding of patient self-triage abilities is needed to inform health policies targeting how and where patients access acute care services within the health care system. Objectives: To determine the accuracy of a patient’s self-assessment of urgency compared against triage nurses. Methods: Setting: ED patients are assigned a score by trained nurses according to the Canadian Emergency Department Triage and Acuity Scale (CTAS). We present a cross-sectional survey of a random patient sample from 12 urban/regional EDs conducted during the winters of 2007 and 2009. This previously validated questionnaire, based on the British Healthcare Commission Survey, was distributed according to a modified Dillman protocol. Exclusion criteria consisted of: age 0– 15 years, left prior to being seen/treated, died during ED visit, no contact information, presented with a privacy-sensitive case. Alberta Health Services provided linked non-survey administrative data. Results: 21,639 surveys distributed with a response rate of 46%. Patients rated health problems as life-threatening (6%), possibly life-threatening (22%), urgent (30%), somewhat urgent (37%), or not urgent (5%). Triage nurses assigned the same patients CTAS scores of I (<1%), II (20%), III (45%), IV (29%) or V (5%). Patients self-rated their condition as 3 or 4 points less urgent than the assigned CTAS score (<1% of the time), 2 points less urgent (5%), 1 point less urgent (25%), exactly as urgent (38%), 1 point more urgent (24%), 2 • www.aemj.org S97 points more urgent (7%), or 3 or 4 points more urgent (1%, respectively). Among CTAS I or II patients, 54% described their problem as life-threatening/possibly life-threatening, 26% as urgent (risk of permanent damage), 18% as urgent (needed to be seen that day), and 2% as not urgent (wanted to be but did not need to be seen that day). Conclusion: The majority of ED patients are generally able to accurately assess the acuity of their problem. Encouraging patients with low-urgency conditions to self-triage to lower-acuity sources of care may relieve stress on EDs. However, physicians and patients must be aware that a small minority of patients are unable to self-triage safely. 179 Intervention to Integrate Health and Social Services for Frequent ED Users with Alcohol Use Disorders Ryan McCormack, Lily Hoffman, Lewis Goldfrank NYU School of Medicine/Bellevue Hospital, New York, NY Background: The ED is a point of frequent contact for medically vulnerable, chronically homeless patients with alcohol use disorders, or chronic public inebriates (CPI). Despite this population’s exposure to health and social agencies, its outcomes suffer due, in part, to lack of stable housing and fragmented, ‘treat and street’ medical care. Objectives: NYU School of Medicine and the Bellevue Hospital Center ED partnered with the Department of Homeless Services (DHS) to implement a multifaceted pilot initiative. This integration of services is hypothesized to improve access to housing and comprehensive medical care resulting in reduced costly ED and inpatient admissions, and homelessness. Engaging the ED as a point of intervention, a cohort of CPIs received needs assessments, enhanced care management, and coordination with DHS outreach. Methods: CPIs were identified primarily through an administrative database search and chart reviews. At the time of this 10-month analysis, 20 of the 56 patients who met inclusion criteria were enrolled. Enrolled S98 2012 SAEM ANNUAL MEETING ABSTRACTS patients had a minimum of 20 ED visits in a 24-month period with at least one visit within 5 months of the pilot commencement in January 2011 and met the DHS standard for chronic homelessness. Preference was given to those with greater visit frequency, co-morbidities, or staff referral. The intervention for enrolled patients included the ongoing implementation of individualized multidisciplinary action plans, case management, and coordination with the housing outreach team upon discharge. Results: Eighteen of the 20 enrolled patients were placed in housing. After first housing placement (mean length, 4.7 months), monthly ED and inpatient use declined 48% and 40%, respectively. ED and inpatient use by the nonenrolled remained stable throughout the study period. Prior to intervention, hospital use had increased over time for the enrolled patients (Figures 1,2). Conclusion: ED-based collaboration amongst medical and social services for a small cohort of CPIs resulted in housing placements and reduced ED and inpatient visits. While promising, the results of this interim pilot data are limited by the non-random sampling method, power, duration, and singular location. Further study is needed to determine the intervention’s effect on public health expenditures and patient outcomes. 180 A Comparison Of Hand-On-Syringe Versus Hand-On-Needle Technique For UltrasoundGuided Nerve Blocks Brian Johnson, Arun Nagdev, Michael Stone, Andrew Herring Alameda County Medical Center, Oakland, CA Background: Ultrasound-guided regional nerve blocks are increasingly used in emergency care. The hand-onsyringe (HS) needling technique is ideally suited to the fast-paced ED setting because it allows a single operator to perform the block without assistance. In the anesthesia literature, the HS technique is commonly assumed to provide less needle control than the alternative two operator hand-on-needle (HN) technique; however, this assumption has never been directly tested. Objectives: To compare needle control under ultrasound guidance by emergency medicine (EM) residents using HN and HS techniques on a standardized phantom simulation model. Methods: This prospective, randomized study evaluated task performance on a simulated ultrasoundguided nerve block phantom model comparing HN and HS techniques. Participants were EM residents at a large, urban, academic hospital. Each participant performed a set of structured needling maneuvers (both simple and difficult) on a standardized phantom simulator. Parameters of task performance and needle control were evaluated including time to task completion, needle visualization during advancement, and placement of the needle tip at target. Additionally, resident technique preference was assessed using a post-task survey. Results: Sixty tasks performed by 10 EM residents were analyzed. The HN technique did not demonstrate superior control compared to the HS technique. There was no difference in time to complete the simple model (HN vs. HS, 18 seconds vs. 18 seconds, p = 0.92). There was no difference in time to complete the difficult model (HN vs. HS, 56 seconds vs. 50 seconds, p = 0.63). There was no difference in first pass success between the HN (8%) and HS (12%) technique. Needle visualization was similar for both techniques. There were more instances of advancing the needle tip into the target in the HN (4/60) versus the HS (1/60) technique. Most residents preferred the HS technique (60%), 40% preferred the HN technique, and 10% had no preference. Conclusion: For EM residents learning ultrasoundguided nerve blocks, the HN technique did not provide superior needle control. Our results suggest that the single-operator HS technique provides equivalent needle control when compared to the two-operator HN technique. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 181 Does Level of Training Matter When EM Residents Provide Patient Care While Distracted? Dustin Smith1, Jeffrey Cukor2, Gloria Kuhn3, Daniel G. Miller4 1 Loma Linda University Medical Center, Loma Linda, CA; 2University of Massachusetts, Worcester, MA; 3Wayne State University, Detroit, MI; 4University of Iowa Hospitals and Clinics, Iowa City, IA • 182 Background: Work interruptions during patient care have been correlated with error. It is unknown at what level of training EM physicians become competent to execute patient care tasks despite distractions. Objectives: We hypothesized that level of training affects EM resident physicians’ ability to execute required patient care tasks in the simulation environment when a second patient with a STEMI interrupts a septic shock case. Methods: The study was a multisite prospective observational cohort study. The study population consisted of EM residents in their first 3 years of EM training. Data were collected spring 2011. Each subject performed a standardized simulated encounter by evaluating and treating a patient in septic shock. At a predetermined point in every case the septic patient became hypotensive. Shortly thereafter, the subject was given a STEMI ECG for a separate chest pain patient in triage and required to verbalize an interpretation and action. Data were collected on the subjects’ treatment of the septic shock patient with acceptable interventions defined as administration of appropriate antibiotics and fluids per EGDT and recognition of STEMI. Results: 91 subjects participated (30 PGY1s, 32 PGY2s, and 29 PGY3s). 87 properly managed the patient with septic shock (90.0% PGY1s, 100% PGY2s, 96.6% PGY 3s; p-value 0.22). Of the 87 who successfully managed the septic shock, 80 correctly identified STEMI on the simulated STEMI patient (86.7% PGY1s, 96.9% PGY2s, 93.1% PGY3s; p-value 0.35). PGY2s were 5.39 times more likely than PGY1s to correctly identify a STEMI (CI 0.56–51.5). PGY3s were 2.26 times more likely than PGY1s to correctly identify a STEMI (CI 0.38–13.5). Conclusion: When management of septic shock was interrupted with a STEMI ECG in simulation we observed no significant difference in completion of early goal directed therapy or recognition of STEMI when compared across different years of training. www.aemj.org S99 Developing a Perfused Cadaver Training Model for Invasive Lifesaving Procedures: Uncontrolled Hemorrhage Robert A. De Lorenzo, John A. Ward, Syed H. Husaini, Allison L. Abplanalp, Suzanne McCall Brooke Army Medical Center, Fort Sam Houston, TX Background: This is a project to develop a perfused human cadaver model to train military medical personnel in invasive lifesaving procedures. The simulation evaluated was hemorrhage control, and a swine carcass model served as a prototype. Objectives: Develop a perfused swine carcass model for hemorrhage control which can be used to develop a perfused human cadaver model for military medical training purposes. Methods: Carcasses were exsanguinated, eviscerated, and refrigerated overnight. The next day, arteries supplying the neck and iliac regions were cannulated. Regional circulations were perfused with red fluid using a pulsatile blood pump. Surgical wounds were made in the neck, thigh, and popliteal fossa to open major arteries and simulate hemorrhage. QuickTime digital video files were recorded for evaluation. Results: Twenty-four, 30 to 60 kg female swine carcasses were used to develop the model and one fresh frozen human cadaver was used to test the concept. Web belts and combat application tourniquets (CAT) were used for validation testing. After establishing simulated hemorrhage, pulse pressure oscillations were observed. The tourniquet was tightened until hemorrhage stopped. When the tourniquet was released, blood spurted from the injured artery as hydrostatic pressure decayed. Pressure and flow were recorded in three animals (see table). The concept was proof-tested in a single fresh frozen human cadaver with perfusion through the femoral artery and hemorrhage from the popliteal artery. The results were qualitatively and quantitatively similar to the swine carcass model. Conclusion: A perfused swine carcass can simulate exsanguinating hemorrhage for training purposes and serves as a prototype for a fresh-frozen human cadaver model. Additional research and development are required before the model can be widely applied. Table - Abstract 182: Pressure and flow measurements from perfused swine carcass models Pressure (mmHg) Animal 1 2 3 3 Perfused Artery Femoral a. Thoracic aorta Thoracic aorta Thoracic aorta Flow (L/min) Statham Site Low High Probe Site Low High Rate (strokes/min) Femoral a. Thoracic aorta Thoracic aorta Abdominal aorta 40 73 57 80 215 84 99 85 Thoracic aorta Thoracic aorta Thoracic aorta Abdominal aorta 0.00 0.00 0.07 0.00 0.60 1.03 0.78 1.20 34.3 36.9 S100 183 2012 SAEM ANNUAL MEETING ABSTRACTS Development of a Simulation-Enhanced Multidisciplinary Teamwork Training Program in a Pediatric Emergency Department Susan Duffy, Linda Brown, Frank Overly Brown University, Providence, RI Background: In the pediatric emergency department (PED), clinicians must work together to provide safe and effective care. Crisis resource management (CRM) principles have been used to improve team performance in high-risk clinical settings, while simulation allows practice and feedback of these behaviors. Objectives: To develop a multidisciplinary educational program in a PED using simulation-enhanced teamwork training to standardize communication and behaviors and identify latent safety threats. Methods: Over 6 months a workgroup of physicians and nurses with experience in team training and simulation developed an educational program for clinical staff of a tertiary PED. Goals included: create a didactic curriculum to teach the principles of CRM, incorporate principles of CRM into simulation-enhanced team training in-situ and center-based exercises, and utilize assessment instruments to evaluate for teamwork, completion of critical actions, and presence of latent safety threats during in-situ SIM resuscitations. Results: During Phase I, 130 clinicians, divided into teams, participated in 90-minute pre-training assessments of PALS-based in-situ simulations. In Phase II, staff participated in a 6-hour curriculum reviewing key CRM concepts, including team training exercises utilizing simulation and expert debriefing. In Phase III, staff participated in post-training 90 minute teamwork and clinical skills assessments in the PED. In all phases, critical action checklists (CAC) were tabulated by simulation educators. In-situ simulations were recorded for later review using the assessment tools. After each simulation, educators facilitated discussion of perceptions of teamwork and identification of systems issues and latent hazards. Overall, 54 in-situ simulations were conducted capturing 97% of the physicians and 84% of the nurses. CAC data were collected by an observer and compared to video recordings. Over 20 significant systems issues, latent hazards, and knowledge deficits were identified. All components of the program were rated highly by 90% of the staff. Conclusion: A workgroup of PEM, simulation, and team training experts developed a multidisciplinary team training program that used in-situ and centerbased simulation and a refined CRM curriculum. Unique features of this program include its multidisciplinary focus, the development of a variety of assessment tools, and use of in-situ simulation for evaluation of systems issues and latent hazards. This program was tested in a PED and findings will be used to refine care and develop a sustainment program while addressing issues identified. 184 ACLS Training: Does High-Fidelity Simulation Matter? Lauren N. Weinberger Hospital of the University of Pennsylvania, Philadelphia, PA Background: An emerging area of simulation research seeks to identify what type of simulation technology offers the greatest benefit to the learner. Our study intends to quantitatively assess the performance of learners while varying the fidelity of simulation technology used to teach Advanced Cardiac Life Support (ACLS). Objectives: Our hypothesis is that participants trained on high-fidelity mannequins will perform better than participants trained on low-fidelity mannequins on both the ACLS written exam and in performance of critical actions during megacode testing. Methods: The study was performed in the context of an ACLS Initial Provider course for new PGY1 residents at the Penn Medicine Clinical Simulation Center and involved three training arms: 1) low fidelity (low-fi): Torso-Rhythm Generator; 2) mid-fidelity (mid-fi): Laerdal SimMan turned OFF; and 3) high-fidelity (high-fi): Laerdal SimMan turned ON. Training in each arm of the study followed standard AHA protocol. Educational outcomes were evaluated by written scores on the ACLS written examination and expert rater reviews of ACLS megacode videos performed by trainees during the course. A sample of 54 subjects were randomized to one of the three training arms: low-fi (n = 18), mid-fi (n = 18), or high-fi (n = 18). Results: Statistical significance across the groups was determined using analysis-of-variance (ANOVA). The three groups had similar written pre-test scores [low-fi 0.4 (0.1), mid-fi 0.5 (0.1), and high-fi 0.4 (0.2)] and written post-test scores [low-fi 0.9 (0.1), mid-fi 0.9 (0.1), and high-fi 0.8 (0.1)]. Similarly, test improvement was not significantly different. After completion of the course, high-fi subjects were more likely to report they felt comfortable in their simulator environment (p = 0.005). Low-fi subjects were less likely to perceive a benefit in ACLS training from high-fi technology (p < 0.001). ACLS Instructors were not rated significantly different by the subjects using the Debriefing Assessment for Simulation in Healthcareª (DASH) student version except for element 6, where the high-fi group subjects reported lower scores (6.1 vs 6.6 and 6.7 in the other groups, p = 0.046). Conclusion: Overall, there was no difference among the three groups in test scores or test improvement. Evaluations of the training modality were different in regards to user comfort and utility of simulator training. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 185 Using Heart Rate Variability as a Physiologic Marker of Stress During the Performance of Complex Tasks Douglas Gallo1, Walter Robey III1, Carmen Russoniello2, Matthew Fish2, Kori Brewer1 1 Pitt County Memorial Hospital, Greenville, NC; 2 East Carolina University, Greenville, NC Background: Managing critically ill patients requires that complex, high-stress tasks be performed rapidly and accurately. Clinical experience and training have been assumed to translate into increased proficiency with these tasks. Objectives: We sought to determine if stress associated with the performance of a complex procedural task can be affected by level of medical training. Heart rate variability (HRV) is used as a measure of autonomic balance, and therefore an indicator of the level of stress. Methods: Twenty-one medical students and emergency medicine residents were enrolled. Participants performed airway procedures on an airway management trainer. HRV data were collected using a continuous heart rate variability monitoring system. Participant HRV was monitored at baseline, during the unassisted first attempt at endotracheal intubation, during supervised practice, and then during a simulated respiratory failure clinical scenario. Standard deviation of beat to beat variability (SDNN), very low frequency (VLF), total power (TP), and low frequency (LF) was analyzed to determine the effect of practice and level of training on the level of stress. A Cohen’s d test was used to determine differences between study groups. Results: SDNN data showed that second-year residents were less stressed during all stages than were fourthyear medical students (avg d = 1.12). VLF data showed third-year residents exhibited less sympathetic activity than did first-year residents (avg d = )0.68). The opportunity to practice resulted in less stress for all participants. TP data showed that residents had a greater degree of control over their autonomic nervous system (ANS) than did medical students (avg d = 0.85). LF data showed that subjects were more engaged in the task at hand as the level of training increased indicating autonomic balance (avg d = 0.80). Conclusion: Our HRV data show that stress associated with the performance of a complex procedural task is reduced by increased training. HRV may provide a quantitative measure of physiologic stress during the learning process and thus serve as a marker of when a subject is adequately trained to perform a particular task. 186 An Experimental Comparison of Endotracheal Intubation During Ongoing CPR With Manual Compression versus Automated Compression Bob Cambridge, Amy Chelin, Austin Lamb, John Hafner OSF St. Francis Medical Center, Peoria, IL Background: The ACLS recommendations for CPR are that chest compressions should be uninterrupted to • www.aemj.org S101 help maintain perfusion pressures. Every time CPR stops the perfusion pressure drops and it takes time to get back up to a clinically helpful level. One common reason CPR is stopped in the ED or in the prehospital setting is to place a definitive airway through tracheal intubation. Intubation during ongoing compressions is difficult due to the randomness of tracheal motion secondary to the chest compressions. Objectives: We seek to examine whether intubation during CPR can be done as efficiently as intubation without ongoing CPR. The hypothesis is that the predictable movement of an automated chest compression device will make intubation easier than the random movement from manual CPR. Methods: The project was an experimental controlled trial and took place in the emergency department at a tertiary referral center in Peoria, Illinois. Emergency medicine residents, attendings, paramedics, and other ACLS trained staff were eligible for participation. In randomized order, each participant attempted intubation on a mannequin with no CPR ongoing, during CPR with a human compressor, and during CPR with an automatic chest compression device (Physio Control Lucas 2). Participants could use whichever style laryngoscope they felt most comfortable with and they were timed during the three attempts. Success was determined after each attempt. Results: There were 43 participants in the trial. The success rate in the control group and the automated CPR group were both 88% (38/43) and the success rate in the manual CPR group was 74% (32/43). The differences in success rates were not statistically significant (p = 0.99 and p = 0.83). The automated CPR group had the fastest average time (13.6 sec; p = 0.019). The mean times for intubation with manual CPR and no CPR were not statistically different (17.1 sec, 18.1 sec; p = 0.606). Conclusion: The success rate of tracheal intubation with ongoing chest compression was the same as the success rate of intubation without CPR. Although intubation with automatic chest compression was faster than during other scenarios, all methods were close to the 10 second timeframe recommended by ACLS. Based on these findings, it may not always be necessary to hold CPR to place a definitive airway; however, further studies will be needed. 187 Fibroblast Growth Factor 2 Affects Vascular Remodeling After Acute Myocardial Infarction Stacey L. House, Thomas Belanger, Carla Weinheimer, David Ornitz Washington University in St. Louis, St. Louis, MO Background: After acute myocardial infarction, vascular remodeling in the peri-infarct area is essential to provide adequate perfusion, prevent additional myocyte loss, and aid in the repair process. We have previously shown that endogenous fibroblast growth factor 2 (FGF2) is essential to the recovery of contractile function and limitation of infarct size after cardiac S102 2012 SAEM ANNUAL MEETING ABSTRACTS ischemia-reperfusion (IR) injury. The role of FGF2 in vascular remodeling in this setting is currently unknown. Objectives: Determine the role of endogenous FGF2 in vascular remodeling in a clinically relevant, closed-chest model of acute myocardial infarction. Methods: Mice with a targeted ablation of the Fgf2 gene (Fgf2 knockout) and wild type controls were subjected to a closed-chest model of regional cardiac IR injury. In this model, mice were subjected to 90 minutes of occlusion of the left anterior descending artery followed by reperfusion for either 1 or 7 days. Immunofluorescence was performed on multiple histological sections from these hearts to visualize capillaries (endothelium, anti-CD31 antibody), larger vessels (venules and arterioles, antismooth muscle actin antibody), and nuclei (DAPI). Digital images were captured, and multiple images from each heart were measured for vessel density and vessel size. Results: Sham-treated Fgf2 knockout and wild type mice show no differences in capillary or vessel density suggesting no defect in vessel formation in the absence of endogenous FGF2. When subjected to closed-chest regional cardiac IR injury, Fgf2 knockout hearts had normal capillary and vessel number and size in the peri-infarct area after 1 day of reperfusion compared to wild type controls. However, after 7 days, Fgf2 knockout hearts showed significantly decreased capillary and vessel number and increased vessel size compared to wild type controls (p < 0.05). Conclusion: These data show the necessity of endogenous FGF2 in vascular remodeling in the peri-infarct zone in a clinically relevant animal model of acute myocardial infarction. These findings may suggest a potential role for modulation of FGF2 signaling as a therapeutic intervention to optimize vascular remodeling in the repair process after myocardial infarction. 188 The Diagnosis of Aortic Dissections by ED Physicians is Rare Scott M. Alter, Barnet Eskin, John R. Allegra Morristown Medical Center, Morristown, NJ Background: Aortic dissection is a rare event. The most common symptom of dissection is chest pain, but chest pain is a frequent emergency department (ED) chief complaint and other diseases that cause chest pain, such as acute coronary syndrome and pulmonary embolism, occur much more frequently. Furthermore, 20% of dissections are without chest pain and 6% are painless. For all these reasons, diagnosing dissection can be difficult for the ED physician. We wished to quantify the magnitude of this problem in a large ED database. Objectives: Our goal was to determine the number of patients diagnosed by ED physicians with aortic dissections compared to total ED patients and to the total number of patients with a chest pain diagnosis. Methods: Design: Retrospective cohort. Setting: 33 suburban, urban, and rural New York and New Jersey EDs with annual visits between 8,000 and 75,000. Participants: Consecutive patients seen by ED physicians from January 1, 1996 through December 31, 2010. Observations: We identified aortic dissections using ICD-9 codes and chest pain diagnoses by examining all ICD-9 codes used over the period of the study and selecting those with a non-traumatic chest pain diagnosis. We then calculated the number of total ED patients and chest pain patients for every aortic dissection diagnosed by emergency physicians. We determined 95% confidence intervals (CIs). Results: From a database of 9.5 million ED visits, we identified 782 (0.0082%) aortic dissections, or one for every 12,200 (95% CI 11,400 to 13,100) visits. The mean age of aortic dissection patients was 58 ± 19 years and 57% were female. Of the total visits there were 763,000 (8%) with a chest pain diagnosis. Thus there is one aortic dissection diagnosis for every 980 (95% CI 910 to 1,050) chest pain diagnoses. Conclusion: The diagnosis of aortic dissections by ED physicians is rare. An ED physician seeing 3,000 to 4,000 patients a year would diagnose an aortic dissection approximately once every 3 to 4 years. An aortic dissection would be diagnosed once for approximately every 1,000 ED chest pain patients. 189 Prevalence and ECG Findings for Patients with False-positive Cardiac Catheterization Laboratory Activation among Patients with Suspected ST-Segment Elevation Myocardial Infarction Kelly N. Sawyer1, Audra L. Robinson2, Charlotte S. Roberts2, Michael C. Kurz2, Michael C. Kontos2 1 William Beaumont Hospital, Royal Oak, MI; 2 Virginia Commonwealth University, Richmond, VA Background: Patients presenting with an initial electrocardiogram (ECG) consistent with ST-elevation myocardial infarction (STEMI) represent a cardiovascular emergency. Low levels of false cardiac catheterization laboratory (CCL) activation are acceptable to ensure maximum sensitivity. Objectives: To describe the patients with false positive CCL activation presenting to our institution with potential STEMI. Methods: This study was a case series conducted from June 2006 to December 2010 of consecutive CCL activations from the ED at our urban, academic, tertiary care hospital. Patients were excluded if they suffered a cardiac arrest, were transferred from another hospital, or if the CCL was activated for an inpatient or from EMS in the field. FP CCL activation was defined as 1) a patient for whom activation was cancelled in the ED and ruled out for MI or 2) a patient who went to catheterization but no culprit vessel was identified and MI was excluded. ECGs for FP patients were classified using standard criteria. Demographic data, cardiac biomarkers, and all relevant time intervals were collected according to an on-going quality assurance protocol. Results: A total of 506 CCL activations were reviewed, with 68% male, average age 57, and 59% black. There were 210 (42%) true STEMIs and 86 (17%) FP activations. There were no significant differences between the FP patients who did and did not have catheterization. For those FP patients who had a catheterization ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 (13%), ‘‘door to page’’ and ‘‘door to lab’’ times were significantly longer than the STEMI patients (see table), but there was substantial overlap. There was no difference in sex or age, but FP patients were more likely to be black (p = 0.02). A total of 82 FP patients had ECGs available for review; findings included anterior elevation with convex (21%) or concave (13%) elevation, ST elevation from prior anterior (10%) or inferior (11%) MI, pericarditis (16%), presumed new LBBB (15%), early repolarization (5%), and other (9%). Conclusion: False CCL activation occurred in a minority of patients, most of whom had ECG findings warranting emergent catheterization. The rate of false CCL activation appears acceptable. Table - Abstract 189: Time Intervals (minutes) STEMI Patients (n = 210) Median (IQR) FP Patients (n = 86) Median (IQR) p-value Door to ECG Door to Page Door to CCL 9 (5, 15) 14.5 (8, 23) 45 (32, 59) 12 (7, 17.5) 21.5 (10.5, 41) 55 (40, 70)* 0.281 0.001 0.004 *n = 59 patients who went to the CCL and MI was excluded 190 An Evaluation of an Atrial Fibrillation Clinic for the Follow-up of Patients Presenting to the Emergency Department with Newly Diagnosed or Symptomatic Arrhythmia Brandon Hone1, Eddy Lang2, Anne Gillis2, Renee Vilneff2, Trevor Langhan2, Russell Quinn2, Vikas Kuriachin2, Laurie Burland2, Beverly Arnburg2 1 University of Alberta, Edmonton, AB, Canada; 2 University of Calgary, Calgary, AB, Canada Background: Atrial fibrillation (AF) is the most common cardiac arrhythmia treated in the ED, leading to high rates of hospitalization and resource utilization. Dedicated atrial fibrillation clinics offer the possibility of reducing the admission burden for AF patients presenting to the ED. While the referral base for these AF clinics is growing, it is unclear to what extent these clinics contribute to reducing the number of ED visits and hospitalizations related to AF. Objectives: To compare the number of ED visits and hospitalizations among discharged ED patients with a primary diagnosis of AF who followed up with an AF clinic and those who did not. Methods: A retrospective cohort study and medical records review including three major tertiary centres in Calgary, Canada. A sample of 600 patients was taken representing 200 patients referred to the AF clinic from the Calgary Zone EDs and compared to 400 matched control ED patients who were referred to other providers for follow-up. The controls were matched for age and sex. Inclusion criteria included patients over 18 years of age, discharged during the index visit, and seen by the AF clinic between January 1, 2009 and October 25, 2010. Exclusion criteria included non-residents • www.aemj.org S103 and patients hospitalized during the index visit. The number of cardiovascular-related ED visits and hospitalizations was measured. All data are categorical, and were compared using chi-square tests. Results: Patients in the control and AF clinic cohorts were similar for all baseline characteristics except for a higher proportion of first episode patients in the intervention arm. In the six months following the index ED visit, 55 study group patients (27.5%) visited an ED on 95 occasions, and 12 (6%) were hospitalized on 16 occasions. Of the control group, 122 patients (30.5%) visited an ED on 193 occasions, and 44 (11%) were hospitalized on 55 occasions. Using a chi-square test we found no significant difference in ED visits (p = 0.5063) or hospitalizations (p = 0.0664) between the control and AF clinic cohorts. Conclusion: Based on our results, referral from the ED to an AF clinic is not associated with a significant reduction in subsequent cardiovascular related ED visits and hospitalizations. Due to the possibility of residual confounding, randomized trials should be performed to evaluate the efficacy of AF clinics. Table - Abstract 190: Cardiovascular-related ED visits and hospitalizations after index ED visit AF Clinic Group (N = 200) Subsequent CV related ED visits Number (%) of patients 55 (27.5) Total number of visits 95 Number (%) of patients with 51 (25.5) 1 or more arrhythmic event leading to ED visit Subsequent CV related hospitalizations Number (%) of patients 12 (6) Total number of admissions 16 191 Usual Care (N = 400) P-value 122 (30.5) 193 98 (24.5) 44 (11) 55 0.5063 0.8673 0.0664 Repolarization Abnormalities in Previous Electrocardiograms of Adult Victims of Non-Traumatic Sudden Cardiac Death Michael C. Plewa Mercy St. Vincent Medical Center, Toledo, OH Background: Repolarization abnormalities, such as long or short QTc interval (LQTI, SQTI), early repolarization (ER), Brugada syndrome (BrS), increased angle between R- and T-axes (QRS-T angle), and a prolonged interval between the peak and end of the T-wave (TPE) on the electrocardiogram (ECG) may increase the risk of ventricular arrhythmia leading to sudden cardiac death (SCD). Objectives: To describe the incidence of LQTI, SQTI, ER, BrS, wide QRS-T angle, and prolonged TPE interval on the most recent previous ECG of adult SCD cases. Methods: Retrospective, structured medical record review of all adult SCD cases for a 5-year period from 8/2006–7/2011 of a 63,000 visit emergency department (ED). Excluded were cases of age <18, trauma, overdose, hemorrhage, terminal illness, lack of prior ECG, paced rhythm, and bundle branch block, but not heart or renal disease or upper age limit. Records were S104 2012 SAEM ANNUAL MEETING ABSTRACTS reviewed for age, sex, race, and initial SCD rhythm. Twelve-lead ECG tracings obtained at 25 mm/sec with MAC 5500 ECG system (Marquette Medical Systems, Inc) were reviewed for computer-derived QTc, R-axis, and T-axis, and interpreted for LQTI, SQTI, ER, BrS, planar QRS-T angle, and TPE interval according to published criteria. Data are expressed as mean ± standard deviation, percentage, and 95% confidence interval. Results: A total of 164 cases were reviewed, average age 62 ± 16 years (range 25–99), with SCD initial rhythm of ventricular fibrillation in 49%. Previous ECG was an average of 36 ± 48 months prior to SCD, with 97% sinus rhythm and 3% atrial fibrillation. Average QTc was 446 ± 38 ms, with QTc ‡ 500 ms in 7% (3–12%), LQTI in 39% (31–47%), and SQTI in none (0– 2%). Previous ECG revealed ER in 12% (8–18%) and BrS type II morphology (without full BrS criteria) in 0.6% (0.3–4%). Average QRS-T angle was 56 ± 54o with 22% (16–29%) widened ‡ 90o. Average TPE was 114 ± 22 ms, with 36% (29–44%) prolonged ‡ 120 ms. Conclusion: Repolarization abnormalities, especially LQTI, wide QRS-T angle ‡ 90o, and prolonged TPE ‡ 120 ms, but not SQTI or Brugada syndrome morphology, are relatively common in the previous ECG of this population of older adults with SCD. Further research, including adults with similar medical illnesses and medications, is needed to clarify if these repolarization abnormalities are truly predictive of SCD. 192 The Association of Health Literacy, Self Care Behaviors, and Knowledge with Emergency Department Readmission Rates for Heart Failure Carolyn Overman, Daniel C. Hootman, Lydia Odenat, Douglas S. Ander Emory University School of Medicine, Atlanta, GA Background: Readmission rates at 30 days for heart failure (HF) are estimated at 13%. Changes to the Medicare payment policy for HF support study of readmission factors. Little is known regarding patient factors that affect HF readmission. Objectives: Objective: Determine the influence of health literacy and self-care knowledge on HF readmission rates. Methods: Prospective observational study of patients with the clinical diagnosis of HF in an inner-city ED. Forty-nine patients were enrolled in a 6-month period. Patient assessment included the Health Literacy Assessment (HLA), Heart Failure Knowledge Test (HFKT), and self-care behavior scores (SCBS). HLA, HFKT, and SCBS were compared to 30-day, 31–90 day, and 90-day readmission via independent binary logistic regressions. Chi-square tests were performed to identify associations between variables. The study was sufficiently powered to show a difference in readmission rates. Results: Of all participants, 69.4% were male, with a mean age of 56.1 years (sd = 11.9), 93.9% were African American, 38.8% were single, 26.5% were divorced or separated, and 36.7% had completed some high school education, with 30.6% having earned a high school degree or equivalency diploma. Additionally, 68.6% reported an income of less than $10,000. There were no significant associations between sex, race, marital status, education level, income, insurance status, and subsequent 30- and-90 day readmission rates. HLA score was not found to be significantly related to readmission rates. The mean HLA score was 18.9 (sd = 7.87), equivalent to less than 6th grade literacy, meaning these patients may not be able to read prescription labels. For each unit increase in HFKT score, the odds of being readmitted within 30 days decreased by 0.219 (p < 0.001) and for 31–90 days decreased by 0.440 (p < 0.001). For each unit increase in SCBS score, the odds of being readmitted within 90 days decreased by 0.949 (p = 0.038). Conclusion: Health care literacy in our patient population is not associated with readmission, likely related to the low literacy rate of our study population. Better HF knowledge and self-care behaviors are associated with lower readmission rates. Greater emphasis should be placed on patient education and self-care behaviors regarding HF as a mechanism to decrease readmission rates. 193 Comparison of Door to Balloon Times in Patients Presenting Directly or Transferred to a Regional Heart Center with STEMI Jennifer Ehlers, Adam V. Wurstle, Luis Gruberg, Adam J. Singer Stony Brook University, Stony Brook, NY Background: Based on the evidence, a door-to-balloon-TIME (DTBT) of less than 90 minutes is recommended by the AHA/ACC for patients with STEMI. In many regions, patients with STEMI are transferred to a regional heart center for percutaneous coronary intervention (PCI). Objectives: We compared DTBT for patients presenting directly to a regional heart center with those for patients transferred from other regional hospitals. We hypothesized that DTBT would be significantly longer for transferred patients. Methods: Study Design-Retrospective medical record review. Setting-Academic ED at a regional heart center with an annual census of 80,000 that includes a catchment area of 12 hospitals up to 50 miles away. Patients-Patients with acute STEMI identified on ED 12-lead ECG. Measures-Demographic and clinical data including time from triage to ECG, from ECG to activation of regional catheterization lab, and from initial triage to PCI (DTBT). Outcomes-Median DTBT and percentage of patients with a DTBT under 90 minutes. Data Analysis-Median DTBT compared with Mann Whitney U tests and proportions compared with chi-square tests. Results: In 2010 there were 379 catheterization lab activations for STEMI: 183 were in patients presenting directly, and 196 in transferred patients. Thrombolytics were administered in 19 (9.7%) transfers. Compared with patients presenting directly to the heart center, transferred patients had longer median [IQR] DTBT (127 [105–151] vs. 64 [49–80]; P < 0.001). Transferred patients also had longer door to ECG (9 [5–18] vs. 5 [2–8]; P < 0.001) and ECG to catheterization lab activation ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 times (18 [12–38] vs. 8 [4–17]; P < 0.001). The percentages of patients with a DTBT within 90 minutes in direct and transfer patients were 83% vs. 17%; P < 0.001. Conclusion: Most patients transferred to a regional heart center do not meet national DTBT guidelines. Consideration should be given to administering thrombolytics in transfer patients, especially if the transport time is prolonged. 194 A Comparison of the Management of ST-elevation Myocardial Infarction Between Patients Who Are English And Non-English Speaking Scott G. Weiner1, Kathryn A. Volz2, Matthew B. Mostofi1, Leon D. Sanchez2, John J. Collins3 1 Tufts Medical Center, Boston, MA; 2Beth Israel Deaconess Medical Center, Boston, MA; 3 Robert Wood Johnson University Hospital, New Brunswick, NJ Background: Prompt treatment with revascularization in the catheterization lab (cath) is essential to preserve cardiac function for patients who present with ST-elevation myocardial infarction (STEMI). Objectives: To determine if there are disparities between patients who speak English or do not speak English who present with STEMI in times from door to first EKG (D2E), door to cath (D2C), and door to intravascular balloon deployment (D2B). Methods: The study was performed in an inner-city academic ED between 1/1/07 and 12/31/10. Every patient for whom ED activation of our STEMI system occurred was included. All times data from a pre-existing quality assurance database were collected prospectively. Patient language was determined retrospectively by chart review. Results: There were 132 patients between 1/1/07 and 12/31/10. 21 patients (16%) were deemed too sick or unable to provide history and were excluded, leaving 111 patients for analysis. 85 (77%) spoke English and 26 (23%) did not. In the non-English group, Chinese was the most common language, in 22 (20%) patients. There was no difference in mode of arrival (EMS 44% English vs. 46% non-English, p = 0.85) or arrival between 5:00 pm–9:00 am (47% English vs. 58% non-English, p = 0.34). English patients were more likely to have a documented chief complaint of chest pain or pressure (81% vs. 42%, p < 0.001). D2E times were not different (English median 8 min [2.5–13.5] vs. non-English median 11 min [4.5–17], p = 0.10). English speakers were more likely to go to cath lab (79% vs. 54%, p = 0.01). Of those who had went to cath, there was no significant D2C time difference (English median 53 min [33–71] vs. non-English median 45.5 min [36–87], p = 0.39). For those who had revascularization, D2B times were also similar (English median 78.5 min [57.8–94.5] vs. nonEnglish median 65 min [55–126], p = 0.44.) The percentage of patients with D2B time <90 min was not different (72% English vs. 64% non-English, p = 0.60). Conclusion: We found that non-English speakers are much less likely to complain of chest pain as an initial chief complaint and are less likely to go to cath. Other parameters such as time and mode of arrival are not • www.aemj.org S105 different. We discovered that our D2E, D2C, and D2B times are not significantly different between these two patient populations. Further research is needed to determine why non-English speaking patients were less likely to be taken to cath. 195 A Four-year Population-Based Analysis Of Emergency Department Syncope: Predictors Of Admission/Readmission, And Regional Variations In Practice Patterns Xin Feng1, Zhe Tian2, Brian Rowe3, Andrew McRae1, Venkatesh Thiruganasambandamoorthy4, Rhonda Rosychuk3, Robert Sheldon1, Eddy Lang1 1 University of Calgary, Calgary, AB, Canada; 2 McGill University, Montreal, QC, Canada; 3 University of Alberta, Edmonton, AB, Canada; 4 University of Ottawa, Ottawa, ON, Canada Background: Syncope is a common, potentially highrisk ED presentation. Hospitalization for syncope, although common, is rarely of benefit. No populationbased study has examined disparities in regional admission practices for syncope care in the ED. Moreover, there are no population-based studies reporting prognostic factors for 7- and 30-day readmission of syncope. Objectives: 1) To identify factors associated with admission as well as prognostic factors for 7- and 30-day readmission to these hospitals; 2) To evaluate variability in syncope admission practices across different sizes and types of hospitals. Methods: DESIGN - Multi-center retrospective cohort study using ED administrative data from 101 Albertan EDs. PARTICIPANTS/SUBJECTS - patients >17 years of age with syncope (ICD10: R55) as a primary or secondary diagnosis from 2007 to June 2011. Readmission was defined as return visits to the ED or admission <7 days or 7–30 days after the index visit (including against medical advice and left without being seen during the index visit). OUTCOMES - factors associated with hospital admission at index presentation, and readmission following ED discharge, adjusted using multivariable logistic regression. Results: Overall, 44521 syncope visits occurred over 4 years. Increased age, increased length of stay (LoS), performance of CXR, transport by ground ambulance, and treatment at a low-volume hospital (non-teaching or non-large urban) were independently associated with index hospitalization. These same factors, as well as hospital admission itself, were associated with 7-day readmission. Additionally, increased age, increased LoS, performance of a head CT, treatment at a low-volume hospital, hospital admission, and female sex were independently associated with 7–30 day readmission. Arrival by ground ambulance was associated with a decreased likelihood of both 7- and 7–30 day readmission. Conclusion: Our data identify variations in practice as well as factors associated with hospitalization and readmission for syncope. The disparity in admission and readmission rates between centers may highlight a gap in quality of care or reflect inappropriate use of resources. Further research to compare patient out- S106 2012 SAEM ANNUAL MEETING ABSTRACTS Table 1 - Abstract 195: Multivariate analysis of prognostic factors for admission and readmission Characteristic Increased age Increased length of stay Treatment at a low-volume hospital (non-teaching or not large urban) compared to high-volume hospital (teaching or large urban) CXR performed Head CT performed Ground ambulance transportation Hospital admission Sex (M:F) Admission (odds ratio, 95% CI) 7-day re-admission (odds ratio, 95% CI) 7–30 day re-admission (odds ratio, 05% CI) 1.03 (1.02–1.05) per year 1.04 (1.03–1.05) per hour 1.32 (1.09–1.60) 1.006 (1.001–1.01) per year 1.025 (1.005–1.05) per hour 1.93 (1.78–2.09) 1.009 (1.001–1.03) per year 1.026 (1.004–1.06) per hour 1.97 (1.85–2.10) 1.54 (1.27–1.86) Not significant 1.31 (1.07–1.60) Not applicable Not significant 1.07 (1.009–1.127) Not significant 0.90 (0.83–0.97) 1.79 (1.38–2.33) Not significant Not significant 1.08 (1.008–1.165) 0.87 (0.82–0.93) 2.15 (1.73–2.67) 0.93 (0.88–0.99) comes and quality of patient care among urban and non-urban centers is needed. Table 2 - Abstract 195: Descriptive statistics on syncope in Alberta from 2007 to 2011 (statistical outliers removed) Sex Age Admission rate from ED 7-day readmission rate 7–30 day readmission rate Male 20559 (46.2%) Mean 54.2 years Female 23962 (53.8%) Median 55 years High-volume: teaching or large urban Mean (range) 19% (9%–27%) High-volume: teaching or large urban Mean (range) 10% (3%–17%) High-volume: teaching or large urban Mean (range) 12% (3%–30%) Low-volume: non-teaching, non-large urban Mean (range) 15% (11%–43%) Low-volume: non-teaching, non-large urban Mean (range) 16% (2%–29%) Low-volume: non-teaching, non-large urban Mean (range) 16% (8%–34%) Range 18–104 years was the 4-hour DS change. Two clinical outcome measures were obtained: 1) the number of days hospitalized or dead within 30 days of the index visit (30-day outcome), and 2) the number of days hospitalized or dead within 90 days of the index visit (90-day outcome). Results: Data on 86 patients were analyzed. The median 30-day outcome variable was 6 days with an interquartile range (IQR) of 3 to 16. The median 90-day outcome variable was 10 days (IQR 4 to 27.5). The median 1-hour DS change was 2.6 cm (IQR 0.3 to 6.7). The median 4-hour DS change was 4.9 cm (IQR 2.2 to 8.2). The 30-day and 90-day mortality rates were 9% and 13% respectively. The spearman rank correlations and 95% confidence intervals are presented in the table below. Conclusion: While the point estimates for the correlations were below 0.5, the 95% CI for two of the correlations extended above 0.5. These pilot data support change in DS as a valid outcome measure for AHF when measured over 4 hours. A larger prospective study is needed to obtain a more accurate point estimate of the correlations. Table - Abstract 196: Spearman Rank Correlations 196 Correlation Between Change in Dyspnea Severity and Clinical Outcome in Patients with Acute Heart Failure Howard Smithline Baystate Medical Center, Springfield, MA Background: Change in dyspnea severity (DS) is a frequently used outcome measure in trials of acute heart failure (AHF). However, there is limited information concerning its validity. Objectives: To assess the predictive validity of change in dyspnea severity. Methods: This was a secondary analysis of a prospective observational study of a convenience sample of AHF patients presenting with dyspnea to the ED of an academic tertiary referral center with a mixed urban/ suburban catchment area. Patients were enrolled weekdays, June through December 2006. Patients assessed their DS using a 10-cm visual analog scale at three times: the start of ED treatment (baseline) as well as at 1 and 4 hours after starting ED treatment. The difference between baseline and 1 hour was the 1-hour DS change. The difference between baseline and 4 hours 1-hour DS change vs 30-day outcome: 4-hour DS change vs 30-day outcome: 1-hour DS change vs 90-day outcome: 4-hour DS change vs 90-day outcome: 197 0.027 (95% CI -0.186 to 0.238) )0.314 (95% CI -0.514 to -0.082) 0.016 (95% CI -0.196 to 0.227) )0.307 (95% CI -0.508 to -0.073) Ability of a Triage Decision Rule for Rapid Electrocardiogram (ECG) to Identify Patients with Suspected ST-elevation Myocardial Infarction (STEMI) Karim Ali, Anwar D. Osborne, James P. Capes, Douglas Lowery-North, Matthew Wheatley, Rachel E. O’Malley, George Leach, Vicki Hertzberg, Franks M. Nicole, Ryan Stroder, Stephen R. Pitts, Michael A. Ross Emory University, Atlanta, GA Background: ACC/AHA guidelines for STEMI state that an ECG should be performed upon presentation to the ED within 10 minutes. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Objectives: To determine the performance of a previously published rapid ECG screening criteria in a population of suspected STEMI patients. This rule was originally designed to identify patients with acute myocardial infarction needing a rapid ECG at ED triage based on presenting complaints in the lytic therapy era. We hypothesize that it would not have identified all patients in which STEMI was suspected. Methods: Three trained physician reviewers retrospectively applied the decision rule to 430 consecutive patients from a database of emergent cardiac catheterization lab (CCL) activations by ED physicians. The decision rule recommends that patients between the ages of 30 and 49 received a rapid ECG if they complained of chest pain and those aged 50 years or older when they complained of chest pain, shortness of breath, palpitations, weakness, or syncope. The triage note or earliest medical contact documentation was used to determine if the patient’s complaints would have resulted in a rapid ECG by the decision rule. Acute myocardial infarction (AMI) was defined as high-grade • www.aemj.org S107 stenosis on the subsequent emergent cardiac catheterization. A single data collection Microsoft Excel spreadsheet was used and descriptive statistics were performed in Excel. Results: The triage ECG rule would have identified 97% patients causing activation of the CCL (see figure). Among these patients, the rule was 98% sensitive (95%CI 95%–98%) for identifying patients who had a high grade stenois on catheterization (see table). Of the 430 STEMI activation patients, 412 patients would have been identified by this rule. Of the 18 patients who would not have been indentified by the rule, four cases were unwitnessed cardiac arrests that could not provide a history. The remaining 14 patients largely presented with nausea/vomiting or were under 30 years old. Conclusion: This triage rule would have identified almost all of our patients where STEMI was suspected. It would have failed to identify patients needing a rapid EKG who presented with nausea and vomiting as a sole complaint as well as patients who were under 30 years old. S108 198 2012 SAEM ANNUAL MEETING ABSTRACTS Health Care Resource Utilization Among Patients with Acute Decompensated Heart Failure Managed by Two University Affiliated Emergency Department Observation Units, 2007–2011 Justin Schrager, Matthew Wheatley, Stephen Pitts, Daniel Axelson, Anwar Osborne, Andreas Kalogeropoulos, Vasiliki Georgiopoulou, Javed Butler, Michael Ross Emory University School of Medicine, Atlanta, GA Background: Emergency department observation units (EDOU) have the potential to reduce costs and unnecessary admissions in the adult heart failure population. It is not known whether EDOU treatment yields similar readmission rates, which could limit this benefit. Objectives: To compare readmission rates and resource utilization between patients admitted following EDOU treatment of acute decompensated heart failure (ADHF) and those successfully discharged from OU. Methods: DESIGN - Retrospective observational cohort study. SETTING - Two university-affiliated EDOUs. SUBJECTS - 358 patients treated for ADHF in two protocol-driven OUs from 10/01/07–6/30/11. Thirtyone patients were excluded for a final diagnosis other than ADHF. OBSERVATIONS - The exposure was admission or discharge following OU treatment. The outcome was readmission within 30 days of either OU discharge or hospital discharge if initially admitted from OU. Descriptive statistical analyses of covariates included age, race, sex, clinical site, ED length of stay (LOS), and OU LOS, as well as B-type natriuretic peptide (BNP), blood urea nitrogen (BUN), serum creatinine (Cr), and ejection fraction (EF). Time to readmission analysis was performed with Cox proportional hazards regression. We also examined resource utilization. The study was powered to show a difference in rate of readmission of 40%. Results: Patients did not differ significantly by exposure based on age, race, sex, ED LOS, or OU LOS. Admitted patients had a higher median BNP (1063 pg/ ml vs. 708 pg/ml, p = 0.0019), and higher BUN (19 mg/ dL vs. 17 mg/dL, p = 0.0445). Admitted patients had a lower median EF (22.5% vs. 35%, p = 0.0020). In adjusted Cox proportional hazards models, the 30-day readmission rate was not significantly different between those admitted and those discharged from OU (HR = 0.95; 95% CI 0.46–1.99). Patients discharged from OU spent a median of 1.7 days as inpatients compared to 3.5 days among those admitted from OU, within 30 days (p < 0.0001). Among all readmitted patients the total median inpatient time was not significantly different. Conclusion: ADHF patients treated in the EDOU and discharged were not more likely to be readmitted within 30 days than those admitted to the hospital. Patients successfully treated in the EDOU for ADHF used fewer hospital bed-days. 199 Emergency Department Case Volume and Short-term Outcomes in Patients with Acute Heart Failure Chu-Lin Tsai1, Wen-Ya Lee1, George L. Delclos1, Carlos A. Camargo2 1 Division of Epidemiology, Human Genetics and Environmental Sciences, The University of Texas School of Public Health, Houston, TX; 2 Department of Emergency Medicine, Massachusetts General Hospital, Harvard Medical School, Boston, MA Background: The majority of volume-quality research has focused on surgical outcomes in the inpatient setting; very few studies have examined the effect of emergency department (ED) case volume on patient outcomes. Objectives: To determine whether ED case volume of acute heart failure (AHF) is associated with short-term patient outcomes. Methods: We analyzed the 2008 Nationwide Emergency Department Sample (NEDS) and Nationwide Inpatient Sample (NIS), the largest, all-payer, ED and inpatient databases in the US. ED visits for AHF were identified with a principal diagnosis of ICD-9-CM code 428.xx. EDs were categorized into quartiles by ED case volume of AHF. The outcome measures were early inpatient mortality (within the first 2 days of admission), overall inpatient mortality, and hospital length of stay (LOS). Results: There were an estimated 946,000 visits for AHF from approximately 4,700 EDs in 2008; 80% were hospitalized. Of these, the overall inpatient mortality rate was 3.2%, and the median hospital LOS was 4 days. Early inpatient mortality was lower in the highest-volume EDs, compared with the lowest-volume EDs (0.8% vs. 2.1%; P < 0.001). Similar patterns were observed for overall inpatient mortality (3.0% vs. 4.1%; P < 0.001). In a multivariable analysis adjusting for 37 patient and hospital characteristics, early inpatient mortality remained lower in patients admitted through the highest-volume EDs (adjusted odds ratios [OR], 0.70; 95% confidence interval [CI], 0.52–0.96), as compared with the lowest-volume EDs. There was a trend towards lower overall inpatient mortality in the highest-volume EDs; however, this was not statistically significant (adjusted OR, 0.92; 95%CI, 0.75–1.14). By contrast, using the NIS data including various sources ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S109 in the Midwest (76.0%) and West (74.8%). Total monies spent on ED services were highest in the South ($69,078,042) followed by the Northeast ($18,233,807), West ($6,360,315) and Midwest ($5,899,481). Conclusion: This large retrospective ED cohort suggests a very high national admission rate with significant regional variation in both disposition decisions as well as total monies spent on ED services for patients with a primary diagnosis of heart failure. Examining these estimates and variations further may provide strategies to reduce the economic burden of heart failure. Table - Abstract 200: Disposition of admissions, a higher case volume of inpatient AHF patients predicted lower overall inpatient mortality (adjusted OR, 0.51; 95%CI, 0.40–0.65). The hospital LOS in patients admitted through the highest-volume EDs was slightly longer (adjusted difference, 0.7 day; 95%CI, 0.2–1.2), compared with the lowest-volume EDs. Conclusion: ED patients who are hospitalized for AHF have an approximately 30% reduced early inpatient mortality if they were admitted from an ED that handles a large volume of AHF cases. The ‘‘practice-makesperfect’’ concept may hold in emergency management of AHF. 200 Emergency Department Disposition and Charges for Heart Failure: Regional Variability Alan B. Storrow, Cathy A. Jenkins, Sean P. Collins, Karen P. Miller, Candace McNaughton, Naftilan Allen, Benjamin S. Heavrin Vanderbilt University, Nashville, TN Background: High inpatient admission rates for ED patients with acute heart failure are felt partially responsible for the large economic burden of this most costly cardiovascular problem. Objectives: We examined regional variability in ED disposition decisions and regional variability in total dollars spent on ED services for admitted patients with primary heart failure. Methods: The 2007 Nationwide Emergency Department Sample (NEDS) was used to perform a retrospective, cohort analysis of patients with heart failure (ICD9 code of 428.x) listed as the primary ED diagnosis. Demographics and disposition percentages (with SE) were calculated for the overall sample and by region: Northeast, South, Midwest, and West. To account for the sample design and to obtain national and regional estimates, a weighted analysis was conducted. Results: There were 941,754 weighted ED visits with heart failure listed as the primary diagnosis. Overall, over eighty percent were admitted (see table). Fifty-two percent of these patients were female; mean age was 72.7 years (SE 0.20). Hospitalization rates were higher in the Northeast (89.1%) and South (81.2%) than Treated and released Admitted to same hospital Transferred ED death Discharged alive, destination unknown Unknown 201 Weighted Frequency (SD) Percent (SE) 150,667 (4,449) 759,005 (19,599) 27,392 (2,023) 1,180 (94) 44 (24) 16.0 (0.41) 80.6 (0.482) 2.9 (0.22) 0.13 (0.01) 0.005 (0.003) 3,466 (859) 0.37 (0.09) Hospital-Based Shootings in the United States: 2000–2010 Gabor D. Kelen, Christina L. Catlett, Joshua G. Cubit, Yu-Hsiang Hsieh Johns Hopkins University, Baltimore, MD Background: Workplace violence in health care settings is a frequent occurrence. Gunfire in hospitals is of particular concern. However, information regarding such workplace violence is limited. Accordingly, we characterized U.S. hospital-based shootings from 2000–2010. Objectives: To determine extent of hospital-based shootings in the U.S. and involvement of emergency departments. Methods: Using LexisNexis, Google, Netscape, PubMed, and ScienceDirect, we searched reports for acute care hospital shooting events from January 2000 through December 2010, and those with at least one injured victim were analyzed. Results: We identified 140 hospital-related shootings (86 inside the hospital, 54 on hospital grounds), in 39 states, with 216 victims, of whom 98 were perpetrators. In comparison to external shootings, shootings within the hospital have not increased over time (see figure). Perpetrators were from all age groups, including the elderly. Most of the events involved a determined shooter: grudge (26%), suicide (19%), ‘‘euthanizing’’ an ill relative (15%), and prisoner escape (12%). Ambient societal violence (8%) and mentally unstable patients (4%) were comparatively infrequent. The most common injured was the perpetrator (45%). Hospital employees comprised only 21% of victims; physician (3%) and nurse (5%) victims were relatively infrequent. The emergency department was the most common site (29%), followed by patient rooms (20%) and the parking lot (20%). In 13% of shootings within hospitals, the weapon was a security officer’s gun grabbed by the perpetrator. ‘‘Grudge’’ motive was the only factor determinative of hospital staff victims (OR = 4.34, 95% CI 1.85–10.17). S110 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: Although hospital-based shootings are relatively rare, emergency departments are the most likely site. The unpredictable nature of this type of event represents a significant challenge to hospital security and deterrence practices, as most perpetrators proved determined, and many hospital shootings occur outside the building. 202 Impact of Emergency Physician Board Certification on Patient Perceptions of ED Care Quality Albert G. Sledge IV1, Carl A. Germann1, Tania D. Strout1, John Southall2 1 Maine Medical Center, Portland, ME; 2Mercy Hospital, Portland, ME Background: The Hospital Value-Based Purchasing Program mandated by the Affordable Care Act is the latest example of how patients’ perceptions of care will affect the future practice environment of all physicians. The type of training of medical providers in the emergency department (ED) is one possible factor affecting patient perceptions of care. A unique situation in a Maine community ED led to the rapid transition from non-emergency medicine (EM) residency trained physicians to all EM residency trained and American Board of Emergency Medicine (ABEM) certified providers. Objectives: The purpose of this study was to evaluate the effect of the implementation of an all EM-trained, ABEM-certified physician staff on patient perceptions of the quality of care they received in the ED. Methods: We retrospectively evaluated Press Ganey data from surveys returned by patients receiving treatment in a single, rural ED. Survey items addressed patient’s perceptions of physician courtesy, time spent listening, concern for patient comfort, and informativeness. Additional items evaluated overall perceptions of care and the likelihood that the respondent would recommend the ED to another. Data were compared for the three years prior to and following implementation of the all trained, certified staff. We used the independent samples t-test to compare mean responses during the two time periods. Bonferroni’s correction was applied to adjust for multiple comparisons. Results: During the study period, 3,039 patients provided surveys for analysis: 1,666 during the pre-certification phase and 1,373 during the post-certification phase. Across all six survey items, mean responses increased following transition to the board-certified staff. These improvements were noted to be statistically significant in each case: courtesy p < 0.001, time listening p < 0.001, concern for comfort p < 0.001, informativeness p < 0.001, overall perception of care p < 0.001, and likelihood to recommend p < 0.001. Conclusion: Data from this community ED suggest that transition from a non-residency trained, ABEM certified staff to a fully trained and certified model has important implications for patient’s perceptions of the care they receive. We observed significant improvement in rating scores provided by patients across all physicianoriented and general ED measures. 203 Electronic, Verbal Discussion-Optional Signout for Admitted Patients: Effects on Patient Safety and ED Throughput Christopher M. Fischer, Julius Yang, Carrie Tibbles, Elizabeth O’Donnell, Ethan Ellis, Larry Nathanson, Leon D. Sanchez Beth Israel Deaconess Medical Center/Harvard Medical School, Boston, MA Background: Transfer of care from the ED to the inpatient floor is a critical transition when miscommunication places patients at risk. The optimal form and content of handoff between providers has not been defined. In July 2011, ED-to-floor signout for all admissions to the medicine and cardiology floors was changed at our urban, academic, tertiary care hospital. Previously, signout was via an unstructured telephone conversation between ED resident and admitting housestaff. The new signout utilizes a web-based ED patient tracking system and includes: 1) a templated description of ED course is completed by the ED resident; 2) when a bed is assigned, an automated page is sent to the admitting housestaff; 3) ED clinical information, including imaging, labs, medications, and nursing interventions (figure) is reviewed by admitting housestaff; 4) if housestaff has specific questions about ED care, a telephone conversation between ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S111 S112 2012 SAEM ANNUAL MEETING ABSTRACTS the ED resident and housestaff occurs; 5) if there are no specific questions, it is indicated electronically and the patient is transferred to the floor. Objectives: To describe the effects on patient safety (floor-to-ICU transfer in 24 hours) and ED throughput (ED length of stay (LOS) and time from bed assignment to ED departure) resulting from a change to an electronic, discussion-optional handoff system. Methods: Review of all patients admitted from the ED to the medicine and cardiology floor from July-October 2010 and 2011. Rate of floor to ICU transfer in 24 hours, ED LOS, and time from bed assignment to ED departure were calculated by review of medical records. Results: There were 3334 admissions in 2010 and 3347 in 2011. Patient characteristics are in the table. After July 2011, 28.2% of patients had a verbal signout between the ED and inpatient teams. For the remaining 71.8%, there was review of clinical information and no verbal discussion. The rate of floor-to-ICU transfer was 2.0% in 2010, and 2.2% in 2011 (p = 0.46). ED LOS was similar (6:13 vs 6:10, p = 0.18). Median time from bed assignment to ED departure decreased 5 minutes (1:36 in 2010 vs 1:31 in 2011, p < 0.01). Conclusion: Transition to a system in which signout of admitted patients is accomplished by accepting housestaff review of ED clinical information supplemented by verbal discussion when needed resulted in no significant change in rate of floor-to-ICU transfer or ED LOS and reduced time from bed assignment to ED departure. Table - Abstract 203: Patient Characteristics And Outcomes 2010 Total admissions Percent male Age (median, IQR) Percent white Hospital LOS (median, IQR) ED LOS (median, IQR) Percent verbal discussion Time from bed assign to ED departure (median, IQR) % transfer from floor to ICU within 24 hours of admission 2011 3334 48.5% 64 (50–79) 66.8% 3 (2–5) 3347 44.8% 64 (50–79) 68.6% 2 (1–4) 6:13 (4:52–8:04) 100% 6:10 (4:45–8:02) 28.2% 1:36 (1:15–2:02) 1:31 (1:14–1:56) 2.0% 2.2% 204 Does the Nature of Chief Complaint, Gender, or Age Affect Time to be Seen in the Emergency Department? Ayesha Sattar1, John Marshall2, Kenneth Sable2, Antonios Likourezos2, Christian Fromm2 1 Stanford University School of Medicine, Stanford, CA; 2Maimonides Medical Center, Brooklyn, NY Background: Emergency physicians may be biased against patients presenting with nonspecific complaints or those requiring more extensive work-ups. This may result in patients being seen less quickly than those with more straightforward presentations, despite equal triage scores or potential for more dangerous conditions. Objectives: The goal of our study was to ascertain which patients, if any, were seen more quickly in the ED based on chief complaint. Methods: A retrospective report was generated from the EMR for all moderate acuity (ESI 3) adult patients who visited the ED from January 2005 through December 2010 at a large urban teaching hospital. The most common complaints were: abdominal pain, alcohol intoxication, back pain, chest pain, cough, dyspnea, dizziness, fall, fever, flank pain, headache, infection, pain (nonspecific), psychiatric evaluation, ‘‘sent by MD,’’ vaginal bleeding, vomiting, and weakness. Non-parametric independent sample tests assessed median time to be seen (TTBS) by a physician for each complaint. Differences in the TTBS between genders and based on age were also calculated. Chi-square testing compared percentages of patients in the ED per hour to assess for differences in the distribution of arrival times. Results: We obtained data from 116,194 patients. Patients with a chief complaint of weakness and dizziness waited the longest with a median time of 35 minutes and patients with flank pain waited the shortest with 24 minutes (p < 0.0001) (Figure 1). Overall, males waited 30 minutes and females waited 32 minutes (p < 0.0001). Stratifying by gender and age, younger females between the ages of 18–50 waited significantly longer times when presenting with a chief complaint of abdominal pain (p < 0.0001), chest pain (p < 0.05), or flank pain (p < 0.0001) as compared to males in the same age group (Figure 2). There was no difference in the distribution of arrival times for these complaints. Conclusion: While the absolute time differences are not large, there is a significant bias toward seeing young ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 male patients more quickly than women or older males despite the lower likelihood of dangerous conditions. Triage systems should perhaps take age and gender better into account. Patients might benefit from efforts to educate EM physicians on the delays and potential quality issues associated with this bias in an attempt to move toward more egalitarian patient selection. 205 The Impact of a Neurology Consult in Patients Placed in Observation for Syncope or Near Syncope Simon Katrib, Margarita E. Pena, Robert B. Takla, Patrick Frank, Susan Szpunar St. John Hospital and Medical Center, Detroit, MI Background: In patients placed in an ED-staffed observation unit (OU) for syncope or near syncope (S/NS), some physicians’ practice pattern includes a routine neurology consult (NC) for all patients, including those without a previous seizure history. Objectives: We tested the hypothesis that very few S/NS patients without a previous seizure history placed in the OU would have a neurologic etiology, and a NC would not contribute significantly to diagnosis or recidivism, but may add to length of stay (LOS) and cost. Methods: This is a retrospective chart review study of all adult patients placed in an OU staffed by ED physicians with a primary diagnosis of S/NS from October 2009 to October 2011. Patients with and without a NC were compared. Patient data collected included demographics, past seizure history, NC, discharge summary reports, tests ordered by neurology, hospital LOS, 48-hour return to the ED after discharge, 30-day readmission, and death. Financial data collected included direct, indirect, and total cost. Chi-square analysis was used to examine associations between the two groups and Student’s t-test to examine differences between mean values. All data were analyzed with SPSS v. 19.0. Results: A NC was obtained in 38.1% of 247 study patients. Of these, 12.1% (11/94) were diagnosed with a neurologic etiology for their S/NS; 10/11 had a final diagnosis of recurrent seizures and 1/11 had a new neurologic diagnosis (subacute stroke seen on CT after patient with new unilateral weakness arrived in OU). Mean age was similar in both groups (p = 0.327). Length of stay for discharged patients with a NC was • www.aemj.org S113 not affected (1.89 days vs. 1.83 days, p = 0.448), but did increase for admitted patients with a NC (5.24 days vs. 2.89 days, p = 0.01). Patients with a NC were less likely to be discharged (77.7% vs. 93.2%, p = 0.001) but tended to be more likely to return to the ED within 48 hours (p = 0.06). There was no difference in 30-day readmission between the two groups (p = 0.293). There were no deaths. A NC adds to direct, indirect, and total costs (p = 0.001, 0.004, and 0.001 respectively). Conclusion: In patients presenting to an OU for evaluation of S/NS, only a small percentage have a neurologic etiology, most of these due to recurrent seizures. A NC does add to cost, increases the likelihood of admission, and prolongs LOS, but does not contribute significantly to a neurologic diagnosis or decrease recidivism. 206 Time Burden of Emergency Department Hand Hygiene with Glove Use Joseph M. Reardon1, Josephine E. Valenzuela1, Siddharth Parmar2, Arjun Venkatesh3, Jeremiah D. Schuur2, Daniel J. Pallin2 1 Harvard Medical School, Boston, MA; 2Brigham and Women’s Hospital, Boston, MA; 3Brigham and Women’s Hospital-Massachusetts General Hospital-Harvard Affiliated Emergency Medicine Residency, Boston, MA Background: Guidelines require ED personnel to perform hand hygiene before and after patient contact, whether nonsterile gloves are used or not. This requirement is based on old studies in the operating room setting. ED staff are known to be less likely to comply when using gloves vs. when not using gloves, perhaps because this seems arbitrary and burdensome. Knowledge of time and materials costs, and benefits measured in decreased disease transmission, are key to understanding and encouraging proper hand hygiene. Objectives: To measure the time burden of alcoholbased handrub use before and after non-sterile glove use among ED staff. Methods: Research assistants counted PGY-2 and -3 EM residents’ glove donning events per hour for 42 hours of observation during clinical shifts. S114 2012 SAEM ANNUAL MEETING ABSTRACTS Table - Abstract 206: Average time (seconds) for alcohol-based handrub use, glove donning, and glove removal Times for Provider Type: Attending Resident Physician Nurse Student Mean (95%CI) Hand rub before gloves Glove donning 9 8 8.8 13.4 7 10.8 (9.0 – 12.6) 16.5 23 17 22.2 23 21.7 (17.7 – 25.7) Glove removal Hand rub after gloves Glove donning alone Glove removal alone Added time for handrub before gloves Added time for handrub after gloves 2.5 3.9 3.2 3.7 2.3 3.6 (3.0 – 4.1) 12.5 6.7 9 7.8 7.7 7.8 (6.9 – 8.7) 17 14.6 11.2 15.4 20 15.2 (12.8 – 17.6) 4 3.1 2.8 6.6 3.3 4.9 (2.8 – 7.0) 8.5 16.4 14.5 13.9 17.3 17.3 (12.9 – 21.7) 11 7.5 9.5 4.8 6.7 6.5 (4.5 – 8.5) Separately, in a controlled setting, the observers timed 40 ED physicians, physician assistants, residents, and nurses donning and removing gloves with and without handrub. We report glove donning events per hour, and donning times and removal times with and without handrub, with 95%CI. Setting: Urban, academic ED, census 60,000. Results: Residents used gloves 0.83 times per hour (95%CI 0.6–1.1) at an ED occupancy rate of 71%. Simultaneously, the average number of ideal gloving events based on a checklist of clinical necessity was 0.69 per hour (95%CI 0.51–1.06). Handrub use added a mean of 17 seconds (95%CI 13–22) before gloving and 8 seconds (95%CI 2–14) after gloving. Thus, handrub use added 14 seconds (95%CI 8–20) per physician per hour. Among the 44% of residents who used sanitizer before putting on gloves (95%CI 26–62), compliance with WHO guidelines for minimum 30 seconds of rub was 3% (95%CI 0–8). Conclusion: Alcohol-based handrub use represents a small time burden for providers when combined with nonsterile glove use. These data may be helpful in efforts to motivate increased hand hygiene compliance. 207 Door-to-Balloon Times for Primary Percutaneous Coronary Intervention: How Do Freestanding Emergency Departments Perform? Erin L. Simon, Peter Griffin, Thomas Lloyd Akron General Medical Center, Akron, OH Background: Freestanding emergency departments (FEDs) have become increasingly popular as the need for emergency care continues to grow. It is unclear how door-to-balloon (D2B) times are affected or if the D2B goal of <90 minutes is met when ST-elevation myocardial infarction patients present to a FED. Objectives: The objective of this study was to determine the proportion of STEMI patients transported by ground ambulance from two different FEDs to a single percutaneous coronary intervention (PCI) center who meet the 90-minute goal for D2B for PCI (utilizing a rapid STEMI protocol). Secondary aims included analyzing individual time components from arrival to the FED until the completion of the PCI and their effect on D2B times. Methods: We conducted a retrospective cohort review and included all patients who presented to two FEDs through April 2011 with a STEMI since the opening of these facilities in July 2007 and August 2009, respectively. Demographic information and key time points were abstracted and statistical evaluation was performed using chi-square analysis. Results: Thirty-five patients met inclusion criteria. The mean arrival time to initial ECG time was 2.77 minutes. The average door-to-transfer time was 32.5 minutes. The average D2B time was 84.97 minutes (SDEV ± 13.04 minutes), with 74.3% of patients having a D2B time less than 90 minutes. The time between the ECG recording and the STEMI alert call was found to be highly significant (p < 0.005), as was the transport time from the ambulance arrival to the PCI center to the catheterization lab itself (p < 0.005). The door-to-transfer time, door-to-doortime, and length of transport time were also significant (p < 0.005; p < 0.001; p < 0.001 respectively). The average door-to-transfer time from the FED was 29.5 minutes in the group achieving D2B times under 90 minutes, and was 39.3 minutes in the group who did not meet the 90 minute goal. Conclusion: A total of 74.3% of STEMI patients seen in two FEDs achieved D2B times under 90 minutes. Factors associated with D2B times under 90 minutes were time from initial ECG to calling a STEMI alert, door-totransfer time, ambulance arrival at the hospital to arrival in the catheterization lab, and overall transport time. Minimizing delays in these areas may decrease D2B times from FEDs. 208 Emergency Department Holding Orders Reduce ED Length Of Stay By Decreasing Time To Bed Order Samir A. Haydar, Joel Botler, Tania D. Strout, Karen D. Taylor Maine Medical Center, Portland, ME Background: Detailed analysis of emergency department (ED) event data identified the time from completion of emergency physician evaluation (Doc Done) to the time patients leave the ED as a significant contributor to ED length of stay (LOS) and boarding at our institution. Process flow mapping identified the time from Doc Done to the time inpatient beds were ordered (BO) as an interval amendable to specific process improvements. Objectives: The purpose of this study was to evaluate the effect of ED holding orders for stable adult ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 inpatient medicine (AIM) patients on: a) the time to BO and b) ED LOS. Methods: A prospective, observational design was used to evaluate the study questions. Data regarding the time to BO and LOS outcomes were collected before and after implementation of the ED holding orders program. The intervention targeted stable AIM patients being admitted to hospitalist, internal medicine, and family medicine services. ED holding orders were placed following the admission discussion with the accepting service and special attention was paid to proper bed type, completion of the emergent work-up and the expected immediate course of the patient’s hospital stay. Holding orders were of limited duration and expired 4 hours after arrival to the inpatient unit. Results: During the 6-month study period, 7321 patients were eligible for the ED holding orders intervention; 6664 (91.0%) were cared for using the standard adult medicine order set and 657 (9.0%) received the intervention. The median time from Doc Done to BO was significantly shorter for patients in the ED holding orders group, 41 min (IQR 19, 88) vs. 95 min (IQR 53, 154) for the standard adult medicine group, p < 0.001. Similarly, the median ED LOS was significantly shorter for those in the ED holding orders group, 413 min (IQR 331, 540) vs. 456 min (IQR 346, 581) for the standard adult medicine group, p < 0.001. No lapses in patient care were reported in the intervention group. Conclusion: In this cohort of ED patients being admitted to an AIM service, placing ED holding orders rather than waiting for a traditional inpatient team evaluation and set of admission orders significantly reduced the time from the completion of the ED workup to placement of a BO. As a result, ED LOS was also significantly shortened. While overall utilization of the intervention was low, it improved with each month. 209 Emergency Department Interruptions in the Age of Electronic Health Records Matthew Albrecht, John Shabosky, Jonathan de la Cruz Southern Illinois University School of Medicine, Springfield, IL Background: Interruptions of clinical care in the emergency department (ED) have been correlated with increased medical errors and decreased patient satisfaction. Studies have also shown that most interruptions happen during physician documentation. With the advent of the electronic health record and computerized documentation, ED physicians now spend much of their clinical time in front of computers and are more susceptible to interruptions. Voice recognition dictation adjuncts to computerized charting boast increased provider efficiency; however, little is known about how data input of computerized documentation affects physician interruptions. Objectives: We present here observational interruptions data comparing two separate ED sites, one that uses • www.aemj.org S115 computerized charting by conventional techniques and one assisted by voice recognition dictation technology. Methods: A prospective observational quality initiative was conducted at two teaching hospital EDs located less than 1 mile from each other. One site primarily uses conventional computerized charting while the other uses voice recognition dictation computerized charting. Four trained observers followed ED physicians for 180 minutes during shifts. The tasks each ED physician performed were noted and logged in 30 second intervals. Tasks listed were selected from a predetermined standardized list presented at observer training. Tasks were also noted as either completed or placed in queue after a change in task occurred. A total of 4140 minutes were logged. Interruptions were noted when a change in task occurred with the previous task being placed in queue. Data were then compared between sites. Results: ED physicians averaged 5.33 interruptions/ hour with conventional computerized charting compared to 3.47 interruptions/hour with assisted voice recognition dictation (p = 0.0165). Conclusion: Computerized charting assisted with voice recognition dictation significantly decreased total per hour interruptions when compared to conventional techniques. Charting with voice recognition dictation has the potential to decrease interruptions in the ED allowing for more efficient workflow and improved patient care. 210 Physician Documentation of Critical Care Time While Working in the Emergency Department Jonathan W. Heidt, Richard Griffey Washington University School of Medicine in Saint Louis, Saint Louis, MO Background: The Current Procedural Terminology (CPT) Code 99291, describing the first 30–74 minutes of critical care provision, has been called ‘‘the most underreported code in the emergency department (ED).’’ We are not aware of publications describing successful interventions to improve documentation and coding in this area within the ED. We evaluated the effect of having medical coders provide feedback to emergency physicians (EPs), identifying their ED visits that may have qualified for code 99291, in order to improve critical care billing. Objectives: The aim of this study was to first determine the proportion of emergency department charts that would qualify for code 99291 based on services/care rendered, but lacked necessary documentation. The effect on proper critical care documentation was then determined after physicians were provided with individualized examples of patient records that may have qualified for critical care billing but lacked documentation. Methods: We conducted a retrospective record review at an urban academic ED with 90,000 annual patient visits, consisting of two 3-month periods, preceding and following an intervention consisting of feedback provided by ED coders to providers on their critical care documentation. We queried our electronic medical S116 2012 SAEM ANNUAL MEETING ABSTRACTS record (EMR) system for eligible visits, which included all ED patients admitted to an intensive care unit (ICU) during the study period. The intervention consisted of individualized e-mail feedback to EPs on their patient encounters that may have qualified for 99291 billing based upon complexity of services/care rendered but lacked supporting documentation. This feedback was ongoing during the post-intervention phase. The primary outcome measure was the proportion of visits documenting critical care time (99291) before as compared to after the intervention. Results: Among the 501 ICU admissions identified in the 3-month pre-intervention period, 88 (18%) documented critical care time. Following the intervention, among 382 ICU admissions identified, 243 (64%) charts included such documentation (P < 0.0001). Conclusion: In this single-center study, documentation of critical care time (99291) among ED patients admitted to an ICU significantly improved following individualized e-mail feedback by medical coders providing physician education. 211 Attitudes Toward Health Care Robot Assistants In The ED: A Survey Of ED Patients And Visitors Karen F. Miller, Wesley H. Self, Candace D. McNaughton, Lorraine C. Mion, Alan B. Storrow Vanderbilt University Medical Center, Nashville, TN Background: Using robot assistants in health care is an emerging strategy to improve efficiency and quality of care while optimizing the use of human work hours. Robot prototypes capable of performing vital signs and assisting with ED triage are under development. However, ED users’ attitudes toward robot assistants are not well studied. Understanding of these attitudes is essential to design user-friendly robots and to prepare EDs for the implementation of robot assistants. Objectives: To evaluate the attitudes of ED patients and their accompanying family and friends toward the potential use of robot assistants in the ED. Methods: We surveyed a convenience sample of adult ED patients and their accompanying adult family members and friends at a single, university-affiliated ED, 9/ 26/11–10/27/11. The survey consisted of eight items from the Negative Attitudes Towards Robots Scale (Normura et al.) modified to address robot use in the ED. Response options included a 5-point Likert scale. A summary score was calculated by summing the responses for all 8 items, with a potential range of 8 (completely negative attitude) to 40 (completely positive attitude). Research assistants gave the written surveys to subjects during their ED visit. Internal consistency was assessed using Cronbach’s alpha. Bivariate analyses were performed to evaluate the association between the summary score and the following variables: participant type (patient or visitor), sex, race, time of day, and day of week. Results: Of 121 potential subjects approached, 113 (93%) completed the survey. Participants were 37% patients, 63% family members or friends, 62% women, 79% white, and had a median age of 45.5 years (IQR 18–84). Cronbach’s alpha was 0.94. The mean summary score was 22.2 (SD = 0.87), indicating subjects were between ‘‘occasionally’’ and ‘‘sometimes’’ comfortable with the idea of ED robot assistants (see table). Men were more positive toward robot use than women (summary score: 24.6 vs 20.8; p = 0.033). No differences in the summary score were detected based on participant type, race, time of day, or day of week. Conclusion: ED users reported significant apprehension about the potential use of robot assistants in the ED. Future research is needed to explore how robot designs and strategies to implement ED robots can help alleviate this apprehension. Table - Abstract 211: *Reverse scored. Item Mean score (SD) I would feel relaxed talking to robots in the ED.* I would feel uneasy if I were in an ED where I had to use robots. I would feel nervous about using a robot in front of other people in the ED. I would dislike the idea that nurses and doctors rely on robots to make judgments in ED care. I feel that if nurses and doctors depend on robots too much in the ED bad things might happen. I would feel uncomfortable with a robot watching me in the ED. I am concerned robots would be a bad influence on my care in the ED. I feel that in the future, robots will be a valuable part of ED patient care.* Total (summary score) 212 2.3 (1.3) 2.9 (1.5) 3.1 (1.6) 2.5 (1.5) 2.7 (1.3) 3.1 (1.6) 3.0 (1.5) 2.8 (1.4) 22.2 (0.87) The Effectiveness Of A Nurse Telephone Triage Protocol For Emergency Department Disposition During The H1N1 Epidemic Of 2009 Jeremy R. Monroe, Christopher M. Verdick, John W. Hafner, Huaping Wang University of Illinois College of Medicine at Peoria, Peoria, IL Background: The H1N1 influenza epidemic affected already overcrowded ED resources and highlighted the need for effective patient triage during widespread contagious outbreaks. Adequate public triage is necessary to avoid dangerous ED overcrowding while identifying patients needing advanced medical care. Nurse-assisted public telephone triage represents a simple but possibly effective tool that can screen large numbers of patients with influenza-like illness. Objectives: This study evaluates the effectiveness of a nurse telephone triage protocol during the H1N1 epidemic of 2009. Methods: A retrospective observational cohort trial was conducted of all patients contacting a hospitalbased nurse telephone triage service during the 2009 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 H1N1 epidemic (peak community health department prevalence 9/28/09–11/9/09). Patients were screened using an adapted CDC criteria protocol for influenzalike illness and further assessed for illness severity and complicating medical conditions; triage logs were recorded in an associated hospital EMR (Epic 2010, Epic Systems Inc.). Triage calls utilizing the protocol, as well as any associated outpatient or ED visits over the next 24 hours, were queried. Patient demographics, interventions, and disposition were abstracted. Ten nurse triage dispositions were grouped into four categories (home care, physician notification, outpatient physician visit, and ED visit). Group differences were analyzed using chi-square tests. Results: Three-hundred fifty triage calls (287 pediatric, mean age 7.8 years; 63 adult, mean age 41.3 years) were documented. Overall, 254 (72.6%) patients followed the recommend disposition. Patients triaged to outpatient physician visits had higher compliance than those triaged to the ED (85.9% vs. 53.3%, p < 0.01).Twenty-seven patients (18.5%) were evaluated in the ED and 20/27 (74.1%) were diagnosed with influenza-related illness; no ED visits required hospital admission. The telephone triage protocol was highly specific (0.95, 95% CI 0.94–0.97) but poorly sensitive (0.26, 95% CI 0.13–0.42) for predicting ultimate patient disposition. Conclusion: In our population, during the 2009 H1N1 influenza epidemic, a nurse telephone triage was effective in delineating ultimate patient disposition. Nurse telephone triage may be one means to adequately distribute medical resources during epidemics. 213 ED Impact of Rh Factor Testing in Patients with First Trimester Vaginal Bleeding Raviraj Patel, Nicholas Genes Mount Sinai School of Medicine, New York, NY Background: Management of patients with first trimester vaginal bleeding in the emergency department commonly involves checking blood for Rh factor and, if negative, dosing Rho(D) immune globulin, to prevent isoimmunization and protect future instances of Rh hemolytic disease. Despite the lack of evidence supporting the effectiveness, this practice continues. The effect of Rh testing on ED resource utilization has not been studied, however, and if significant, may help prompt a change in practice. Objectives: We sought to determine the effect on ED resources of Rh factor testing on patients with first trimester vaginal bleeding. Methods: We retrospectively reviewed operations data from our large urban academic ED from June 1 to September 30, 2011, to determine the number of patients discharged who presented with first trimester bleeding, their length of stay compared to other discharged patients, the turnaround time for blood typing, and frequency of Rh immune globulin administration. Results: Based on discharge diagnoses during the study period, we identified 311 patients with first trimester vaginal bleeding. Their ED length-of-stay (LOS) • www.aemj.org S117 was 292 minutes (SD ± 176 min), 42.9 minutes longer than the average LOS of 16,946 discharged patients in that period (95% CI 11.8–73.9 minutes). Blood type turnaround time was 198 minutes (SD ± 82.3 minutes). Rh immune globulin was administered 15 times. Conclusion: Patients discharged with first trimester vaginal bleeding have a significantly longer than average LOS; the long turnaround time for Rh factor determination likely plays a major role in LOS. Just 4.8% of patients with first trimester bleeding received Rh immune globulin, reflecting the low prevalence of Rh negativity in the population and underscoring the inefficiency of current practice. If blood type turnaround time was subtracted from LOS for these patients, a total of 29,300 bed minutes would be saved over the study period, or 38.0 hours / week. (Originally submitted as a ‘‘late-breaker.’’) 214 Are We Punishing Hospitals for Progressive Treatment of Atrial Fibrillation? Nicole E. Piela, Alfred Sacchetti, Darius Sholevar, Reginald Blaber, Steven Levi Our Lady of Lourdes Medical Center, Camden, NJ Background: Emergency department cardioversion (EDC) of recent-onset atrial fibrillation or flutter (AF) patients is an increasingly common management approach to this arrhythmia. Patients who qualify for EDC generally have few co-morbidities and are often discharged directly from the ED. This results in a shift towards a sicker population of patients admitted to the hospital with this diagnosis. Objectives: To determine whether hospital charges and length of stay (LOS) profiles are affected by emergency department discharge of AF patients. Methods: Patients receiving treatment at an urban teaching community hospital with a primary diagnosis of atrial fibrillation or flutter were identified through the hospital’s billing data base. Information collected on each patient included date of service, patient status, length of stay, and total charges. Patient status was categorized as inpatient (admitted to the hospital), observation (transferred from the ED to an inpatient bed but placed in an observation status), or ED (discharged directly from the ED). The hospital billing system automatically defaults to a length of stay of 0 for observation patients. ED patients were assigned a length of stay of 0. Total hospital charges and mean LOS were determined for two different models: a standard model (SM) in which patients discharged from the ED were excluded from hospital statistics, and an inclusive model (IM) in which discharged ED patients were included in the hospital statistics. Statistical analysis was through ANOVA. Results: A total of 317 patients were evaluated for AF over an 18–month period. Of these, 197 (62%) were admitted, 22 (7%) were placed in observation status, and 98 (31%) were discharged from the ED. Hospital charges and LOS in days are summarized in the table. All differences were statistically significant at (p < 0.001). S118 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: Emergency department management can lead to a population of AF patients discharged directly from the ED. Exclusion of these patients from hospital statistics skews performance profiles effectively punishing institutions for progressive care. Table - Abstract 214: Group Standard Model Inclusive Model 215 Mean Hospital Charges (95% CI) Mean LOS (95% CI) $47,542 ($41–54,000) $34,063 ($29–39,000) 3.42 (3.0–3.8) 2.37 (2.0–2.7) A Comparison of Two Hospital Electronic Medical Record Systems and Their Effects on the Relationship Between Physician Charting and Patient Contact John Shabosky, Matthew Albrecht, Jonathan de la Cruz Southern Illinois University School of Medicine, Springfield, IL Background: Recent health care reform has placed an emphasis on the electronic health record (EHR). With the advent of the EHR it is common to see ED providers spending more time in front of computers documenting and away from patients. Finding strategies to decrease provider interaction with computers and increase time with patients may lead to improved patient outcomes and satisfaction. Computerized charting adjuncts, such as voice recognition software, have been marketed as ways to improve provider efficiency and patient contact. Objectives: We present here observational data comparing two separate ED sites, one where computerized charting is done by conventional techniques and one that is assisted with voice recognition dictation, and their effects on physican charting and patient contact. Methods: A prospective observational quality initiative was conducted at two teaching hospitals located less than 1 mile from each other. One site primarily uses conventional computerized charting while the other uses voice recognition dictation. Four trained quality assistants observed ED physicians for 180 minutes during shifts. The tasks each physician performed were noted and logged in 30 second intervals. Tasks listed were identified from a predetermined standardized list presented at observer training. A total of 4140 minutes were logged. Time allocated to charting and that allocated to direct patient care were then compared between sites. Results: ED physicians spent 28.6% of their time charting using conventional techniques vs 25.7% using voice recognition dictation (p = 0.4349). Time allocated to direct patient care was found to be 22.8% with conventional charting vs 25.1% using dictation (p = 4887). In total, ED physicians using conventional charting techniques spent 668/2340 minutes charting. ED physicians using voice recognition dictation spent 333/1800 minutes dictating and an additional 129.5/1800 minutes reviewing or correcting their dictations. Conclusion: The use of voice recognition assisted dictation rather than conventional techniques did not significantly change the amount of time physicians spent charting or with direct patient care. Although voice recognition dictation decreased initial input time of documenting data, a considerable amount of time was required to review and correct these dictations. 216 Emergency Department Rectal Temperatures are Frequently Discordant from Initial Triage Temperatures Daniel Runde1, Daniel Rolston1, Graham Walker2, Jarone Lee3 1 St. Luke’s Roosevelt, New York, NY; 2Stanford University, Palo Alto, CA; 3Massachusetts General Hospital, Boston, MA Background: Fever in patients can provide important clues to the etiology of a patient’s symptoms. Non-invasive temperature sites (oral, axillary, temporal) may be insensitive due to a variety of factors. This has not been well-studied in adult emergency department (ED) patients. Objectives: For our primary objective, we studied whether emergency department triage temperatures detected fever adequately when compared to a rectal temperature. As secondary objectives, we examined the temperature differences when a rectal temperature was taken within an hour of non-invasive temperature, temperature site (oral, axillary, temporal), and also examined the patients that were initially afebrile but were found to be febrile by rectal temperature. Methods: We performed an electronic chart review at our inner city, academic emergency department with an annual census of 110,000 patients. We identified all patients over the age of 18 who received a non-invasive triage temperature and a subsequent rectal temperature while in the ED from January 2002 through February 2011. Specific data elements included many aspects of the patient’s medical record (e.g. subject demographics, temperature, and source). We analyzed our data with standard descriptive statistics, t-tests for continuous variables, and Pearson chi-square tests for proportions. Results: A total of 27,130 patients met our inclusion criteria. The mean difference in temperatures between the initial temperature and the rectal temperature was 1.3F, with 25.9% having higher rectal temperatures ‡2F, and 5.0% having higher rectal temperatures ‡4F. The mean temperature difference among the 10,313 patients who an initial noninvasive temperature and a rectal temperature within one hour was 1.4F. The mean difference among patients that received oral, axillary, and temporal temperatures was 1.2F, 1.8F, and 1.2F respectively. Approximately one in five patients (18.1%) were initially afebrile and found to be febrile by rectal temperature, with an average temperature difference of 2.5F. These patients had a higher rate of admission, and were more likely to be admitted to the intensive care unit. Conclusion: There are significant differences between rectal temperatures and non-invasive triage temperatures in this emergency department cohort. In almost ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 one in five patients, fever was missed by triage temperature. 217 Direct Bedding, Bedside Registration, and Patient Pooling to Improve Pediatric Emergency Department Length of Stay Niel F. Miele, Neelam R. Patel, Rachel D. Grieco, Ernest G. Leva University of Medicine and Dentistry of New Jersey, New Brunswick, NJ Background: Pediatric emergency department (PED) overcrowding has become a national crisis, and has resulted in delays in treatment, and patients leaving without being seen. Increased wait times have also been associated with decreased patient satisfaction. Optimizing PED throughput is one means by which to handle the increased demands for services. Various strategies have been proposed to increase efficiency and reduce length of stay (LOS). Objectives: To measure the effect of direct bedding, bedside registration, and patient pooling on PED wait times, length of stay, and patient satisfaction. Methods: Data were extracted from a computerized ED tracking system in an urban tertiary care PED. Comparisons were made between metrics for 2010 (23,681 patients) and the 3 months following process change (6,195 patients). During 2010, patients were triaged by one or two nurses, registered, and then sent either to a 14-bed PED or a physically separate 5-bed fast-track unit, where they were seen by a physician. Following process change, patients were brought directly to a bed in the 14-bed PED, triaged and registered, then seen by a physician. The fast-track unit was only utilized to accommodate patient surges. Results: Anticipating improved efficiencies, attending physician coverage was decreased by 9%. After instituting process changes, improvements were noted immediately. Although daily patient volume increased by 3%, median time to be seen by a physician decreased by 20%. Additionally, median LOS for discharged patients decreased by 15%, and median time until the decisionto-admit decreased by 10%. Press-Ganey satisfaction scores during this time increased by greater than 5 mean score points, which was reported to be a statistically significant increase. Conclusion: Direct bedding, bedside registration, and patient pooling were simple to implement process • www.aemj.org S119 changes. These changes resulted in more efficient PED throughput, as evidenced by decreased times to be seen by a physician, LOS for discharged patients, and time until decision-to-admit. Additionally, patient satisfaction scores improved, despite decreased attending physician coverage and a 30% decrease in room utilization. Table - Abstract 217: Patients/day Median time to be seen–M.D. Median LOS for D/C Median decision to admit 218 Year 2010 April 2011 May 2011 June 2011 65 0:39 2:00 3:12 67 0:31 1:42 3:12 68 0:28 1:42 2:54 67 0:30 1:42 2:36 Are Emergency Physicians More CostEffective in Running an Observation Unit? Margarita E. Pena, Robert B. Takla, Susan Szpunar, Steve Kler St. John Hospital and Medical Center, Detroit, MI Background: There are no studies comparing costeffectiveness when an observation unit (OU) is managed and staffed by EM physicians versus non-EM physicians. Objectives: To compare cost-effectiveness when the same OU is managed and staffed by EM physicians versus non-EM physicians. Methods: This was an observational, retrospective data collection study of a 30-bed OU in an urban teaching hospital. Three time periods were compared: November 2007 to August 2008 (period 1), November 2008 to August 2009 (period 2), and November 2010 to August 2011 (period 3). During period 1, the OU was managed by the internal medicine department and staffed by primary care physicians and physician assistants. During periods 2 and 3, the OU was managed and staffed by EM physicians. Data collected included OU patient volume, length of stay (LOS) for discharged and admitted patients, admission rates, and 30-day readmission rates for discharged patients. Cost data collected included direct, indirect, and total cost per patient encounter. Data were compared using chi-square and ANOVA analysis followed by multiple pairwise comparisons using the Bonferroni method of p-value adjustment. Table - Abstract 218: Characteristic OU volume/month % of ED volume LOS (hours) discharged LOS (hours) admitted Admission rate 30-day readmission rate Direct cost Indirect cost Total cost Period 1 576.2 ± 7.1% 27.3 ± 20.7 ± 32.5% 11.6% 1367.9 ± 817.4 ± 2185.4 ± 10.4 1.7 2.2 1055.0 552.5 1579.6 Period 2 620.1 ± 7.1% 17.3 ± 16.5 ± 21.6% 7.7% 1018.7 ± 592.8 ± 1611.5 ± 66.7 1.3 3.0 759.6 462.9 1156.3 Period 3 758.0 ± 7.9% 16.9 ± 15.0 ± 19.6% 7.9% 938.0 ± 938.0 ± 1592.3 ± 34.2 0.4 0.44 743.0 743.0 1199.8 p-values 1 vs 3, 2 vs 3, p < 0.0001 p < 0.0001 1 vs 2, 1 vs 3, p < 0.0001 1 vs 2 p = 0.001, 1 vs 3, p < 0.0001 p < 0.0001 p < 0.0001 All comparisons, p < 0.0001 All comparisons, p < 0.0001 All comparisons, p < 0.0001 S120 2012 SAEM ANNUAL MEETING ABSTRACTS Results: See table. The OU patient volume and percent of ED volume was greater in period 3 compared to periods 1 and 2. Length of stay, admission rates, 30-day readmission rates, and costs were greater in period 1 compared to periods 2 and 3. Conclusion: EM physicians provide more cost-effective care for patients in this large OU compared to non-EM physicians, resulting in shorter LOS for admitted and discharged patients, greater rates of patients discharged, and less 30-day readmission rates for discharged patients. This is not affected by an increase in OU volume and shows a trend towards improvement. 219 A Long Term Analysis of Physician Screening in the Emergency Department Jonathan G. Rogg1, Benjamin A. White2, Paul Biddinger2, Yuchiao Chang2, David F. M. Brown2 1 Harvard Affiliated Emergency Medicine 2 Residency, Boston, MA; Massachusetts General Hospital, Boston, MA Background: Emergency department (ED) crowding continues to be a problem, and new intake models may represent part of the solution. However, little data exist on the sustainability and long-term effects of physician triage and screening on standard ED performance metrics, as most studies are short-term. Objectives: We examined the hypothesis that a physician screening program (START) sustainably improves standard ED performance metrics including patient length of stay (LOS) and patients who left without completing assessment (LWCA). We also investigated the number of patients treated and dispositioned by START without using a monitored bed and the median patient door-to-room time. Methods: Design and Setting: This study is a retrospective before-and-after analysis of START in a Level I tertiary care urban academic medical center with approximately 90,000 annual patient visits. All adult patients from December 2006 until November 2010 are included, though only a subset was seen in START. START began at our institution in December 2007. Observations: Our outcome measures were length of stay for ED patients, LWCA rates, patients treated and dispositioned by START without using a monitored bed, and door-to-room time. Statistics: Simple descriptive statistics were used. P-values for LOS were calculated with Wilcoxon test and p-value for LWCA was calculated with chi-square. Results: Table 2 shows median length of stay for ED patients was reduced by 56 minutes/patient (p-value <0.0001) when comparing the most recent year to the year before START. Patients who LWCA were reduced from 4.8% to 2.9% (p-value <0.0001) during the same time period. We also found that in the first half-year of START, 18% of patients screened in the ED were treated and dispositioned without using a monitored bed and by the end of year 3, this number had grown to 29%. Median door-to-room time decreased from 18.4 minutes to 9.9 minutes over the same period of time. Conclusion: A START system can provide sustained improvements in ED performance metrics, including a significant reduction in ED LOS, LWCA rate, and doorto-room time. Additionally, START can decrease the need for monitored ED beds and thus increase ED capacity. Table 1 - Abstract 219: Provides demographic data for patients eligible for START from December 2006–November 2007 and actual START volume after December 2007. Dec 09 – Nov 07 START volume Age, median (IQR) Male, population (%) Hospital characteristics ED volume Boarders per day, median (IQR) Boarding hours per day, median (IQR) Boarding hours per patient, median (IQR) Dec 07 – Nov 08 39142 43 (24–61) 19245 (49.2) 42723 43 (24–61) 20952 (49) 81578 44 (39–49) 179.5 (118.6–246.9) 2.31 (0.92–5.46) 85551 44 (37–50) 180.7 (115.3–258.1) 2.21 (0.88–5.47) Dec 08 – Nov 09 48756 41 (22–60) 23944 (49.1) 91428 43 (36–50) 172 (111.7–250.3) 2.05 (0.78–5.19) Dec 09 – Nov 10 50249 43 (23–61) 24966 (49.7) 91395 43 (35–48) 179 (100.6–246.7) 2.08 (0.8–5.76) Table 2 - Abstract 219: Length of Stay and LWCA ED length of Stay overall (min), median (IQR) Discharged Patients (min), median (IQR) Admitted Patients (min), median (IQR) Other Disposition (min), median (IQR) LWCA Dec 06–Nov 07 Dec 07–Nov 08 Dec 08–Nov 09 Dec 09–Nov 10 diff* p-value 362 (234–544) 342 (216–348) 310 (193–478) 306 (189–477) 56 <0.0001 317 (208–475) 293 (188–447) 263 (168–399) 257 (163–394) 60 <0.0001 461 (320–678) 456 (311–666) 431 (298–635) 425 (287–639) 36 <0.0001 208 (115–329) 152 (82–267) 138 (74–247) 139 (70–251) 69 <0.0001 4.80% 3.10% 2.70% 2.90% 1.90% <0.0001 * difference is comparing year prior to START to the most recent year ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 220 Professional Translation Does Not Result In Decreased Length Of Stay For Spanish Speaking Patients With Abdominal Pain Otar Taktakishville, Gregory Garra, Adam J. Singer Stony Brook University, Stony Brook, NY Background: Language discordance is the most frequently reported communication barrier (CB) with patients. CBs are associated with decreased diagnostic confidence, increased diagnostic test utilization, and increased ED length of stay (LOS). Objectives: Our primary objective was to determine whether professional translation results in decreased ED LOS for patient with abdominal pain. Our secondary objective was to determine differences in test/ consult utilization and disposition. Our hypothesis was that professional translation service would result in a 1hour decrease in LOS. Methods: Study design: Prospective observational. Setting: University ED with 90,000 visits/ yr. Subjects: Spanish-speaking patients presenting to the ED for abdominal pain. Measures: An anonymous survey tool was completed by the treating physician. Data collected included demographics, triage time, disposition time, type of translation method utilized, and ancillary testing and consultations obtained in the ED. Analysis: descriptive statistics. Continuous variables were compared with analysis of variance (ANOVA), and binary variables were compared with phi coefficient. Results: Ninety-two patients were enrolled; mean age was 35 (IQR 27–42), 76% were female. The median ED LOS was 270 min (IQR 199–368). Labs were obtained in 98%, CT in 37%, US in 30%, and consultation in 23%. 18% of the cohort was admitted to the hospital. The most commonly utilized source of translation was a layman (35%). A professional translator was used in 9% and translation service (language line, MARTY) in 30%. The examiner was fluent in the patient’s language in 11%. Both the patient and examiner were able to maintain basic communication in 11%. There were 47 patients in the professional/ fluent translation group and 44 patients in the lay translation group. There was no difference in ED LOS between groups 288 vs 304 min; p = 0.6. There was no difference in the frequency of lab tests, computerized tomography, ultrasound, consultations, or hospital admission. Frequencies did not differ by sex or age. Conclusion: Translation method was not associated with a difference in overall ED LOS, ancillary test use, or specialist consultation in Spanish-speaking patients presenting to the ED for abdominal pain. 221 Emergency Department Patients on Warfarin - How Often Is the Visit Due to the Medication? Jim Killeen, Edward Castillo, Theodore Chan, Gary Vilke UCSD Medical Center, San Diego, CA Background: Warfarin has important therapeutic value for many patients, but has been associated with signi- • www.aemj.org S121 ficant bleeding complications, hypersensitivity reactions, and drug-drug interactions, which can result in patients seeking care in the emergency department (ED). Objectives: To determine how often ED patients on warfarin present for care as a result of the medication itself. Methods: A multi-center prospective survey study in two academic EDs over 6 months. Patients who presented to the ED taking warfarin were identified, and ED providers were prospectively queried at the time of disposition regarding whether the visit was the result of a complication or side effect associated with warfarin. Data were also collected on patient demographics, chief complaint, triage acuity, vital signs, disposition, ED evaluation time, and length of stay (LOS). Patients identified with a warfarin-related cause for their ED visit were compared with those who were not. Statistical analysis was performed using descriptive statistics. Results: During the study period, 31,500 patients were cared for by ED staff, of whom 594 were identified as taking warfarin as part of their medication regimen. Of these, providers identified 54.7% (325 patients) who presented with a warfarin-related complication as their primary reason for the ED visit. 56.9% (338) of patients were seen at an academic facility and 43.1% (256) patients were seen at a community hospital. 53.4% (317) were male patients and 46.6% (277) were females, with 42.3% (251) over the age of 65 vs. 57.7% (343) under the age of 65. Providers admitted 33.8% (201) of patients to the hospital while 63.1% (375) were discharged home. Providers identified 8.1% (48) of the admitted patients were attributed to a bleeding complication. Patients with a warfarin-related ED visit were more likely to be triaged at an urgent level 75.3% (447) vs. emergent 14.3% (85) or non-urgent 10.4% (62). There was no significant difference between ED evaluation times by patients with active bleeding complaints vs. non-active bleeding complaints. There was no statistical difference between ED evaluation time for all patients identified as taking warfarin from triage. Conclusion: Half of all patients identified as taking warfarin from triage were in the ED for a complication related to warfarin use. 222 Ultrasound in Triage in Patients at Risk for Ectopic Pregnancy Decreases Emergency Department Length of Stay Kenneth J. Cody, Daniel Jafari, Nova L. Panebianco, Olan A. Soremekun, Anthony J. Dean University of Pennsylvania, Philadelphia, PA Background: First trimester abdominal pain and vaginal bleeding are common ED complaints. ED overcrowding leads to long wait times, delayed treatment, and high left without being seen (LWBS) rates in these patients. Objectives: To determine whether receiving a pelvic ultrasound in triage (TUS group) decreased ED length of stay (LOS) and LWBS rate compared to routine ED care (EDUS group). Methods: We prospectively enrolled a convenience sample of patients and matched with historic controls S122 2012 SAEM ANNUAL MEETING ABSTRACTS receiving routine care. Study setting: urban academic ED with an annual census of 58,000. Inclusion criteria were women ages 16 to 49 arriving during pre-determined high volume periods with a positive urine b-HCG, Emergency Severity Index of 3, and one or more of the following: abdominal or pelvic pain, vaginal bleeding, dizziness, or syncope. Exclusion criteria: documented intrauterine pregnancy (IUP), hemodynamic instability, or assisted reproductive technique use. After initial triage evaluation, eligible patients received a pelvic ultrasound while still in triage. Controls were matched to cases by time of visit and ED census. Routine care for EDUS controls consisted of triage, then transfer to the main ED when a bed became available. Pelvic ultrasound was performed after room placement. ED physicians performed all ultrasounds for cases and controls. After the ultrasound, both groups received similar care: those with an identified IUP were discharged. Those with no IUP received radiology ultrasound and gynecology evaluation. LOS was defined as patient intake time to disposition time. LWBS rate was determined during enrollment and control periods. Results: 27 TUS cases were enrolled and compared to 27 EDUS controls. The groups were similar with respect to age, ethnicity, parity, gestational age, presenting symptoms, medical history, ED census, and final diagnosis. T-test results showed TUS group LOS was significantly shorter than that of EDUS (249 minutes [95%CI 203–293] vs. 497 minutes [95%CI 416–579]; p = 0.0003). The LWBS rate for EDUS trended higher than that of TUS (6% vs. 20%; p = 0.1, Fisher’s exact test). Conclusion: Emergency ultrasound in triage significantly reduced LOS in patients presenting with possible ectopic pregnancy. There was a trend towards decreased LWBS rate in the TUS group. 223 Boarding and Press Ganey Patient Satisfaction Scores Among Discharged Patients - Quantifying the Relationship Paris B. Lovett, Frederick T. Randolph, Rex G. Mathew Thomas Jefferson University, Philadelphia, PA Background: Long wait times, long length of stay, use of hallway beds, and physical crowding have all been reported to negatively affect patient satisfaction. There is a need for quantitative assessment of the relationship among patients discharged from the ED. Objectives: To describe the association between total boarding hours for given calendar days, and mean Press Ganey patient satisfaction raw scores (PGs) on those days. To determine a quantitative coefficient for the relationship. Methods: We measured total hours of boarding (stays greater than two hours after admission decision) for each calendar day in a nine month period. We obtained mean PGs for the same dates (by date of visit). A linear regression analysis was performed. Results: Scatter plots with regression lines are shown in the figure. The relationships were statistically significant. See the table for regression data. Conclusion: Our research supports prior reports that boarding has a negative affect on patient satisfaction. Each 100 hours of daily boarding is associated with a drop of 1.3 raw score points in both PG metrics. These seemingly small drops in raw scores translate into major changes in rankings on Press Ganey national percentile scales (a difference of as much as 10 percentile points). Our institution commonly has hundreds of hours of daily boarding. It is possible that patient-level measurements of boarding impact would show stronger correlation with individual satisfaction scores, as opposed to the daily aggregate measures we describe here. Our research suggests that reducing the burden of boarding on EDs will improve patient satisfaction. 224 Effect of Day of Week on Number of Patients Transferred to a Tertiary Care Emergency Department Wendy L. Woolley, Daniel K. Pauze, Denis R. Pauze, Dennis P. McKenna, Wayne R. Triner Albany Medical Center, Albany, NY Background: The lack of subspecialty coverage in many EDs often results in patient transfers for higher levels of care. Many hospitals do not have consistent daily coverage by such specialists, and frequently lack weekend coverage. As the gaps in coverage widen, transfer rates to tertiary referral EDs are thought to increase. This can significantly effect both the sending and receiving EDs’ throughput, efficiency, and ultimately patient satisfaction and safety. Objectives: We sought to determine whether the day of the week has any effect on the number of patients transferred and their disposition. Methods: A retrospective chart-review of patient transfers from January - December 2010 into a Level I, tertiary care ED with an annual patient volume of 72,000 visits. Transfer center database was queried for day ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 and date of transfer, requested specialty service, and patient disposition. Results: In 2010 a total of 3888 patients were received in transfer. Those cases where no specific specialty service was requested by the transferring provider were excluded. The remaining 3707 patients were analyzed using descriptive and statistical analysis. 33.4% of these transfers (1238) were received on Saturday or Sunday (weekend) as opposed to 66.6% (2469) patients received Monday through Friday (weekday). Mean number of patient transfers per day were as follows: Monday 9.8, Tuesday 9.3, Wednesday 9.1, Thursday 9.1, Friday 9.9, Saturday 11.5, Sunday 12.3. When dichotomized to weekend or weekday, the mean for total number of transfers per day was greater for weekends (p < 0.01). The admission rate on weekends was 75.9% compared to 80.5% on weekdays (risk ratio 1.23, CI 1.09, 1.29). Conclusion: Our data show a significantly disproportionate increase in the number of patients received in transfers to our tertiary care ED on the weekends with a lower likelihood of hospital admission. Maintaining subspecialty coverage is a national challenge that seems to be worsening with the current health care crisis. The potential for increased volume due to patient transfers on the weekends should be considered by tertiary care centers when making staffing and on-call decisions. Further discussion may center on the types of specialty coverage needed during these times. 225 The Impact Of Increased Output Capacity Interventions On Emergency Department Length Of Stay And Patient Flow Hallam M. Gugelmann1, Olanrewaju A. Soremekun1, Elizabeth M. Datner1, Asako C. Matsuura1, Jesse M. Pines2 1 Department of Emergency Medicine, University of Pennsylvania, Philadelphia, PA; 2 Departments of Emergency Medicine and Health Policy, George Washington University, Washington, DC Background: Prolonged emergency department (ED) boarding is a key contributor to ED crowding. The • www.aemj.org S123 effect of output interventions (moving boarders out of the ED into an intermediate area prior to admission or adding additional capacity to an observation unit) has not been well studied. Objectives: We studied the effect of a combined observation-transition (OT) unit, consisting of observation beds and an interim holding area for boarding ED patients, on the length of stay (LOS) for admitted patients, as well as secondary outcomes such as LOS for discharged patients, and left without being seen rates. Methods: We conducted a retrospective review (12 months pre-, 12 months post-design) of an OT unit at an urban teaching ED with 59,000 annual visits (study ED). We compared outcomes to a nearby communitybased ED with 38,000 annual visits in the same health system (control ED) where no capacity interventions were performed. The OT had 17 beds, full monitoring capacity, and was staffed 24 hours per day. The number of beds allocated to transition and observation patients fluctuated throughout the course of the intervention, based on patient demands. All analyses were conducted at the level of the ED-day. Wilcoxon rank-sum and analysis of covariance tests were used for comparisons; continuous variables were summarized with medians. Results: In unadjusted analyses, median daily LOS of admitted patients at the study ED was 31 minutes lower in the 12 months after the OT opened, 6.98 to 6.47 hours (p < 0.0001). Control site daily LOS for admitted patients increased 26 minutes from 4.52 to 4.95 hours (p < 0.0001). Results were similar after adjusting for other covariates (day of week, ED volume, and triage level). LOS of discharged patients at study ED decreased by 14 minutes, from 4.1 hours to 3.8 hours (p < 0.001), while the control ED saw no significant changes in discharged patient LOS (2.6 hours to 2.7 hours, p = 0.06). Left without being seen rates did not decrease at either site. Conclusion: Opening an OT unit was associated with a 30-minute reduction in average daily ED LOS for admitted patients and discharged patients in the study ED. Given the large expense of opening an OT, future studies should compare capacity-dependent (e.g., OT) vs. capacity-independent (e.g, organizational) interventions to reduce ED crowding. S124 226 2012 SAEM ANNUAL MEETING ABSTRACTS The Epidemiology of Pelvic Inflammatory Disease in a Pediatric Emergency Department Fran Balamuth, Katie Hayes, Cynthia Mollen, Monika Goyal Children’s Hospital of Philadelphia, Philadelphia, PA Background: Lower abdominal pain and genitourinary problems are common chief complaints in adolescent females presenting to emergency departments. Pelvic inflammatory disease (PID) is a potentially severe complication of lower genital tract infections, which involves inflammation of the female upper genital tract secondary to ascending STIs. PID has been associated with severe sequelae including infertility, ectopic pregnancy, and chronic pelvic pain. We describe the prevalence and microbial patterns of PID in a cohort of adolescent females presenting to an urban emergency department with abdominal or genitourinary complaints. Objectives: To describe the prevalence and microbial patterns of PID in a cohort of adolescent patients presenting to an ED with lower abdominal or genitourinary complaints. Methods: This is a secondary analysis of a prospective study of females ages 14–19 years presenting to a pediatric ED with lower abdominal or genitourinary complaints. Diagnosis of PID was per 2006 CDC guidelines. Patients underwent Chlamydia trachomatis (CT) and Neisseria gonorrhea (GC) testing via urine APTIMA Combo 2 Assay and Trichomonas vaginalis (TV) testing using the vaginal OSOM Trichomonas rapid test. Descriptive statistics were performed using STATA 11.0. Results: The prevalence of PID in this cohort of 328 patients was 19.5% (95% CI 15.2%, 23.8%), 37.5% (95% CI 25.3%, 49.7%) of whom had positive sexually transmitted infection (STI) testing: 25% (95% CI 14.1%, 35.9%) with CT, 7.8% (95% CI 1.1, 14.6%) with GC, and 12.5% (95% CI 4.2%, 20.8%) with TV. 84.4% (95% CI 75.2, 93.5%) of patients diagnosed with PID received antibiotics consistent with CDC recommendations. Patients with lower abdominal pain as their chief complaint were more likely to have PID than patients with genitourinary complaints (OR 3.3, 95% CI 1.7, 6.4). Conclusion: A substantial number of adolescent females presenting to the emergency department with lower abdominal pain were diagnosed with PID, with microbial patterns similar to those previously reported in largely adult, outpatient samples. Furthermore, appropriate treatment for PID was observed in the majority of patients diagnosed with PID. 227 Impact Of Maternal Ultrasound Implementation In Rural Clinics In Mali Melody Eckardt1, Roy Ahn1, Raquel Reyes1, Elizabeth Cafferty1, Kathryn L. Conn1, Alison Mulcahy2, Jean Crawford1, Thomas F. Burke1 1 Department of Emergency Medicine, Division of Global Health and Human Rights, Massachusetts General Hospital, Boston, MA; 2 Alameda County Medical Center, Highland Hospital, Oakland, CA Background: In resource-poor settings, maternal health care facilities are often underutilized, contributing to high maternal mortality. The effect of ultrasound in these settings on patients, health care providers, and communities is poorly understood. Objectives: The purpose of this study was to assess the effect of the introduction of maternal ultrasound in a population not previously exposed to this intervention. Methods: An NGO-led program trained nurses at four remote clinics outside Koutiala, Mali, who performed 8,339 maternal ultrasound scans over three years. Our researchers conducted an independent assessment of this program, which involved log book review, sonographer skill assessment, referral follow-up, semi-structured interviews of clinic staff and patients, and focus groups of community members in surrounding villages. Analyses included the effect of ultrasound on clinic function, job satisfaction, community utilization of prenatal care and maternity services, alterations in clinical decision making, sonographer skill, and referral frequency. We used QRS NVivo9 to organize qualitative findings, code data, and identify emergent themes, and GraphPad software (La Jolla, CA) and Microsoft Excel to tabulate quantitative findings Results: -Findings that triggered changes in clinical practice were noted in 10.1% of ultrasounds, with a 3.5% referral rate to comprehensive maternity care facilities. -Skill retention and job satisfaction for ultrasound providers was high. -The number of patients coming for antenatal care increased, after introduction of ultrasound, in an area where the birth rate has been decreasing. -Over time, women traveled from farther distances to access ultrasound and participate in antenatal care. -Very high acceptance among staff, patients and community members. -Ultrasound was perceived as most useful for finding fetal position, sex, due date, and well-being. -Improved confidence in diagnosis and treatment plan for all cohorts. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 -Improved compliance with referral recommendations. -No evidence of gender selection motivation for ultrasound use. Conclusion: Use of maternal ultrasound in rural and resource-limited settings draws women to an initial antenatal care visit, increases referral, and improves job satisfaction among health care workers. 228 Predicting Return Visits for Patients Evaluated in the Emergency Department with Nausea and Vomiting of Pregnancy Brian Sharp, Kristen Sharp, Suzanne DooleyHash University of Michigan Hospital, Ann Arbor, MI Background: Nausea and vomiting in pregnancy is a condition experienced by up to 50% of pregnant women. Despite adequate ED treatment of their symptoms and frequent utilization of observational emergency medicine, these patients have high rates of subsequent repeat ED visits. Objectives: To evaluate what factors are predictive of return visits to the ED in patients presenting with nausea and vomiting of pregnancy. Methods: A retrospective database analysis was conducted using the electronic medical record from a single, large academic hospital. ED patients who received a billing diagnosis of ‘‘nausea and vomiting of pregnancy’’ or ‘‘hyperemesis gravidarum’’ between 1/1/10 and 12/31/10 were selected. A manual chart review was conducted with demographic and treatment variables collected. Statistical significance was determined using multiple regression analysis for a primary outcome of return visit to the emergency department for nausea and vomiting of pregnancy. Results: 113 patients were identified. The mean age was 27.1 years (SD±5.25), mean gravidity 2.90 (SD±1.94), and mean gestational age 8.78 weeks (SD±3.21). The average length of ED evaluation was 730 min (SD±513). Of the 113 patients, 38 (33.6%) had a return ED visit for nausea and vomiting of pregnancy, 17 (15%) were admitted to the hospital, and 49 (43%) were admitted to the ED observation protocol. Multiple regression analysis showed that the presence of medical co-morbidity (p = 0.039), patient gravditity (p = 0.016), gestational age (p = 0.038), and admission to the hospital (p = 0.004) had small but significant effects on the primary outcome (return visits to the emergency department). No other variables were found to be predictive of return visits to the ED including admission to the ED observation unit or factors classically thought to be associated with severe forms of nausea and vomiting in pregnancy including ketonuria, electrolyte abnormalities, or vital sign abnormalities. Conclusion: Nausea and vomiting in pregnancy has a high rate of return ED visits that can be predicted by young patient age, low patient gravidity, early gestational age, and the presence of other comorbidities. These patients may benefit from obstetric consultation and/or optimization of symptom management after discharge in order to prevent recurrent utilization of the ED. • 229 www.aemj.org S125 Prevalence of Human Trafficking in Adult Sexual Assault Victims David E. Slattery1, Jeff Westin1, Jerri Dermanelian2, Wesley Forred2, Toshia Shaw2, Dale Carrison1 1 University of Nevada School of Medicine, Las Vegas, NV; 2University Medical Center of Southern Nevada, Las Vegas, NV Background: Sex trafficking, the major form of human trafficking (HT), is defined as ‘‘the recruitment, harboring, transportation, provision, or obtaining of a person for the purpose of a commercial sex act’’. It is estimated that 2 million people are trafficked internationally and 25,000 within US borders each year. The majority are women and children. It is important for emergency physicians to recognize the signs of HT; however, there is a paucity of data regarding how commonly these victims present to the ED. Objectives: To determine the prevalence of human trafficking (HT) in adult sexual assault (SA) victims in an urban ED. Methods: IRB-approved, prospective, observational study conducted by sexual assault nurse examiners (SANE) in an urban, academic ED which is the sole provider of care and forensics for adult SA in our community. Inclusion criteria: Adult, SA victims evaluated by the SANE nurses. Exclusion criteria: Victims known to be pregnant, in custody, or psychiatric emergencies. Using convenience sampling, at the end of the examination, four questions were asked. The data were not associated with the medical record and there was no link to any individual. The primary measure was the prevalence of HT defined as a positive response to the question ‘‘have you ever exchanged sex for money, drugs, housing, transportation, clothes, food’’. Secondary measures were the prevalence of ever working in the adult industry and characteristics of involvement. Results: During the 15 month study period, 644 patients were seen by a SANE. Of those patients, 296 were screened for HT. For the primary outcome measure, 73 patients (31%; 95%CI 25,38) met the HT criteria. Of those, 64 (22%) admitted to involvement in the adult industry, and an additional 33 admitted to exchanging sex for material goods or needs. See table. Limitations: Convenience sampling, question results not directly linked to victims’ demographics. Conclusion: There is a high prevalence of HT in adult SA victims. Although our study design and data do not allow us to make any inferences regarding causation, this first report of HT ED prevalence suggests the opportunity to clarify this relationship and the potential opportunity to intervene. Table - Abstract 229: N (%) Adult Industry Involvement Dancer Escort Massage Phone Sex Prostitution Pornography 65 (22) 25 5 1 3 41 2 (8.4) (1.7) (0.34) (1) (13.9) (0.68) N (%) Material Goods for Sex Money Drugs Housing Transportation Clothes Food 89 (30) 68 56 21 17 17 24 (22.9) (18.9) (7.1) (5.7) (5.7) (8.1) S126 230 2012 SAEM ANNUAL MEETING ABSTRACTS Should Empiric Treatment of Gonorrhea and Chlamydia Be Used in the Emergency Department? Scarlet Reichenbach, Leon D. Sanchez, Kathryn A. Volz BIDMC, Boston, MA Background: Sexually transmitted infections (STI) are a significant public health problem. Because of the risks associated with STIs including PID, ectopic pregnancy, and infertility the CDC recommends aggressive treatment with antibiotics in any patient with a suspected STI. Objectives: To determine the rates of positive gonorrhea and chlamydia (G/C) screening and rates of empiric antibiotic use among patients of an urban academic ED with >55,000 visits in Boston, MA. Methods: A retrospective study of all patients who had G/C cultures in the ED over 12 months. Chi-square was used in data analysis. Sensitivity and specificity were also calculated. Results: A positive rate of 9/712 (1.2%) was seen for gonorrhea and 26/714 (3.6%) for chlamydia. Females had positive rates of 2/602 (0.3%) and 17/603 (2.8%) respectively. Males had higher rates of 7/110 (6.4%) (p =< 0.001) and 9/111 (8.1%) (p = 0.006). 284 patients with G/C sent received an alternative diagnosis, the most common being UTI (63), ovarian pathology (35), vaginal bleeding (34), and vaginal candidiasis (33); 4 were excluded. This left 426 without definitive diagnosis. Of these, 24.2% (87/360) of females were treated empirically with antibiotics for G/C, and a greater percentage of males (66%, 45/66) were treated empirically (p < 0.001). Of those empirically treated, 109/132 (82.6%) had negative cultures. Meanwhile 9/32 (28.1%) who ultimately had positive cultures were not treated with antibiotics during their ED stay. Sensitivity of the provider to predict presence of disease based on decision to give empiric antibiotics was 71.9 (CI 53.0–85.6). Specificity was 72.3 (CI 67.6–76.6). Conclusion: Most patients screened in our ED for G/C did not have positive cultures and 82.6% of those treated empirically were found not to have G/C. While early treatment is important to prevent complications, there are risks associated with antibiotic use such as allergic reaction, C difficile infection, and development of antibiotic resistance. Our results suggest that at our institution we may be over-treating for G/C. Furthermore, despite high rates of treatment, 28% of patients who ultimately had positive cultures did not receive antibiotics during their ED stay. Further research into predictive factors or development of a clinical decision rule may be useful to help determine which patients are best treated empirically with antibiotics for presumed G/C. 231 Impact of Airline Travel on Outcome in NHL and NFL Players Immediately Post mTBI: Increased Recovery Times David Milzman1, Jeremy Altman2, Matt Milzman2, Carla Tilchin3, Greg Larkin4, Jordy Sax5 1 Georgetown University School of Medicine, Washington, DC; 2Georgetown University, Washington, DC; 3Bates, Lewis, ME; 4Yale University School of Medicine, New Haven, CT; 5 Johns Hopkins Dept of EM, Baltimore, MD Background: Air travel may be associated with unmeasured neurophysiological changes in an injured brain that may affect post-concussion recovery. No study has compared the effect of commercial airtravel on concussion injuries despite rather obvious decreased oxygen tension and increased dehydration effect on acute mTBI. Objectives: To determine if air travel within 4–6 hours of concussion is associated with increased recovery time in professional football and hockey players. Methods: Prospective cohort study of all active-roster National Football League and National Hockey League players during the 2010–2011 seasons. Internet website review of league sties for injury identification of concussive injury and when player returned to play solely for mTBI. Team schedules and flight times were also confirmed to include only players who flew immediately following game (within 4–6 hr). Multiple injuries were excluded as were players who had injury around all-star break for NHL and scheduled off week in NFL. Results: During the 2010–2011 NFL and NHL seasons, 122 (7.2%) and 101 (13.0%) players experienced a concussion (percent of total players), in the respective leagues. Of these, 68 NFL players (57%) and 39 NHL players (39%) flew within 6 hours of the incident injury. The mean distance flown was shorter for NFL (850 miles, SD 576 vs. NHL 1060, SD 579) miles and all were in a pressurized cabin. The mean number of games missed for NFL and NHL players who traveled by air immediately after concussion was increased by 29% and 24% (respectively) than for those who did not travel by air NFL: 3.8 (SD 2.2) vs. 2.6 games (SD 1.8) and NHL: 16.2 games (SD 22.0) vs.12.4 (SD 18.6); p < 0.03. Conclusion: This is an initial report of an increased rate of recovery in terms of more games missed, for professional athletes flying commercial airlines postmTBI compared to those that do not subject their recently injured brains to pressurized airflight. The obvious changes of decreased oxygen tension with altitude equivalent of 7,500 feet, decreased humidity with increased dehydration, and duress of travel accompanying pressurized airline cabins all likely increase the concussion penumbra in acute mTBI. Early air travel post concussion should be further evaluated and likely postponed 48–72 hr. until initial symptoms subside. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 232 Regional Differences in Emergency Medical Services Use For Patients with Acute Stroke: Findings from the National Hospital Ambulatory Medical Care Survey Emergency Department Data file Prasanthi Govindarajan, Kristin Kuzma, Ralph Gonzales, Judith Maselli, S Claiborne Johnston, Jahan Fahimi, Sharon Poisson, John Stein UCSF Medical Center, San Francisco, CA Background: Previous studies have shown better in-hospital stroke time targets for those who arrive by ambulance compared to other modes of transport. However, regional studies report that less than half of stroke patients arrive by ambulance. Objectives: Our objectives were to describe the proportion of stroke patients who arrive by ambulance nationwide, and to examine regional differences and factors associated with the mode of transport to the emergency department (ED). Methods: This is a cross-sectional study of all patients with a primary discharge diagnosis of stroke based on previously validated ICD-9 codes abstracted from the National Hospital Ambulatory Medical Care Survey for 2007–2009. We excluded subjects <18 years of age and those with missing data. The study related survey variables included patient demographics, community characteristics, mode of transport to the hospital, and hospital characteristics. Results: 566 patients met inclusion criteria, representing 2,153,234 patient records nationally. Of these, 50.4% arrived by ambulance. After adjustment for potential confounders, patients residing in the west and south had lower odds of arriving by ambulance for stroke when compared to northeast (Southern Region, OR 0.45, 95% CI 0.26–0.76, Western Region, OR 0.45, 95% CI 0.25–0.84, Midwest Region, OR 0.56, 95% CI 0.31– 1.01). Compared to the Medicare population, privately insured and self insured had lower odds of arriving by ambulance (OR for private insurance 0.48, 95% CI 0.28– 0.84 and OR for self payers 0.36, 95% CI 0.14–0.93). Age, sex, race, urban or rural location of ED, or safety net status were not independently associated with ambulance use. Conclusion: Patients with stroke arrive by ambulance more frequently in the northeast than in other regions of the US. Identifying reasons for this regional difference may be useful in improving ambulance utilization and overall stroke care nationwide. 233 Effect of Race and Ethnicity on the Presentation of Stroke Among Adults Presenting to the Emergency Department Bradley Li, Sayed Hussain, Hani Judeh, Kruti Joshi, Michael S. Radeos New York Hospital Queens, Bayside, NY Background: Stroke is recognized as a time-urgent disease that requires prompt diagnosis and treatment. Management differs depending on whether the stroke is hemorrhagic or ischemic. • www.aemj.org S127 Objectives: We sought to determine whether there was a difference in type of stroke presentation based upon race. We further sought to determine whether there is an increase in hemorrhagic strokes among Asian patients with limited English proficiency. Methods: We performed a retrospective chart review of all stroke patients age 18 and older for 1 year of patients that were diagnosed with cerebral vascular accident (CVA) or intracranial hemorrhage (ICH). We collected data on patient demographics, and past medical history. We then stratified patients according to race (white, black, Latino, Asian, and other). We classified strokes as ischemic, intracranial hemorrhage (ICH), subarachnoid hemorrhage (SAH), subdural hemorrhage (SDH), and other (e.g., bleeding into metatstatic lesions). We used only the index visit. We present the data percentages, medians and interquartile ranges (IQR). We tested the association of the outcome of intracranial hemorrhage against demographic and clinical variables using chi-square and Kruskal-Wallis tests. We performed a logistic regression model to determine factors related to presentation with an intracranial hemorrhage (ICH). Results: A total of 457 patients presented between 7/1/ 09 and 6/30/10. The median age was 74 years (IQR 60 to 83). 251 (55%) were female; 194 (42%) were white, 58 (12%) black, 76 (17%) Latino, 111 (24%) Asian/Pacific Islander (of these 14% were Chinese and 6% were Korean, 3% Indian, and 2% other Asian/Pacific Islander), 2% missing, and 2% other race. 94 patients (21%) had a primary language other than English. Of all strokes, 353 (77%) were ischemic, 69 (15%) were ICH, 17 (4%) were SAH, 14 (3%) were SDH, and 4 (1%) were other. There was no association between the presentation of ICH and race, either in the univariate analysis (OR 1.31, 95% CI 0.80 to 2.15) or in a model adjusted for age, sex, non-English speaking status and comorbidities (OR 1.29, 95% CI 0.49 to 3.40). Conclusion: Asian patients, when compared to nonAsian patients, had no detectable difference in the rate of ICH on ED presentation. Further research in this area should continue to focus on primary prevention equally to all races while searching for risk factors that may be present in certain races. 234 Non-febrile Seizures In The Pediatric Emergency Department: To Draw Labs And CT Scan Or Not? Vikramjit S. Gill, Ashley Strobel University of Maryland, Baltimore, MD Background: The practice of obtaining laboratory studies and routine CT scan of the brain on every child with a seizure has been called into question in the patient who is alert, interactive, and back to functional baseline. There is still no standard practice for the management of non-febrile seizure patients in the pediatric emergency department (PED). Objectives: We sought to determine the proportion of patients in whom clinically significant laboratory studies and CT scans of the brain were obtained in children who presented to the PED with a first or recurrent S128 2012 SAEM ANNUAL MEETING ABSTRACTS non-febrile seizure. We hypothesize that the majority of these children do not have clinically significant laboratory or imaging studies. If clinically significant values were found, the history given would warrant further laboratory and imaging assessment despite seizure alone. Methods: We performed a retrospective chart review of 93 patients with first-time or recurrent non-febrile seizures at an urban, academic PED between July 2007 to June 2011. Exclusion criteria included children who presented to the PED with a fever and age less than 2 months. We looked at specific values that included a complete blood count, basic metabolic panel, and liver function tests, and if the child was on antiepileptics along with a level for a known seizure disorder, and CT scan. Abnormal laboratory and CT scan findings were classified as clinically significant or not. Results: The median age of our study population is 4 years with male to female ratio of 1.7. 70% of patients had a generalized tonic-clonic seizure. Laboratory studies and CT scans were obtained in 87% and 35% of patients, respectively. Five patients had clinically significant abnormal labs; however, one had ESRD, one developed urosepsis, one had eclampsia, and two others had hyponatremia, which was secondary to diluted formula and trileptal toxicity. Three children had an abnormal head CT: two had a VP shunt and one had a chromosomal abnormality with developmental delay. Conclusion: The majority of the children analyzed did not have clinically significant laboratory or imaging studies in the setting of a first or recurrent non-febrile seizure. Of those with clinically significant results, the patient’s history suggested a possible etiology for their seizure presentation and further workup was indicated. stroke-like symptoms fall 2006–summer 2010. The product-moment correlation coefficient was used to estimate the strength of correlation between providers’ NIHSS scores and decisions to give or withhold rtPA. A BlandAltman analysis was used to determine level of agreement (LOA) between NIHSS scores, and frequency analyses were performed on emergency physician’s rationale for withholding rtPA. Results: Correlation coefficients of NIHSS scores between providers were 0.914 for 2006–2007, 0.869 for 2010, and 0.907 for both time periods combined (P < 0.0001). Bland-Altman analysis demonstrated a slight, but statistically insignificant, increase in the difference in emergency department scores over stroke team scores as the average NIHSS scores increased. The overall point estimate for LOA for NIHSS scores showed strong agreement (1.096, 95%CI )4.87–6.77), but 95% limits of the LOA for NIHSS scores were ±5.82. For the decision to treat the correlation coefficients were 0.282 for 2006–2007, 0.204 for 2010, and 0.250 for both time periods combined, and kappa coefficients were 0.245 for 2006–2007, 0.186 for 2010, and 0.226 for all years. Frequency analysis showed the most frequent reason (29%) for EM physicians’ decision to withhold rtPA was the onset of patient’s symptoms >3 hours prior to the decision to give rtPA. Conclusion: EM and neurology decisions to give or withhold rtPA cannot serve as substitutes for each other. Despite very similar initial neurologic assessments, EM physicians and the stroke team may weigh the risks of harm and benefit differently in making the decision to administer thrombolytics. 236 235 Thrombolytics in Acute Ischemic Stroke Assessment and Decision-Making: EM vs. Neurology Margaret Vo1, Yashwant Chathampally2, Raja Malkani3, David Robinson2, Emily Pavlik4 1 Paul L. Foster School of Medicine, El Paso, TX; 2University of Texas Health Science Center at Houston, Houston, TX; 3University of Texas School of Public Health - Austin Regional Campus, Austin, TX; 4Emergency Physicians Affiliate, San Antonio, TX Background: The amount of time elapsed after an ischemic stroke may increase the extent of irreversible neuronal loss, making avoiding delays in reperfusion pertinent to reducing mortality and morbidity, and improving quality of life. Objectives: The purpose of this study was to compare the emergency physician’s and stroke team’s assessments and independent decisions to administer or withhold thrombolytic therapy in patients presenting to the emergency department (ED) with stroke-like symptoms. Methods: Stroke assessment and decisions performed by EM and neurology providers at a tertiary care, university hospital were collected on a convenience sample of the emergency physician and stroke team neurologists for 77 patients presenting to the ED with Lumbar Puncture or CT Angiography Following a Negative CT Scan for Suspected Subarachnoid Hemorrhage: A Decision Analysis? Foster R. Goss, John B. Wong Tufts Medical Center, Boston, MA Background: In patients with a negative CT scan for suspected subarachnoid hemorrhage (SAH), CT angiography (CTA) has emerged as a controversial alternative diagnostic strategy in place of lumbar puncture (LP). Objectives: To determine the diagnostic accuracy for SAH and aneurysm of LP alone, CTA alone, and LP followed by CTA if the LP is positive. Methods: We developed a decision and Bayesian analysis to evaluate 1) LP, 2) CTA, and 3) LP followed by CTA if the LP is positive. Data were obtained from the literature. The model considers probability of SAH (15%), aneurysm (85% if SAH), sensitivity and specificity of CT (92.9% and 100% overall), of LP (based on RBC and xanthochromia), and of CTA, traumatic tap and its influence on SAH detection. Analyses considered all patients and those presenting at less than 6 hours or greater than 6 hours from symptom onset by varying the sensitivity and specificity of CT and CTA. Results: Using the reported ranges of CT scan sensitivity and the specificity, the revised likelihood of SAH ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 following a negative CT ranged from 0.5–3.7%, and the likelihood of aneurysm ranged from 2.3–5.4%. Following any of the diagnostic strategies, the likelihood of missing SAH ranged from 0–0.7%. Either LP strategy diagnosed 99.8% of SAHs versus 83–84% with CTA alone because CTA only detected SAH in the presence of an aneurysm. False positive SAH with LP ranged from 8.5–8.8% due to traumatic taps and with CTA ranged from 0.2–6.0% due to aneurysms without SAH. The positive predictive value for SAH ranged from 5.7–30% with LP and from 7.9–63% with CTA. For patients presenting within 6 hours of symptom onset, the revised likelihood of SAH following a negative CT became 0.53%, and the likelihood of aneurysm ranged from 2.3–2.7%. Following any of the diagnostic strategies, the likelihood of missing SAH ranged from 0.01–0.095%. Either LP strategy diagnosed 99.8% of SAH versus 83–84% with CTA alone. False positive SAH with LP was 8.8% and with CTA ranged from 0.2–5.1%. The positive predictive value for SAH was 5.7% with LP and from 7.9–63% with CTA. CTA following a positive LP diagnosed 8.5–24% of aneurysms. Conclusion: LP strategies are more sensitive for detecting SAH but less specific than CTA because of traumatic taps, leading to lower predictive value positives for SAH with LP than with CTA. Either diagnostic strategy results in a low likelihood of missing SAH, particularly within 6 hours of symptom onset. 237 Abnormalities on CT Perfusion in Patients Presenting with Transient Ischemic Attack (TIA) Sharon N. Poisson1, Jane J. Kim2, Prasanthi Govindarajan1, S. Claiborne Johnston1, Mai N. Nguyen-Huynh3 1 University of California San Francisco, San Francisco, CA; 2Kaiser Permanente Medical Care Plan, San Francisco, CA; 3Kaiser Permanente Division of Research, Oakland, CA Background: Recent studies support perfusion imaging as a prognostic tool in ischemic stroke, but little data exist regarding its utility in transient ischemic attack (TIA). CT perfusion (CTP), which is more available and less costly to perform than MRI, has not been well studied. Objectives: To characterize CTP findings in TIA patients, and identify imaging predictors of outcome. Methods: This retrospective cohort study evaluated TIA patients at a single ED over 15 months, who had CTP at initial evaluation. A neurologist blinded to CTP findings collected demographic and clinical data. CTP images were analyzed by a neuroradiologist blinded to clinical information. CTP maps were described as qualitatively normal, increased, or decreased in mean transit time (MTT), cerebral blood volume (CBV), and cerebral blood flow (CBF). Quantitative analysis involved measurements of average MTT (seconds), CBV (cc/100 g) and CBF (cc/[100g x min]) in standardized regions of interest within each vascular distribution. These were compared with values in the other hemisphere for • www.aemj.org S129 relative measures of MTT difference, CBV ratio, and CBFfratio. MTT difference of ‡2 seconds, rCBV as £0.60, and rCBF as £0.48 were defined as abnormal based on prior studies. Clinical outcomes including stroke, TIA, or hospitalization during follow-up were determined up to one year following the index event. Dichotomous variables were compared using Fisher’s exact test. Logistic regression was used to evaluate the association of CTP abnormalities with outcome in TIA patients. Results: Of 99 patients with validated TIA, 53 had CTP done. Mean age was 72 ± 12 years, 55% were women, and 64% were Caucasian. Mean ABCD2 score was 4.7 ± 2.1, and 69% had an ABCD2 ‡ 4. Prolonged MTT was the most common abnormality (19, 36%), and 5 (9.4%) had decreased CBV in the same distribution. On quantitative analysis, 23 (43%) had a significant abnormality. Four patients (7.5%) had prolonged MTT and decreased CBV in the same territory, while 17 (32%) had mismatched abnormalities. When tested in a multivariate model, no significant associations between mismatch abnormalities on CTP and new stroke, TIA, or hospitalizations were observed. Conclusion: CTP abnormalities are common in TIA patients. Although no association between these abnormalities and clinical outcomes was observed in this small study, this needs to be studied further. 238 Withdrawn 239 Economic Benefit of an Educational Intervention to Improve tPA Use as Treatment For Acute Ischemic Stroke in Community Hospitals: Secondary Analysis of the INSTINCT Trial Cemal B. Sozener, David W. Hutton, William J. Meurer, Shirley M. Frederiksen, Allison M. Kade, Phillip A. Scott University of Michigan, Ann Arbor, MI Background: Prior work demonstrates substantial economic benefit from tPA use in acute ischemic stroke (AIS). Objectives: We hypothesized a T2 knowledge translation (KT) program to increase community tPA treatment in AIS would be cost-effective beyond research funds spent. Methods: Data were utilized from the INcreasing Stroke Treatment through INterventional behavior Change Tactics (INSTINCT) trial, a prospective, cluster randomized, controlled trial involving 24 community hospitals in matched pairs. Within pairs, hospitals were randomly assigned to receive a barrier assessmentinteractive educational intervention (BA-IEI) vs. control. Cost analyses were conducted from a societal perspective for two cases: 1) using total trial costs, and 2) using intervention costs alone (no research overhead) as an estimate of the cost of generalization of the results. Trial costs are defined as total INSTINCT funding combined with opportunity costs of health professionals S130 2012 SAEM ANNUAL MEETING ABSTRACTS attending study events. Savings attributable to increased tPA use were determined by applying published stroke economic data, adjusted for inflation, to the study cohorts. These data were integrated in a Markov model to determine the long-term economic effect of the INSTINCT BA-IEI versus control. Results: The INSTINCT trial cost (US)$3.3 million. Intervention sites treated 2.30% (244/10,627) of patients with tPA compared to 1.59% (160/10,071) at control sites (per protocol analysis). This increase in tPA use resulted in direct savings of approximately $540,000 due to reduced length of hospital and nursing facility stay. Increased tPA usage resulted in an estimated additional 81 quality adjusted life years (QALY), with an incremental costeffectiveness ratio of $34,000/QALY. Using $50,000 as a conservative estimate of societal value per QALY, an additional benefit of $4,100,000, or net societal economic benefit of $1.3 million, was realized. Generalizing the intervention in a similar population (excluding research overhead) would cost an estimated $680,000 and provide a net benefit of $3.9 million, assuming similar effectiveness and treatment outcomes. Conclusion: Due to the underlying cost-effectiveness of tPA, community KT efforts with modest absolute gains in tPA usage produce substantial societal economic returns and are considered good value when compared to spending on other health interventions. 240 Anti-hypertensive Treatment Prolongs tPA Door-to-treatment Time: Secondary Analysis Of The Increasing Stroke Treatment Through Interventional Behavior Change Tactics (INSTINCT) Trial Lesli E. Skolarus, Phillip A. Scott, James F. Burke, Eric E. Adelman, Shirley M. Frederiksen, Allison M. Kade, Jack D. Kalbfleisch, William J. Meurer University of Michigan, Ann Arbor, MI Background: Increased time to tPA treatment is associated with worse outcomes. Thus, identifying modifiable treatment delays may improve stroke outcomes. Objectives: We hypothesized that pre-thrombolytic anti-hypertensive treatment (AHT) may prolong door to treatment time (DTT). Methods: Secondary data analysis of consecutive tPAtreated patients at 24 randomly selected Michigan community hospitals in the INSTINCT trial. DTT among stroke patients who received pre-thrombolytic AHT were compared to those who did not receive pre-thrombolytic AHT. We then calculated a propensity score for the probability of receiving pre-thrombolytic AHT using a logistic regression model with covariates including demographics, stroke risk factors, antiplatelet or beta blocker as home medication, stroke severity (NIHSS), onset to door time, admission glucose, pretreatment systolic and diastolic blood pressure, EMS usage, and location at time of stroke. A paired t-test was then performed to compare the DTT between the propensity-matched groups. A separate generalized estimating equations (GEE) approach was also used to estimate the differences between patients receiving pre-thrombolytic AHT and those who did not while accounting for within-hospital clustering. Results: A total of 557 patients were included in INSTINCT; however, onset, arrival, or treatment times were not able to be determined in 23, leaving 534 patients for this analysis. The unmatched cohort consisted of 95 stroke patients who received pre-thrombolytic AHT and 439 stroke patients who did not receive AHT from 2007–2010 (table). In the unmatched cohort, patients who received pre-thrombolytic AHT had a longer DTT (mean increase 9 minutes; 95% confidence interval (CI) 2–16 minutes) than patients who did not receive pre-thrombolytic AHT. After propensity matching (table), patients who received pre-thrombolytic AHT had a longer DTT (mean increase 10.4 minutes, 95% CI 1.9–18.8) than patients who did not receive pre-thrombolytic AHT. This effect persisted and its magnitude was not altered by accounting for clustering within hospitals. Conclusion: Pre-thrombolytic AHT is associated with modest delays in DTT. This represents a feasible target for physician educational interventions and quality improvement initiatives. Further research evaluating optimum hypertension management pre-thrombolytic treatment is warranted. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 241 Protocol Deviations During and After IV Thrombolysis in Community Hospitals Eric E. Adelman, William Meurer, Lesli E. Skolarus, Allison M. Kade, Shirley M. Frederiksen, Jack D. Kalbfleisch, Phillip A. Scott University of Michigan, Ann Arbor, MI Background: Protocol deviations (PDs) before and immediately after IV thrombolysis for acute ischemic stroke are common. Patient and hospital factors associated with PDs are not well described. Objectives: We aimed to determine which patient or hospital factors were associated with pre- and posttreatment PDs in a cohort of community-treated thrombolysis patients. Methods: The INSTINCT (Increasing Stroke Treatment through Interventional Behavior Change Tactics) study was a multicenter, cluster-randomized controlled trial in 24 Michigan community hospitals evaluating the efficacy of a barrier assessment and educational intervention to increase appropriate tPA use. PDs were defined based on 2007 AHA guidelines with the addition of the 3–4.5 hour treatment window, for which the ECASS III criteria were applied. PDs were categorized as pretreatment (Pre-PDs), post-treatment (Post-PDs), or both. Multi-level logistic regression models were fitted to determine whether patient and hospital variables were associated with Pre-PDs or Post-PDs. The models included all variables specified a priori to be potentially clinically relevant; Pre-PD was included as a covariate in the model for Post-PD. Results: During the study, 557 patients (mean age 70; 52% male; median NIHSS 12) were treated with IV tPA. PDs occurred in 233 (42%) patients: 26% had only Table - Abstract 241: • www.aemj.org S131 Post-PDs, 7% had only Pre-PDs, and 9% had both. The most common PDs included failure to treat post-treatment hypertension (131, 24%), antiplatelet agent within 24 hours of treatment (61, 11%), pre-treatment blood pressure over 185/110 (39, 7%), anticoagulant agent within 24 hours of treatment (31, 6%), and treatment outside the time window (29, 5%). Symptomatic intracranial hemorrhage (SICH) was observed in 7.3% of patients with PDs and 6.5% of patients without any PD. In-hospital case fatality was 12% with and 10% without a PD. In the fully adjusted model, older age was significantly associated with Pre-PDs (Table). When Post-PDs were evaluated with adjustment for Pre-PDs, age was not associated with PDs; however, Pre-PDs were associated with Post-PDs. Conclusion: Older age was associated with increased odds of Pre-PDs in Michigan community hospitals. PrePDs were associated with Post-PDs. SICH and in-hospital case fatality were not associated with PDs; however, the low number of such events limited our ability to detect a difference. 242 CT Is Sensitive For The Detection Of Advanced Leukoaraiosis In Patients With Transient Ischemic Attack Matthew S. Siket1, Ross C. Avery1, Eitan Auriel1, Johanna A. Helenius1, Gyeong Moon Kim1, J. Alfredo Caceres1, Octavio Pontes-Neto1, Hakan Ay2 1 Massachusetts General Hospital, Boston, MA; 2 Massachusetts General Hospital, AA Martinos Center for Biomedical Imaging, Harvard Medical School, Boston, MA Background: MRI has become the gold standard for the detection of cerebral ischemia and is a component of multiple imaging enhanced clinical risk prediction rules for the short-term risk of stroke in patients with transient ischemic attack (TIA). However, it is not always available in the emergency department (ED) and is often contraindicated. Leukoaraiosis (LA) is a radiographic term for white matter ischemic changes, and has recently been shown to be independently predictive of disabling stroke. Although it is easily detected by both CT and MRI, their comparative ability is unknown. Objectives: We sought to determine whether leukoaraiosis, when combined with evidence of acute or old infarction as detected by CT, achieved similar sensitivity to MRI in patients presenting to the ED with TIA. Methods: We conducted a retrospective review of consecutive patients diagnosed with TIA between June 2009 and July 2011 that underwent both CT and MRI as part of routine care within 1 calendar day of presentation to a single, academic ED. CT and MR images were reviewed by a single emergency physician who was blinded to the MR images at the time of CT interpretation. LA was graded using the Van Sweiten scale (VSS), a validated grading scale applicable to both CT and MRI. Anterior and posterior regions were graded independently from 0 to 2. Results: 361 patients were diagnosed with TIA during the study period. Of these, 194 had both CT and MRI S132 2012 SAEM ANNUAL MEETING ABSTRACTS performed within 1 day of presentation. Images from both modalities were available for review in 172. Abnormalities defined as acute infarction, old infarction, or LA were present in 133 (77.3%) by CT compared to 161 (93.6%) by MRI. LA was detected in 115 (66.9%) by CT compared to 145 (84.3%) by MRI (P < 0.0002). CT achieved sensitivity and specificity of 75.9% and 81.5% respectively in the overall detection of LA. Positive and negative predictive values were 95.6% and 38.6% respectively. Advanced LA (VSS 2) was similar in both modalities, with 68 (39.5%) by CT and 63 (36.6%) by MRI (p = 0.577). Conclusion: MRI is significantly more sensitive than CT in detecting cerebral ischemia in patients with TIA. However, CT is sufficiently sensitive in detecting advanced LA, which may serve to increase CT-based stroke risk prediction strategies of TIA patients in the ED. (median 23 vs 21, p = 0.042) but similar CS (39 vs 40, p = 0.16) and TE (19 vs 19, p = 0.38) compared to those working mostly day shifts. Residents with children had similar levels of BO (21 vs 22, p = 0.47) and CS (41 vs 40, p = 0.26), but higher TE (21 vs 19, p = 0.007) than those without children. Emergency medicine residents had a nonsignificant trend towards higher TE, but similar BO and CS compared to those in primarily surgical or medical specialties. Conclusion: Compassion satisfaction was similar in all groups. Burnout was associated with working more hours per week and working predominantly nights. Traumatic experiences appear to be more pronounced in those working in the emergency medicine specialty, those working more hours per week, and those with children. 244 243 Compassion Fatigue: Emotional Exhaustion After Burnout M. Fernanda Bellolio, Daniel Cabrera, Annie T. Sadosty, Erik P. Hess, Kharmene L. Sunga Mayo Clinic, Rochester, MN Background: Helping others is often a rewarding experience but can also come with a ‘‘cost of caring’’ also known as compassion fatigue (CF). CF can be defined as the emotional and physical toll suffered by those helping others in distress. It is affected by three major components: compassion satisfaction (CS), burnout (BO), and traumatic experiences (TE). Previous literature has recognized an increase in BO related to work hours and stress among resident physicians. Objectives: To assess the state of CF among residents with regard to differences in specialty training, hours worked, number of overnights, and demands of child care. We aim to measure associations with the three components of CF (CS, BO, and TE). Methods: We used the previously validated survey, ProQOL 5. The survey was sent to the residents after approval from the IRB and the program directors. Results: A total of 193 responses were received (40% of the 478 surveyed). Five were excluded due to incomplete questionnaires. We found that residents who worked more hours per week had significantly higher BO levels (median 25 vs 21, p = 0.038) and higher TE (22 vs 19, p = 0.048) than those working less hours. There was no difference in CS (42 vs 40, p = 0.73). Eighteen percent of the residents worked a majority of the night shifts. These residents had higher levels of BO Effective Methods To Improve Emergency Department Documentation In A Teaching Hospital To Enhance Education, Billing, And Medical Liability Anurag Gupta, Joseph Habboushe Beth Israel Medical Center, New York, NY Background: Emergency department (ED) billing includes both facility and professional fees. An algorithm derived from the medical provider’s chart generates the latter fee. Many private hospitals encourage appropriate documentation by financially incentivizing providers. Academic hospitals sometimes lag in this initiative, possibly resulting in less than optimal charting. Past attempts to teach proper documentation using our electronic medical record (EMR) were difficult in our urban, academic ED of 80 providers (approximately 25 attending physicians, 36 residents, and 20 physician assistants). Objectives: We created a tutorial to teach documentation of ED charts, modified the EMR to encourage appropriate documentation, and provided feedback from the coding department. This was combined with an incentive structure shared equally amongst all attendings based on increased collections. We hypothesized this instructional intervention would lead to more appropriate billing, improve chart content, decrease medical liability, and increase educational value of charting process. Methods: Documentation recommendations, divided into two-month phases of 2–3 proposals, were administered to all ED providers by e-mails, lectures, and reminders during sign-out rounds. Charts were reviewed by coders who provided individual feedback ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S133 if specific phase recommendations were not followed. Our endpoints included change in total RVU, RVUs/ patient, E/M level distribution, and subjective quality of chart improvement. We did not examine effects on procedure codes or facility fees. Results: Our base average RVU/patient in our ED from 1/1/11–6/30/11 was 2.615 with monthly variability of approximately 2%. Implementation of phase one increased average RVU/patient within two weeks to 2.73 (4.4% increase from baseline, p < 0.05). The second aggregate phase implemented 8 weeks later increased average RVU/patient to 3.04 (16.4% increase from baseline, p < 0.05). Conclusion: Using our teaching methods, chart reviews focused on 2–3 recommendations at a time, and EMR adjustments, we were able to better reflect the complexity of care that we deliver every day in our medical charts. Future phases will focus on appropriate documentation for procedures, critical care, fast track, and pediatric patients, as well as examining correlations between increase in RVUs with charge capture. mentoring evaluation, create a mechanism to identify and reward mentoring. National practice examples offered critical recommendations to address multi-generational attitudes and faculty diversity in terms of gender, race, and culture. Conclusion: Mentoring strategies can be identified to serve a diverse faculty in academic medicine. Interventions to improve mentoring practices should be targeted at the level of the institution, department, and individual faculty members. It is imperative to adopt results such as these to design effective mentoring programs to enhance the success of emergency medicine faculty seeking robust academic careers. Identifying Mentoring ‘‘Best Practices’’ for Medical School Faculty Julie L. Welch, Teresita Bellido, Cherri D. Hobgood Indiana University, Indianapolis, IN Background: For graduating residents applying for academic emergency medicine positions, there is a paucity of information on what attributes emergency department (ED) chairs are seeking. Objectives: To determine which qualities academic ED chairs are looking for when hiring a new physician directly out of residency or fellowship. Methods: An anonymous 15-item web-based survey was sent to the department chairs of all accredited civilian emergency medicine residency programs in March and April of 2011. Respondents were asked to rate different parts of the candidate’s application using a fivepoint Likert scale, rank important attributes, and list any desirable fellowships. They were also asked to give the current number of available job openings. Results: 84 of 152 eligible chairs responded, giving a response rate of 55%. The most important parts of a candidate’s application were the interview (4.79 ± 0.41), another employee’s recommendation (4.70 ± 0.52), and the program director’s recommendation (4.54 ± 0.67). Less weight was given to the reputation of the residency program attended (3.80 ± 0.71) as well as attending a 3- or 4-year program (2.78 ± 1.46). The single most important attribute possessed by a candidate was identified as ‘‘Ability to work in a team.’’ Other common responses included ‘‘Clinical productivity,’’ ‘‘Teaching potential,’’ and ‘‘Substantial publications.’’ Advanced training in ultrasound was listed as the most sought after fellowship by 55% of the chairs. Other fellowships receiving >30% of affirmative votes include critical care, pediatrics, and toxicology. Overall, department chairs did not have a difficult time in recruiting EM-trained physicians (2.5 ± 1.3 on a five-point scale), with 56% of respondents stating that they had no current job openings. Conclusion: This study is the first attempt to examine what academic ED chairs are looking for when they hire new attendings. How a physician relates to others was consistently rated and ranked as the most important part of the candidate’s application with the interview and recommendations taking the lead. As they are 245 Background: Mentoring has been identified as an essential component for career success and satisfaction in academic medicine. Many institutions and departments struggle with providing both basic and transformative mentoring for their faculty. Objectives: We sought to identify and understand the essential practices of successful mentoring programs. Methods: Multidisciplinary institutional stakeholders in the school of medicine including tenured professors, deans, and faculty acknowledged as successful mentors were identified and participated in focused interviews between Mar-Nov 2011. The major area of inquiry involved their experiences with mentoring relationships, practices, and structure within the school, department, or division. Focused interview data were transcribed and grounded theory analysis was performed. Additional data collected by a 2009 institutional mentoring taskforce were examined. Key elements and themes were identified and organized for final review. Results: Results identified the mentoring practices for three categories: 1) General themes for all faculty, 2) Specific practices for faculty groups: Basic Science Researchers, Clinician Researchers, Clinician Educators, and 3) National examples. Additional mentoring strategies that failed were identified. The general themes were quite universal among faculty groups. These included: clarify the best type of mentoring for the mentee, allow the mentee to choose the mentor, establish a panel of mentors with complementary skills, schedule regular meetings, establish a clear mentoring plan with expectations and goals, offer training and resources for both the mentor and mentee at institutional and departmental levels, ensure ongoing 246 A Pilot Study to Survey Academic Emergency Medicine Department Chairs on Hiring New Attendings Ryan D. Aycock, Moshe Weizberg, Brahim Ardolic Staten Island University Hospital, Staten Island, NY S134 2012 SAEM ANNUAL MEETING ABSTRACTS tasked with training the next generation of physicians, academic programs are also concerned with teaching and research ability. Overall, finding a job in academic emergency medicine is difficult, with graduates having limited job prospects. 247 Reliability of the Revised Professional Practice Environment Scale when Used with Emergency Physicians Tania D. Strout, Michael R. Baumann, Julie E.O. Pelletier, Katie W.D. Dolbec Maine Medical Center, Portland, ME Background: The Revised Professional Practice Environment Scale (RPPE) is a 39-item multidimensional measure designed to evaluate eight components of professional clinical practice in the acute care setting. Developed for and used primarily with registered nurses, the RPPE is a valuable tool for evaluating clinicians’ perceptions of the practice environment, the efficacy of changes in practice, for strategic planning, and for developing supports for the IOM’s six dimensions of quality. Objectives: We sought to evaluate the reliability and validity of the RPPE when used with a sample of emergency physicians. Methods: A psychometric evaluation of the RPPE was undertaken with a sample of emergency physicians. Participants completed the 39-item instrument in an anonymous fashion using a pencil and paper format. Analyses included evaluations of: a) internal consistency reliability using Cronbach’s alpha coefficient and item analysis, b) principle components analysis (PCA) with Varimax rotation and Kaiser normalization, and c) internal consistency reliability of the components identified through PCA. Results: The initial Cronbach’s alpha for all 39 items was 0.88; 11 items were then removed from analysis due to low (<0.30) corrected item-to-total correlations. Cronbach’s alpha for the remaining 28 items was 0.905, indicating high internal consistency reliability. The 28 items subjected to PCA yielded a six-component solution explaining 76.1% of observed variance. Cronbach’s alpha coefficient for the individual components comprising the scale ranged from 0.70 to 0.89, indicating that the subscales may be used independently to measure the various aspects of the emergency physician’s professional practice environment. Conclusion: The results of this pilot study suggest that a shortened, 28-item version of the RPPE scale is psychometrically sound when used with emergency physicians to evaluate their perceptions regarding their practice environment. While the RPPE holds promise as a method for evaluating changes to the practice environment and for creating an environment supportive of the IOM’s dimensions of quality, additional research with a larger, more diverse sample of physicians is indicated prior to widespread utilization of the instrument in the physician population. 248 Gender Diversity in Emergency Medicine: Measuring System’s Change Lori Post, Nupur Garg, Gail D’Onofrio Yale University School of Medicine, New Haven, CT Background: Women comprise half of the talent pool from which the specialty of emergency medicine draws future leaders, researchers, and educators and yet only 5% of full professors in US emergency medicine are female. Both research and interventions are aimed at reducing the gender gap, however, it will take decades for the benefits to be realized which creates a methodological challenge in assessing system’s change. Current techniques to measure disparities are insensitive to systems change as they are limited to percentages and trends over time. Objectives: To determine if the use of Relative Rate Index (RRI) better predicts which stage in the system women are not advancing in the academic pipeline than traditional metrics. Methods: RRI is a method of analysis that assesses the percent of sub-populations in each stage relative to their representation in the stage directly prior. Thus, there is a better notion of the advancement given the availability to advance. RRI also standardizes data for ease of interpretation. This study was conducted on the total population of academic professors in all departments at Yale School of Medicine during the academic year of 2010–2011. Data were obtained from the Yale University Provost’s office. Results: N = 1305. There were a total of 402 full, 429 associate, and 484 assistant professors. Males comprised 78%, 59%, and 54% respectively. RRI for the Department of Emergency Medicine (DEM) is 0.67, 1.93, and 0.78, for Full, Associate, and Assistant Professors, respectively while the percentages were 44%, 60%, and 33% respectively. Conclusion: Relying solely on percentages masks improvements to the system. Women are most represented at the associate professor level in DEM, highlighting the importance of systems change evidence. Specifically, twice as many women are promoted to associate professor rank given the number who exists as assistant professors. Within 5 years, the DEM should have an equal system as the numbers of associate professors have dramatically increased and will be eligible to promote to full professor. Additionally, DEM has a better record of retaining and promoting women than other Yale Departments of Medicine at both associate and full professor ranks. Table 1 - Abstract 248: Interpretation of Relative Rate Index (RRI) RRI 1 <1 >1 Interpretation there is an equal rate of progression of females and males from one stage to the next there are fewer females progressing relative to males given their representation in the step prior there are more females progressing relative to males given their representation in the step prior ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 249 Quantifying the Safety Net Role of the Academic Emergency Department Benjamin S. Heavrin, Cathy A. Jenkins Vanderbilt University, Nashville, TN Background: EDs have an important role in providing care to safety net populations. The safety net role for the academic ED has not been quantified at a national level. Objectives: We examine the payer mixes of community non-rehabilitation EDs in metropolitan areas by region to identify the proportion of academic and nonacademic EDs that could be considered safety net EDs. We hypothesize that the proportion of safety net academic • www.aemj.org S135 EDs is greater than that for non-academic EDs and is increasing over time. Methods: This is an ecological study examining US ED visits from 2006 through 2008. Data were obtained from the Nationwide Emergency Department Sample (NEDS). We grouped each ED visit according to the unique hospital-based ED identifier, thus creating a payer mix for each ED. We define a ‘‘Safety Net ED’’ as any ED where the payer mix satisfied any one of the following three conditions: 1) >30% of all ED visits are Medicaid patients; 2) >30% of all ED visits are self-pay patients; or 3) >40% of all ED visits are either Medicaid or self-pay patients. NEDS tags each ED with a hospital-based variable to delineate metropolitan/non-metropolitan locations and academic affiliation. We chose to examine a subpopulation of EDs tagged as either academic metropolitan or non-academic metropolitan, because the teaching status of non-metropolitan hospitals was not provided. We then measured the proportion of EDs that met safety net criteria by academic status and region. Results: We examined 2,821, 2,793, and 2,844 weighted metro EDs in years 2006–2008, respectively. Table 1 presents safety net proportions. The proportions of academic safety net EDs increased across the study period. Widespread regional variability in safety net proportions existed across all years. The proportions of safety net EDs were highest in the South and lowest in the Northeast and Midwest. Table 2 describes these findings for 2008. Conclusion: These data suggest that the proportion of safety-net academic EDs may be greater than that of non-academic EDs, is increasing over time, and is Table 1 - Abstract 249: Hospital Frequencies and Safety Net Proportions by Academic Status Academic Metropolitan EDs Year 2006 2007 2008 Safety Net Academic Metropolitan EDs Frequency Weighted Frequency Frequency Weighted Frequency Academic Safety Net % 95% CI 181 167 172 886.0 805.0 815.0 77 83 85 376.1 416.9 409.4 42.45 51.79 50.23 (36.13, 48.77) (45.06, 58.52) (43.35, 57.11) Non-Academic Metropolitan EDs 2006 2007 2008 Safety Net Non-Academic Metropolitan EDs Frequency Weighted Frequency Frequency Weighted Frequency Non-Academic Safety Net % 95% CI 392 404 411 1935.0 1,988.0 2,029.0 166 167 171 825.7 832.4 842.8 42.67 41.87 41.54 (38.05, 47.30) (37.40, 46.34) (37.23, 45.85) Table 2 - Abstract 249: Safety Net Proportions by Region, 2008 Academic Metropolitan EDs Region Northeast Midwest South West Safety Net Academic Metropolitan EDs Weighted Frequency Weighted Frequency Academic ED Safety Net % 205.0 225.0 241.0 144.0 61.3 109.5 162.0 76.6 29.88 48.68 67.22 53.17 95% CI (15.77, (35.23, (56.33, (34.77, 43.99) 62.13) 78.11) 71.58) Non-Academic Metropolitan EDs Safety Net Non-Academic Metropolitan EDs Weighted Frequency Weighted Frequency Non-Academic ED Safety Net % 95% CI 272.0 460.0 852.0 445.0 39.4 89.2 555.6 158.7 14.47 19.39 65.21 35.65 (5.09, 23.85) (11.30, 27.48) (58.14, 72.27) (25.64, 45.67) S136 2012 SAEM ANNUAL MEETING ABSTRACTS markedly variable by region. Such findings likely have important policy implications. Our study has several limitations: this is a preliminary descriptive analysis; temporal confounders may affect our findings; our definition of ‘‘safety net’’ may not be optimal; and the academic status of the affiliated hospital may not reflect the academic status of the ED. 250 Impact of Health Care Reform in Massachusetts on Emergency Department and Hospital Utilization Peter Smulowitz1, Xiaowen Yang2, James O’Malley3, Bruce Landon3 1 Beth Israel Deaconess Medical Center, Boston, MA; 2Massachusetts Institute of Technology, Department of Economics, Boston, MA; 3 Department of Health Care Policy, Harvard Medical School, Boston, MA Background: Massachusetts’ health care reform dramatically reduced the number of uninsured in the state and served as a model for national health reform legislation. The implementation of MA health reform provides a unique opportunity to study the effect of a large-scale expansion of health insurance on the patterns of seeking health care services. Objectives: To examine the effect of MA health reform implementation on ED and hospital utilization before and after health reform, using an approach that relies on differential changes in insurance rates across different areas of the state in order to make causal inferences as to the effect of health reform on ED visits and hospitalizations. Our hypothesis was that health care reform (i.e. reducing rates of uninsurance) would result in increased rates of ED use and hospitalizations. Methods: We used a novel difference-in-differences approach, with geographic variation (at the zip code level) in the percentage uninsured as our method of identifying changes resulting from health reform, to determine the specific effect of Massachusetts’ health care reform on ED utilization and hospitalizations. Using administrative data available from the Massachusetts Division of Health Care Finance and Policy Acute Hospital Case Mix Databases, we compared a one-year period before health reform with an identical period after reform. We fit linear regression models at the area-quarter level to estimate the effect of health reform and the changing uninsurance rate (defined as self-pay only) on ED visits and hospitalizations. Results: There were 2,562,330 ED visits and 777,357 hospitalizations pre-reform and 2,713,726 ED visits and 787,700 hospitalizations post-reform. The rate of uninsurance decreased from 6.2% to 3.7% in the ED group and from 1.3% to 0.6% in the hospitalization group. A reduction in the rate of the uninsured was associated with a small but statistically significant increase in ED utilization (p = 0.03) and no change in hospitalizations (p = 0.13). Conclusion: We find that increasing levels of insurance coverage in Massachusetts were associated with small but statistically significant increases in ED visits, but no differences in rates of hospitalizations. These results should aid in planning for anticipated changes that might result from the implementation of health reform nationally. 251 Access to Appointments Following a Policy Change to Improve Medicaid Reimbursement in Washington DC Janice Blanchard, Rachelle Pierre Mathieu, Lara Oyedele, Rachel Nash, Adith Sekaran, Lauren Winter, Paige Diamant George Washington, Washington, DC Background: In April, 2009 the District of Columbia adopted a policy in which Medicaid reimbursement rates are now on par with Medicare. It is unclear whether this policy has improved access to care in this population. This policy is similar to what will be adopted as part of the nationwide health reform initiative. Objectives: To evaluate changes in realized access to care since implementation of a local policy change affecting Medicaid reimbursement. We hypothesized that access would be improved among persons with Medicaid in 2011 as compared to years before implementation of the policy change. Methods: We used a scripted hypothetical patient scenario of a patient with hypertension seen in the emergency department requiring close outpatient follow-up. We compared our results in 2011 to prior results from studies conducted with similar methodology in 2005 and 2008. Calls were made to private providers (Medicare, Medicaid, uninsured, and private insurance scenarios) and safety net clinics (Medicaid HMO and uninsured scenarios). Appointment success rates were compared across scenarios using bivariate (chi-square) analysis. Results: Calls were made to a total of 31 private provider offices (Medicaid, private, uninsured, and Medicare scenarios) and 35 safety net clinics (uninsured, Medicaid, and DC Alliance safety net scenarios). When comparing 2011 appointment success rates, the Medicaid fee for service scenario calls to private providers had the lowest success rate (30.8%) as compared to the highest success rate of 50.6% among the private insurance scenario calls (p = 0.09). Analyzing trends over time, access to appointments for private providers for all insurance scenarios combined decreased from 60% in 2005 to 53.8% in 2008 and 46.2% in 2011 (p = 0.02 comparing 2005 to 2011). For the Medicaid fee for service scenario, there was no significant change in appointment success rate before and after implementation of the reimbursement changes. Conclusion: Despite high rates of insurance coverage in Washington DC, our study indicates that accessing care was more difficult for consumers in 2008 and 2011 as compared to 2005. Policy changes designed to improve Medicaid reimbursement to private providers did not improve appointment accessibility among the Medicaid fee for service population. Health reform initiatives that expand insurance coverage should also address realized access to care. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 252 Access to Urgent Pediatric Primary Care Appointments in the District of Columbia Rachelle Pierre Mathieu, Janice Blanchard, Adith Sekaran, Rachel Nash, Lauren Winter, Christine Prideaux George Washington, Washington, DC Background: Timely access to acute primary care appointments after an emergency department (ED) visit has become a challenge for both providers and patients. Previous studies have documented disparities in accessing adult primary care and pediatric specialty care, especially among those lacking private insurance. There are little data regarding urgent pediatric primary care access. Objectives: This study measured pediatric access to urgent primary care appointments within the District of Columbia following an ED visit. We hypothesized there would be a disparity in access for uninsured children and those with Medicaid. Methods: We used mystery caller methodology to evaluate rates of appointment access for pediatric patients. Calls were made to randomly selected private pediatric practices as well as pediatricians at safety net clinics. Research assistants posed as a parent calling to secure an urgent appointment for their child following a recent ED visit for a urinary tract infection using a standardized clinical script varying by insurance status. We calculated rates of appointment success as well as average appointment wait time, and analyzed differences using bivariate (chisquare) analysis. Results: We sampled 57 safety net clinics and 29 private clinics. As compared to private scenario calls made to private providers (36.8% success rate), appointment success rates were lowest for the Medicaid scenario calls made to private providers (27.8%). Calls made to safety net providers for the Medicaid patient scenario (48.8%, p = 0.38) and uninsured patient scenario (47.7%, p = 0.42) had higher appointment success rates but longer wait times. Average appointment wait time at safety net clinics was 12.3 days (95% CI, 3.5 to 21.1) for Medicaid patients and 10.4 days (95% CI, 6.7 to 14.1) for uninsured patients. Average appointment wait times for the privately insured at private practices were 1.9 days (95% CI, 1.0 to 2.7). Conclusion: This study demonstrated disparities in access to urgent pediatric primary care appointments by health insurance in the District. Although appointment success rates were not different by practice setting or insurance type, wait times were significantly longer for callers to safety net providers as compared to private practices. Pediatric provider access needs to be improved with public and private insurance expansions in the wake of health reform. • 253 www.aemj.org S137 Comparison Of Three Prehospital Cervical Spine Protocols With Respect To Immobilization Requirements And Missed Injuries Rick Hong, Molly Meenan, Erin Prince, Ronald Murphy, Caitlin Tambussi, Rick Rohrbach, Brigitte M. Baumann Cooper University Hospital, Camden, NJ Background: The ideal cervical spine immobilization protocol avoids unnecessary immobilization and potential morbidity while properly immobilizing patients at greatest risk of cervical spine injury. Objectives: We compared three immobilization protocols, the Prehospital Trauma Life Support (PHTLS) (mechanism-based), the Domeier protocol (parallels NEXUS criteria), and the Hankins criteria (requires immobilization for those <12 or >65 yrs, with altered consciousness, focal neurologic deficit, distracting injury, and midline or paraspinal tenderness) to determine the number of patients who would require cervical immobilization. Our secondary objective was to determine the percentage of missed cervical spine injuries, had each protocol been followed with 100% compliance. Methods: Design: Cross sectional. Setting: Inner city ED. Subjects: Patients ‡18 yrs transported by EMS after a traumatic mechanism. For patients meeting inclusion criteria, a structured data form which obtained demographics and data for all three protocols was completed by the treating physician immediately after the history and physical exam but before radiologic imaging. Medical record review ascertained cervical spine injuries. Both physicians and data abstractors were blinded to the objective of the study. Analysis: Chi-square. Results: Of the 498 enrolled patients, 58% were male and the mean age was 48 ± 20 yrs. The following proportions of patients would have required cervical spine immobilization based on the respective protocols: PHTLS, 95.4% (95% CI 93.1–96.9%); Domeier, 68.7% (95% CI 64.5–72.6%); Hankins, 81.5% (95% CI 77.9– 84.7%), p < 0.001. There were a total of 16 (3.2%) cervical spine injuries: 11 (2%) vertebral fractures, 2 (0.4%) subluxations/dislocations, and 3 (0.6%) spinal cord injuries. Complete compliance with each of the three protocols would have led to appropriate cervical spine immobilization of all injured patients. Conclusion: The mechanism-based PHTLS protocol required immobilization of the greatest percentage of patients, as compared to the Domeier and Hankins protocols. Although physician-determined presence of cervical spine immobilization criteria cannot be generalized to the findings obtained by EMS personnel, our findings suggest that mechanism-based criteria may result in unnecessary cervical spine immobilization without any benefit to injured patients. S138 254 2012 SAEM ANNUAL MEETING ABSTRACTS The Cost-Effectiveness Of Improvements In Prehospital Trauma Triage In The U.S M. Kit Delgado1, David Spain1, Sharada Weir2, Jeremy Goldhaber-Fiebert1 1 Stanford University School of Medicine, Stanford, CA; 2University of Massachusettes Medical School, Shrewsbury, MA Background: Trauma centers (TC) reduce mortality by 25% for severely injured patients but cost significantly more than non-trauma centers (NTC). The CDC 2009 field triage guidelines set targets to reduce undertriage of these patients to NTC to <5% and reduce overtriage of minor injury patients to TC to <50%. Objectives: Determine the cost-effectiveness of reaching CDC targets for prehospital trauma triage performance in U.S. regions with <1 hour EMS access to Level I TC. Methods: Using a decision-analytic Markov model, we evaluated the effect of incremental improvements in prehospital trauma triage performance on costs and survival given a baseline undertriage rate of major injury patients to NTC of 20% and overtriage rate of minor trauma patients to TC of 75%. The model followed patients from injury through prehospital care, hospitalization, first year post-discharge, and the remainder of life. Patients were trauma victims with a mean age of 43 (range: 18–85) with Abbreviated Injury Scores (AIS) from 1–6. Cost and survival probability inputs were derived from the National Study on the Costs and Outcomes of Trauma for patients with moderate to severe injury (AIS 3–6), National Trauma Data Bank, and published literature for patients with minor injury (AIS 1–2). Outcomes included costs (2009$), quality adjusted life-years (QALY), and incremental costeffectiveness ratios. Results: Reducing undertriage rates from 20% to 5% would yield 4.0 (95% CI 3.5–4.4) QALYs gained (or 0.16 to 0.20 lives saved) per 100 patients transported by EMS. Reducing overtriage rates from 75% to 50% would save $108,000 per 100 patients transported. Reducing undertriage is cost-effective at less than $100,000/QALY as long as overtriage rates do not proportionally increase by a factor >0.7. Ideal simultaneous reductions of undertriage to 5% and overtriage to 50% would be cost-effective at $17,400/QALY gained. Results were only sensitive to situations in which the cost of treating patients with minor injures at TC, relative to NTC, was smaller than expected. Conclusion: Reducing prehospital undertriage of trauma patients would be cost-effective provided overtriage does not proportionally increase by a factor >0.7. Improving prehospital trauma triage should be a national priority; reducing undertriage by 15% could save 7,000–8,800 lives/year and reducing overtriage by 25% could save up to $4.8 billion/year. Emergency Medical Services Compliance With Prehospital Trauma Life Support (PHTLS) Cervical Spine Immobilization Guidelines Rick Hong, Molly Meenan, Erin Prince, Caitlin Tambussi, Rachel Haroz, Michael E. Chansky, Brigitte M. Baumann Cooper University Hospital, Camden, NJ 255 Background: PHTLS provides guidelines for prehospital cervical spine immobilization in trauma patients and is followed by many EMS personnel. Yet, it is unknown how closely EMS providers adhere to these guidelines. Objectives: To determine EMS compliance with the PHTLS guidelines in trauma patients transported to a Level I trauma center ED who did not result in a trauma alert. Our secondary objective was to identify criteria associated with inadequate cervical spine immobilization by EMS which then led to immobilization in the ED. Methods: Design: Prospective cohort. Setting: Urban, academic ED. Subjects: Patients ‡18 years transported by EMS after a traumatic mechanism. Trained research associates screened all EMS patients for inclusion. For patients meeting inclusion criteria, a structured data form which obtained demographics and PHTLS data was completed by the treating physician immediately after the history and physical exam but before radiologic imaging. Both RAs and physicians were blinded to the objective of the study. Analysis: Chi-square and multivariable regression. Results: Of the 498 patients who were enrolled, 58% were male, mean patient age was 48 ± 20 years, and mean GCS was 14.8 ± 0.8. Of the 475 patients with at least one PHTLS criterion, 386 (81%) underwent cervical spine immobilization by EMS. Compliance with PHTLS criteria is presented in the table. Multivariable Table - Abstract 255: Implementation of PHTLS Cervical Spine Immobilization Criteria noted by physician (n = 498) Anatomic deformity of spine Unable to communicate Violent impact to: Head Neck Torso Pelvis Sustained a fall Ejected or fell from vehicle Sudden acceleration, deceleration, or bending forces 2 40 292 223 153 123 193 38 253 C collar placed by EMS (n = 386) 2 26 221 186 132 107 130 31 214 (7) (57) (48) (34) (28) (34) (8) (55) C collar placed by MD (n = 82) 0 12 60 34 19 12 44 6 31 (15) (73) (42) (23) (15) (54) (7) (38) p value 1 0.02 0.01 0.27 0.05 0.01 <0.001 0.83 0.004 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 analysis of PHTLS criteria demonstrated that traumatic impact to the head (OR 2.41, 95% CI 1.30–4.46) and mechanism of fall (OR 1.82, 95% CI 1.04–3.21) were associated with insufficient cervical spine immobilization by EMS. Conclusion: Compliance by EMS with PHTLS guidelines is insufficient and results in nearly one fifth of patients who are not properly immobilized. Trauma to the head and a mechanism of fall are the two PHTLS criteria that resulted in inadequate immobilization by EMS. • A Computer-Assisted Self-Interview Focused on Sexually Transmitted Infections in the Pediatric Emergency Department Is Easy To Use and WellAccepted Fahd A. Ahmad1, Katie Plax1, Karen K. Collins1, Donna B. Jeffe1, Kenneth B. Schechtman1, Jane Garbutt1, Dwight E. Doerhoff2, David M. Jaffe1 1 Washington University in St. Louis School of Medicine, St. Louis, MO; 2St. Louis Children’s Hospital, St. Louis, MO Background: Chlamydia trachomatis and Neisseria gonorrhea are sexually transmitted infections (STIs) with high levels of co-morbidity when untreated in adolescents. Despite broad CDC screening recommendations, many youth do not receive testing when indicated. The pediatric emergency department (PED) is a venue with a high volume of patients potentially in need of STI testing, but assessing risk in the PED is difficult given constraints on time and privacy. We hypothesized that patients visiting a PED would find an Audio-enhanced Computer-Assisted Self-Interview (ACASI) program to establish STI risk easy to use, and would report a preference for the ACASI over other methods of disclosing this information. Objectives: To assess acceptability, ease of use, and comfort level of an ACASI designed to assess adolescents’ risk for STIs in the PED. Methods: We developed a branch-logic questionnaire and ACASI system to determine whether patients aged 15–21 visiting the PED need STI testing, regardless of chief complaint. We obtained consent from participants and guardians. Patients completed the ACASI in private on a laptop. They read a one-page computer introduction describing study details and completed the ACASI. Patients rated use of the ACASI upon completion using five-point Likert scales. Results: 2030 eligible patients visited the PED during the study period. We approached 873 (43%) and enrolled and analyzed data for 460/873 (53%). The median time to read the introduction and complete the ACASI was 8.2 minutes (interquartile range 6.4– 11.5 minutes). 90.7% of patients rated the ACASI ‘‘very easy’’ or ‘‘easy’’ to use, 90.6% rated the wording as ‘‘very easy’’ or ‘‘easy’’ to understand, 60% rated the ACASI ‘‘very short’’ or ‘‘short’’, 60.3% rated the audio as ‘‘very helpful’’ or ‘‘helpful,’’ 82.9% were ‘‘very comfortable’’ or ‘‘comfortable’’ with the system S139 confidentiality, and 71.2% said they would prefer a computer interface over in-person interviews or written surveys for collection of this type of information. Conclusion: Patients rated the computer interface of the ACASI as easy and comfortable to use. A median of 8.2 minutes was needed to obtain meaningful clinical information. The ACASI is a promising approach to enhance the collection of sensitive information in the PED. 257 256 www.aemj.org The Association Between Prehospital Glasgow Coma Scale and Trauma Center Outcomes in Victims of Moderate and Severe TBI: At Statewide Trauma System Analysis Daniel W. Spaite1, Uwe Stolz1, Bentley J. Bobrow2, Vatsal Chikani3, Michael Sotelo2, Joshua B. Gaither1, Chad Viscusi1, David Harden3, Jason Roosa4, Kurt R. Denninghoff1 1 University of Arizona, Tucson, AZ; 2University of Arizona, Phoenix, AZ; 3Arizona Department of Health Services, Phoenix, AZ; 4Maricopa Integrated Health System, Phoenix, AZ Background: The Glasgow Coma Scale (GCS) is utilized widely for evaluation and decision-making in emergency medical services (EMS) systems. However, since linkage of EMS and trauma center (TC) outcome data is highly challenging, there is a paucity of large, multisystem studies that directly assess the association between EMS GCS and distal outcomes. Objectives: To evaluate the association between EMS GCS and outcomes in patients with moderate or severe (m/s) TBI in a statewide EMS system. Methods: The Arizona State Trauma Registry (ASTR) contains EMS and TC data from all trauma patients transported by 300 EMS agencies to any of the eight formally designated Level I TCs in the state. We evaluated the associations between initial EMS GCS and various TC outcomes in all m/s TBI cases based upon final diagnoses (CDC Barell Matrix Type 1, 1/1/09–12/31/10). GCS was grouped into four commonly used categories: 15–13 (mild), 12–9 (mod), 8–4 (severe), 3 (ominous). We compared survival, TC length of stay (LOS), TC charges, and final disposition across groups using Fisher’s exact or Kruskal-Wallis test. Results: There were a total of 6,985 m/s TBIs and 5,375 (77%) had documented EMS GCS (study group). The proportion in each GCS group was 15–13: 68.5%, 12–9: 8.6%, 8–4: 7.9%, 3: 15.1%. Survival was 97.8%, 91.7%, 74.1%, and 41.8% respectively (p < 0.001). Proportion of survivors discharged home vs. to rehab (R) or skilled care facility (SCF) was 82.7%, 59.8%, 38.6%, and 41.0% (p < 0.001). Median LOS (days, interquartile range) was 3.1 (1.6–6.1), 6.8 (3.2–13.3), 7.9 (2.4–15.9), and 2.0 (0.1– 10.3, early death very common) (p < 0.001). Median hospital charges (US$) per patient were $37,024 (21,336– 75,976), $91,861 (47,800–181,137), $125,501 (59,536– 237,886), and $62,006 (21,437–150,553) (p < 0.001). Conclusion: EMS GCS showed strong association with survival and other outcomes, with lower GCS associated with lower survival, greater hospital LOS, higher S140 2012 SAEM ANNUAL MEETING ABSTRACTS hospital charges, and greater proportion of survivors discharged to R/SCF. While initial GCS was strongly associated with outcomes, it is notable that two-thirds of all patients with final diagnosis of moderate or severe TBI had an initial GCS of ‡13. Furthermore, 17.3% of patients with ‘‘mild’’ GCS were discharged to R/SCF. Thus, while EMS GCS is useful for risk stratification, a substantial number of patients with a ‘‘good’’ prehospital GCS actually have significant TBI. Conclusion: The computer-based intervention shows promise for delivering content that decreases moderate dating victimization over 6 months. The therapist BI is promising for decreasing moderate dating victimization over 12 months and severe dating victimization over 3 months. ED-based BIs delivered on a computer addressing multiple risk behaviors could have important public health effects. 259 258 Dating Violence: Outcomes Following a Brief Motivational Interviewing Intervention Among At-Risk Adolescents in an Urban ED Lauren K. Whiteside1, Rebecca Cunningham1, Stephan T. Chermack2, Marc A. Zimmerman2, Jean T. Shope2, C. Raymond Bingham2, Frederick C. Blow2, Maureen A. Walton2 1 University of Michigan and Hurley Medical Center, Ann Arbor and Flint, MI; 2University of Michigan, Ann Arbor, MI Background: Dating violence is a serious cause of emotional and physical injury among adolescents. Studies of nationally representative samples show that one in ten high school students report being the victim of violence from a dating partner. A recent study demonstrated the efficacy of the SafERteens intervention on reducing peer violence among adolescents presenting to the emergency department (ED). Objectives: To determine the efficacy of this ED-based brief intervention (BI) on dating violence one year following the ED visit, among the subsample of adolescents in the intervention reporting past year dating violence. Methods: Patients (14–18 years old) presenting for medical illness or injury were recruited from an urban, Level I trauma center ED. Participants were eligible for the BI if they had past year violence and alcohol use. Participants were randomized to one of three conditions, BI delivered by a computer (CBI), BI delivered by a therapist assisted by a computer (TBI), or control, and completed 3, 6, and 12 month follow-up. In addition to content on alcohol misuse and peer violence, adolescents reporting dating violence received a tailored module on dating violence. The main outcome for this analysis was frequency of moderate and severe dating victimization and aggression at the baseline assessment and 3, 6, and 12 months post ED visit. Results: Among eligible adolescents, 55% (n = 397) reported dating violence and were included in these analyses. Compared to controls, after controlling for baseline dating victimization, participants in the CBI showed reductions in moderate dating victimization at 3 months (OR 0.7; CI 0.51–0.99; p < 0.05, effect size 0.12) and 6 months (OR 0.56; CI 0.38–0.83; p < 0.01, effect size 0.18); models examining interaction effects were significant for the CBI on moderate dating victimization at 3 and 6 months. Significant interaction effects were found for the TBI on moderate dating victimization at 6 and 12 months and severe dating victimization at 3 months. Relationship of Intimate Partner Violence to Health Status and Preventative Screening Behaviors Among Emergency Department Patients Anitha E. Mathew, L. Shakiyla Smith, Brittany Marsh, Debra Houry Emory University, Atlanta, GA Background: Intimate partner violence (IPV) is a health problem that many ED patients experience. It is unclear whether IPV victims engage in health screening behaviors at different rates than other women, placing them at varying risks for other diseases. Objectives: To assess the association of IPV with health status and preventative screening behaviors in an ED population. We hypothesized that IPV victims would report poorer physical and mental health and have lower frequencies of preventative screening behaviors than nonvictims. Methods: Adult female patients who presented to three EDs on weekdays from 11 AM to 7 PM were asked by trained research staff to participate in a computerized survey about ‘‘women’s health’’ over a 14-month period. Women were excluded if they were critically ill, did not speak English, intoxicated, or exhibited signs of psychosis. Validated measures were used, including the Universal Violence Prevention Screen to identify IPV and the Danger Assessment Scale to measure IPV severity. We also assessed respondents’ physical and mental health using the Short Form-12. Patients were asked about chronic disease history, including diagnoses of HIV and diabetes, if they had a regular doctor, and how often they had received pap smears, selfbreast exams, and doctor or nurse-performed breast exams. We used chi-square tests, t-tests, and linear regression analyses to measure associations. Results: 1,474 women out of 3381 approached (43.6%) agreed to take the survey. Age averaged 38 years ± 12.8 (range 18–68), and the majority of participants were black (n = 1218, 83.9%). 153 out of 832 women (18.4%) who had been in a relationship the previous year had experienced IPV. IPV victims were more likely to report positive HIV status (p = 0.017) and less likely to conduct monthly self-breast exams (p = 0.003) than nonvictims. Victims scored significantly lower in physical and mental health (38.6 [SD 3.9] and 45.1 [SD 5.5], respectively; p < 0.001) than the population mean of 50 on the Short Form-12. IPV victims who reported HIV testing had significantly higher Danger Assessment scores than victims who had not been tested (p = 0.048). Conclusion: IPV victims are more likely to report lower measures of health status, higher rates of HIV, and less ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 frequent self-breast exams than other female ED patients, putting them at risk for other chronic diseases. • www.aemj.org S141 data suggest that other cities should consider similar ordinances to prevent unwanted consequences of alcohol. Table - Abstract 260: Comparison of Visits by Year 260 The 21-Only Ordinance Reduced Alcohol-Related Adverse Consequences among College-Aged Adults Michael E. Takacs, Nicholas Edwards, Christopher Peterson, Gregory Pelc University of Iowa, Iowa City, IA Background: In an effort to reduce underage drinking, Iowa City, IA adopted a 21-only ordinance prohibiting persons under 21 from being in bars after 10 PM. This ordinance went into effect on June 1, 2010. Preliminary data studied over a 3-month period showed a significant decrease in alcohol related emergencies. Iowa City is mainly a college town and home of the University of Iowa (UI). Objectives: The primary goal was to determine whether there was a change in alcohol-related (AR) emergency visits over a one year period from the start of the ordinance. Secondary goals were to measure the effect on underage alcohol emergencies, student alcohol-related emergencies, and arrests for public intoxication. Methods: We performed a retrospective cohort study of emergency department patients, ages 18 to 22, presenting for AR reasons from June 1, 2010 to May 31, 2011 (AY2010) and compared data to the same time period a year earlier (AY2009). Data were also obtained from UI student records and public arrest data. Pearson chi square analysis compared categorical variables. Results: There were 2954 total visits in AY2009 and 2989 total visits in AY2010. AR visits decreased from 453 in AY2009 (15.3%) to 359 in 2010 (12.0%, p < 0.001) and underage AR visits decreased from 15.9% to 12.3%, p < 0.005, see table. UI student AR visits decreased by 16% from 218 to 183, and Non-UI student AR visits decreased by 25% from 235 to 176. Public intoxication bookings for 18 to 20 year olds decreased from 598 to 409 (32%), Figure 1. Conclusion: The 21-only ordinance was associated with a significant reduction of AR visits. This ordinance was also associated with reduction in underage AR visits, UI student visits, and public intoxication bookings. These AY2009 AY2010 AY2009 AY2010 261 Total Visits AR Visits Rate per 100 2954 2989 453 359 15.3 12.0 <0.001 Underage Visits Underage AR Visits 1685 1664 268 204 15.9 12.3 <0.005 P Value Factors Affecting Success of Prehospital Intubation in an Air and Land Critical Care Transport Service: Results of a Multivariate Analysis Anna MacDonald1, RD MacDonald2, Jacques S. Lee3 1 University of Toronto, Toronto, ON, Canada; 2 Ornge Transport Medicine, Mississauga, ON, Canada; 3Sunnybrook Health Sciences Centre, Toronto, ON, Canada Background: Prehospital providers perform tracheal intubation in the prehospital environment, and failed attempts are of concern due to the danger of hypoxia and hypotension. Some question the appropriateness of intubation in this setting due to the morbidity risk associated with intubation in the field. Thus it is important to gain an understanding of the factors that predict the success of prehospital intubation attempts to inform this discussion. Objectives: To determine the factors that affect success rates on first attempt of paramedic intubations in a rapid sequence intubation (RSI) capable critical care transport service. Methods: We conducted a multivariate logistic analysis on a prospectively collected database of airway management from an air and land critical care transport service that provides scene responses and interfacility transport in the Province of Ontario. The study population includes all intubations performed by flight paramedics from January 2006 to July 2009. The primary outcome is success on first attempt. A list of potential factors predicting success was obtained from a review of the literature and included age, sex, Glasgow Coma Scale, location of intubation attempt, paralytics and sedation given, a difficult airway prediction score, and type of call (trauma, medical, or cardiac arrest). Results: Data from 549 intubations were analysed. The success rate on first attempt at intubation was 317/549 (57.7%) and the overall success rate was 87.4%. The mean age was 43.5 years and 69.4% were male and 56.4% were trauma patients. Of these, 498 had complete data for all predictive variables and were included in the multivariate analysis. The factors that were found to be statistically significant were age per decade (OR 1.12, CI 1.04–1.2), female sex (OR 1.5, CI 1.03–2.32), paralytics given (OR 2.66, CI 1.5–4.7), and sedation given (OR 0.61, S142 2012 SAEM ANNUAL MEETING ABSTRACTS CI 0.41–0.91). This model demonstrated a good fit (Hosmer Lemeshow = 8.906) with an AUC of 0.632. Conclusion: Use of a paralytic agent, age, and sex were associated with increased success of intubation. The association of sedative use only with decreased success of intubation was unexpected and may be due to confounding related to the indications for sedation, such as patient agitation. Our findings may have implications for RSI-capable paramedics and require further study. 262 Incidence and Predictors of Psychological Distress after Motor Vehicle Collision Gemma C. Lewis1, Timothy F. Platts-Mills1, Robert Swor2, David Peak3, Jeffrey Jones4, Neils Rathlev5, David Lee6, Robert Domieir7, Phyllis Hendry8, Samuel A. McLean1 1 University of North Carolina, Chapel Hill, NC; 2William Beaumont Hospital, Royal Oak, MI; 3Massachusetts General Hospital, Boston, MA; 4Spectrum Health - Butterworth Campus, Grand Rapids, MI; 5Baystate Medical Center, Springfield, MA; 6North Shore University Hospital, Manhasset, NY; 7 St. Joseph’s Mercy Hospital, Ann Arbor, MI; 8 University of Florida, Jacksonville, FL Background: Motor vehicle collisions (MVCs) are one of the most common types of trauma for which people seek ED care. The vast majority of these patients are discharged home after evaluation. Acute psychological distress after trauma causes great suffering and is a known predictor of posttraumatic stress disorder (PTSD) development. However, the incidence and predictors of psychological distress among patients discharged to home from the ED after MVCs have not been reported. Objectives: To examine the incidence and predictors of acute psychological distress among individuals seen in the ED after MVCs and discharged to home. Methods: We analyzed data from a prospective observational study of adults 18–64 years of age presenting to one of eight ED study sites after MVC between 02/ 2009 and 10/2011. English-speaking patients who were alert and oriented, stable, and without injuries requiring hospital admission were enrolled. Patient interview included assessment of patient sociodemographic and psychological characteristics and MVC characteristics. Level of psychological distress in the ED was assessed using the 13-item Peritraumatic Distress Inventory (PDI). PDI scores >23 are associated with increased risk of PTSD and were used to define substantial psychological distress. Descriptive statistics and logistic regression were performed using Stata IC 11.0 (StataCorp LP, College Station, Texas). Results: 9339 MVC patients were screened, 1584 were eligible, and 949 were enrolled. 361/949 (38%) participants had substantial psychological distress. After adjusting for crash severity (severity of vehicle damage, vehicle speed), substantial patient distress was predicted by sociodemographic factors, pre-MVC depressive symptoms, and arriving to the ED on a backboard (table). Conclusion: Substantial psychological distress is common among individuals discharged from the ED after MVCs and is predicted by patient characteristics separate from MVC severity. A better under standing of the frequency and predictors of substantial psychological distress is an important first step in identifying these patients and developing effective interventions to reduce severe distress in the aftermath of trauma. Such interventions have the potential to reduce both immediate patient suffering and the development of persistent psychological sequelae. Table - Abstract 262: Logistic Regression Analysis of Predictors of Peritraumatic Distress Predictor Age Female sex Educational attainment Pre-MVC depressive symptoms (CESD) Patient was driver Extent of vehicle damage Vehicle speed at impact Patient backboarded 263 Odds Ratio 95% CI P value 0.986 3.074 0.698 4.029 0.974–0.997 2.232–4.235 0.599–0.813 2.144–7.569 0.015 0.015 <0.001 <0.001 1.839 1.735 1.077 1.384 1.174–2.882 1.377–2.186 1.003–1.157 1.024–1.871 0.008 <0.001 0.041 0.035 Derivation of a Simplified Pulmonary Embolism Triage Score (PETS) to Predict the Mortality in Patients with Confirmed Pulmonary Embolism from the Emergency Medicine Pulmonary Embolism in the Real World Registry (EMPEROR) Beau Briese1, Don Schreiber2, Brian Lin3, Gigi Liu2, Jane Fansler4, Samuel Z. Goldhaber5, Brian J. O’Neil6, David Slattery7, Brian Hiestand8, Jeff A. Kline9, Charles V. Pollack10 1 Stanford/Kaiser Emergency Medicine Residency Program, Stanford, CA; 2Stanford University School of Medicine, Stanford, CA; 3 University of California, San Francisco, San Francisco, CA; 4Christian Hospital, Saint Louis, MO; 5Brigham and Women’s Hospital, Boston, MA; 6Wayne State University School of Medicine, Detroit, MI; 7University Medical Center of Southern Nevada, Las Vegas, NV; 8 Ohio State University College of Medicine, Columbus, OH; 9Carolinas Medical Center, 10 Charlotte, NC; Pennsylvania Hospital, Philadelphia, PA Background: The Pulmonary Embolism Severity Index (PESI) with 11 variables and simplified PESI (sPESI) with seven variables are the two leading risk stratification tools for mortality in patients with acute pulmonary embolism (PE). Objectives: To derive a simple four-variable prognostic model of mortality for patients with confirmed PE, the ‘‘Pulmonary Embolism Triage Score’’ (PETS), with equal performance characteristics to the more complicated PESI and sPESI. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Methods: PETS was retrospectively derived from 1438 patients with data to compute PESI in the Emergency Medicine Pulmonary Embolism in the Real World Registry (EMPEROR), an observational database of 1880 image-confirmed PE patients prospectively enrolled at 22 U.S. academic and community hospitals from 1/1/ 2005 to 12/29/2008. Logistic regression identified four variables that when positive were independently associated with higher 30-day mortality: presence of massive pulmonary embolism (SBP<90 mm Hg), tachypnea (respiratory rate‡24 breaths per minute), history of cancer, and leukocytosis (WBC>11,000 cells per cubic mm). PETS was defined as HIGH if any of the four variables was positive, LOW if otherwise. (See figure) The predictive characteristics of PETS, PESI, and sPESI for 30-day mortality in EMPEROR, including AUC, negative predictive value, sensitivity, and specificity were calculated. • www.aemj.org S143 Results: The 646 of 1438 patients (44.9%; 95% CI 42.3%–47.5%) classified as PETS LOW had 30-day mortality of 0.5% (95% CI 0.1–1.5%), versus 10.2% (95% CI 8.0%–12.4%) in the PETS HIGH group, statistically similar to PESI and sPESI. PETS is significantly more specific for mortality than the sPESI (47.0% v 37.6%; p < 0.0001), classifying far more patients as low-risk while maintaining a sensitivity of 96% (95% CI 88.3%– 99.0%), not significantly different from sPESI or PESI (p > 0.05). Conclusion: With four variables, PETS in this derivation cohort is as sensitive for 30-day mortality as the more complicated PESI and sPESI, with significantly greater specificity than the sPESI for mortality, placing 25% more patients in the low-risk group. External validation is necessary. S144 264 2012 SAEM ANNUAL MEETING ABSTRACTS Denver Trauma Organ Failure Score Outperforms Traditional Methods of Risk Stratification in Trauma Nicole Seleno, Jody Vogel, Michael Liao, Emily Hopkins, Richard Byyny, Ernest Moore, Craig Gravitz, Jason Haukoos Denver Health Medical Center, Denver, CO Background: The Sequential Organ Failure Assessment (SOFA) Score, base excess, and lactate have been shown to be associated with mortality in critically ill trauma patients. The Denver Emergency Department (ED) Trauma Organ Failure (TOF) Score was recently derived and internally validated to predict multiple organ failure in trauma patients. The relationship between the Denver TOF Score and mortality has not been assessed or compared to other conventional measures of mortality in trauma. Objectives: To compare the prognostic accuracies of the Denver ED TOF Score, ED SOFA Score, and ED base excess and lactate for mortality in a large heterogeneous trauma population. Methods: A secondary analysis of data from the Denver Health Trauma Registry, a prospectively collected database. Consecutive adult trauma patients from 2005 through 2008 were included in the study. Data collected included demographics, injury characteristics, prehospital care characteristics, response to injury characteristics, ED diagnostic evaluation and interventions, and in-hospital mortality. The values of the four clinically relevant measures (Denver ED TOF Score, ED SOFA score, ED base excess, and ED lactate) were determined within four hours of patient arrival, and prognostic accuracies for in-hospital mortality for the four measures were evaluated with receiver operating characteristic (ROC) curves. Multiple imputation was used for missing values. Results: Of the 4,355 patients, the median age was 37 (IQR 26–51) years, median injury severity score was 9 (IQR 4–16), and 81% had blunt mechanisms. Thirty-eight percent (1,670 patients) were admitted to the ICU with a median ICU length of stay of 2.5 (IQR 1–8) days, and 3% (138 patients) died. In the non-survivors, the median values for the four measures were ED SOFA 5.0 (IQR 0.0– 8.0); Denver ED TOF 4.0 (IQR 4.0–5.0); ED base excess 7.0 (IQR 8.0–19.0) mEq/L; and ED lactate 6.5 (IQR 4.5– 11.8) mmol/L. The areas under the ROC curves for these measures are demonstrated in the figure. Conclusion: The Denver ED TOF Score more accurately predicts in-hospital mortality in trauma patients as compared to the ED SOFA Score, ED base excess, or ED lactate. The Denver ED TOF Score may help identify patients early who are at risk for mortality, allowing for targeted resuscitation and secondary triage to improve outcomes in these critically ill patients. 265 The Relationship Between Early Blood Pressure Goals and Outcomes in Post-Cardiac Arrest Syndrome Patients Treated with Therapeutic Hypothermia Maria Beylin, Anne Grossestreuer, Frances Shofer, Benjamin S. Abella, David F. Gaieski University of Pennsylvania School of Medicine, Philadelphia, PA Background: The 2010 American Heart Association Guidelines for Post-Cardiac Arrest Care Consensus recommend immediate treatment of hypotension to maintain adequate tissue perfusion. If mean arterial pressure (MAP) is <65 mmHg, they recommend infusion of fluid and use of vasopressors to restore adequate pressure. However, there is no literature to date examining the relationship between early blood pressure goals and outcomes in post-cardiac arrest syndrome (PCAS) patients treated with therapeutic hypothermia (TH). Objectives: We examined the relationship between MAP at specific time points post-arrest and neurologic and survival outcomes for PCAS patients. Methods: All consecutive PCAS patients treated with algorithmic post-arrest care including hemodynamic optimization and TH at the University of Pennsylvania Health System between May, 2005 and October, 2011 were included. Hemodynamic data, including MAP and number of vasopressors, were analyzed at 1, 6, 12, and 24 hour time points after return of spontaneous circulation. Outcomes data collected included survival to hospital discharge. Data were analyzed using logistic regression analysis and ANOVA in repeated measures over time. Results: 168 patients were included in the analysis; 45% (75/168) survived, and 35% (58/168) had a good neurological outcome at hospital discharge. The majority of the 168 patients were at or above goal MAP (80– 100 mmHg) at all time points and between 51–59% were on at least one vasopressor at any time point. In the linear regression model, increasing vasopressor use (point estimate [PE] 0.41; 95% confidence intervals [95% CI] 0.269–0.626) and increasing age (PE 0.977; 95% CI 0.955–1.00) were associated with worsened survival. In addition, higher MAP showed a trend toward improved survival (PE 1.015; 95% CI 1.0–1.32). In the ANOVA analyses, at all time points MAP was higher in survivors than non-survivors. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • Initial Lactate Level Not Associated With Mortality in Post-Arrest Patients Treated with Therapeutic Hypothermia David F. Gaieski, Anne Grossestreuer, Marion Leary, Lance B. Becker, Benjamin S. Abella University of Pennsylvania School of Medicine, Philadelphia, PA Background: Lactate levels rise during the no- or lowflow states associated with cardiac arrest. Prior studies have demonstrated that initial post-arrest lactate levels correlate with mortality (higher levels associated with worse outcomes). These findings need further validation in health systems delivering state-of-the-art post-arrest care including therapeutic hypothermia (TH), hemodynamic optimization, and other aspects of modern critical care. Objectives: To assess the relationship between initial lactate levels and outcomes in post-cardiac arrest syndrome patients. Methods: A retrospective chart review was performed of 155 post-cardiac arrest syndrome (PCAS) patients admitted to an academic medical center between April 2005 and October 2010 who were treated with bundled post-arrest care and had serial lactate values drawn during the first 24 hours after return of spontaneous circulation (ROSC). Unadjusted (t-test) analyses were performed to determine the association between initial and serial lactate values and mortality in these patients. Results: The mean post-ROSC lactate value (121 patients) was 9.5 mmol/L (SD 5.0; range 1.2–21.9). This value did not correlate with mortality, including when limited to initial lactate values >10 mmol/L. The mean value for the second lactate measured (105 patients) was 6.0 mmol/L (SD 4.6; range 0.7–20.0); higher second lactate values were associated with higher mortality (p = 0.049). The mean value for 12-hour post-ROSC lactate (130 patients) was 3.6 mmol/L (SD 2.8; range 0.3–14.3); higher 12-hour lactate values were associated with higher mortality (p = 0.027). Similar results were seen at 24 hours. Conclusion: Initial lactate levels are not associated with mortality in PCAS patients treated with bundled postarrest care. Subsequent lactate levels (second, 12-hour, and 24-hour) were associated with mortality with lower levels associated with better survival. 267 Prehospital Initiation Of Therapeutic Hypothermia In Adult Patients After Cardiac Arrest Does Not Improve Time To Target Temperature Eric M. Schenfeld1, David A. Pearson1, Jonathan Studnek2, Marcy Nussbaum3, Kathi Kraft4, Alan C. Heffner5 1 Carolinas Medical Center, Department of Emergency Medicine, Charlotte, NC; 2Carolinas Medical Center, The Center for Prehospital S145 Medicine, Charlotte, NC; 3Carolinas Medical Center, Dickson Institute for Health Studies, Charlotte, NC; 4Carolinas Medical Center, Center for Clinical Data Analysis, Charlotte, NC; 5Carolinas Medical Center, Department of Emergency Medicine, Department of Internal Medicine, Division of Critical Care Medicine, Charlotte, NC Conclusion: In comatose PCAS patients undergoing TH, early optimization of MAP as a measure of adequacy of perfusion may improve outcomes. Further prospective studies with specific MAP goals and hemodynamic optimization algorithms need to be performed. 266 www.aemj.org Background: Both animal and human studies suggest that early initiation of therapeutic hypothermia (TH) and rapid cooling improve outcomes after cardiac arrest. Objectives: The objective was to determine if administration of cold IV fluids in a prehospital setting decreased time-to-target-temperature (TT) with secondary analysis of effects on mortality and neurological outcome. Methods: Patients resuscitated after out-of-hospital cardiac arrest (OOHCA) who received an in-hospital post cardiac arrest bundle including TH were prospectively enrolled into a quality assurance database from November 2007 to November 2011. On April 1, 2009 a protocol for intra-arrest prehospital cooling with 4C normal saline on patients experiencing OOHCA was initiated. We retrospectively compared TT for those receiving prehospital cold fluids and those not receiving cold fluids. TT was defined as 34C measured via Foley thermistor. Secondary outcomes included mortality, good neurological outcome defined as Cerebral Performance Category (CPC) score of 1 or 2 at discharge, and effects of pre-ROSC cooling. Results: There were 132 patients who were included in this analysis with 80 patients receiving prehospital cold IV fluids and 52 who did not. Initially, 63% of patients were in VF/VT and 36% asystole/PEA. Patients receiving prehospital cooling did not have a significant improvement in TT (256 minutes vs 271 minutes, p = 0.64). Survival to discharge and good neurologic outcome were not associated with prehospital cooling (54% vs 50%, p = 0.67) and CPC of 1 or 2 in 49% vs 44%, (p = 0.61). Initiating cold fluids prior to ROSC showed both a nonsignificant decrease in survival (48% vs 56%, p = 0.35) and increase in poor neurologic outcomes (42% vs 50%, p = 0.39). 77% of patients received £ 1L of cooled IVF prior to hospital arrival. Patients receiving prehospital cold IVF had a longer time from arrest to hospital arrival (44 vs 34 min, p =< 0.001) in addition to a prolonged ROSC to hospital time (20 vs 12 min, p = 0.005). Conclusion: At our urban hospital, patients achieving ROSC following OOHCA did not demonstrate faster TT or outcome improvement with prehospital cooling compared to cooling initiated immediately upon ED arrival. Further research is needed to assess the utility of prehospital cooling. 268 Assessment of Building Type and Provision of Bystander CPR and AED Use in Public Out-of-Hospital Cardiac Arrest Events Kenneth Jones1, Steven C. Brooks2, Jonathan Hsu3, Jason Haukoos4, Bryan F. McNally5, Comilla Sasson4 1 University of Colorado, Aurora, CO; 2Division of Emergency Medicine, Department of S146 2012 SAEM ANNUAL MEETING ABSTRACTS Medicine, University of Toronto, Toronto, ON, Canada; 3Undergraduate Medical Education Program, Faculty of Medicine, University of Toronto, Toronto, ON, Canada; 4Department of Emergency Medicine, University of Colorado, Aurora, CO; 5Department of Emergency Medicine, Emory University, Atlanta, GA Background: Approximately 20% of all out-of-hospital cardiac arrests (OHCA) occur in a public location. However, little is known about how differences in building type affect the provision of bystander CPR and AED use for OHCA victims. Objectives: To categorize the locations of public OHCA events and to estimate and compare the prevalence of AED use and bystander CPR performance between these building types. Methods: Design: Secondary analysis of the Cardiac Arrest Registry to Enhance Survival (CARES) dataset. Setting: CARES comprises 29 U.S. cities in 17 states with a catchment of over 22 million people. Population: Consecutive arrests restricted to those that occurred in a public location (excluding street/highway) and had an AED used by either layperson or first responder from October 1, 2005 through December 31, 2009. Street address and latitude/longitude information were used to identify the type of building. Buildings were categorized using an adaptation of the City of Toronto Employment Survey Classification Protocols. Satellite and street imagery and property/tax records were used to categorize the business/institution. Results: A total of 20,018 arrests occurred during the study period. Of the 971 public arrests which had an AED used, 77 were unable to be geocoded, while an additional 150 were found to be misclassified. The most common reasons for misclassification were the use of the categories ‘‘Public Building’’ and ‘‘Other’’ for structures that were non-public (e.g. private residences or health care facility). The final sample included 744 OHCA events (see table). An AED was used by a layperson in 224 (30.1%) or first responder in 520 (69.9%) of these arrests, while bystander CPR was performed in 441 arrests (59.3%). Almost one in six arrests where an AED was used occurred in a retail shopping center. Colleges (1.2%, n = 10) and convention centers (1.0%, n = 8) made up a small percentage of the public arrests where an AED was used, but a large proportion of events occurring in these locations received bystander CPR (80, 87%) and had a layperson use an AED (60, 62%). Conclusion: Building types have differing layperson bystander CPR and AED use. Building types in which larger proportions of OHCA victims are receiving these life-saving interventions could be models for successful public education campaigns. 269 Ethnic Disparities In The Utilization Of EMS Resources And The Impact On Doorto-PCI Times In STEMI Patients Nick Testa, Garren MI Low, David Shavelle, Stephanie Hall, Kim Newton, Linda Chan Los Angeles County + USC Medical Center, Los Angeles, CA Background: EMS literature has shown that the use of 9-1-1 services varies with ethnicity. Objectives: In this study, we examined whether ethnic variability exists among STEMI patients utilizing EMS and whether there is a difference in door-to-PCI time between ethnic groups. Methods: We analyzed data prospectively collected from January 2009 through June 2011 to assess the difference in door-to-PCI time between the two largest ethnic groups of STEMI patients (Hispanic and African Americans). Results: Of a total of 494 STEMI patients, 276 (56%) were Hispanic and 95 (19%) were African American. African Americans utilized EMS more than Hispanics (75% versus 40%, p < 0.0001). To assess the difference in door-to-PCI time between the two ethnic groups, we Table - Abstract 268: Building Type and Percentage of Layperson Bystander CPR and AED Use Location Type Retail Shopping / Services Office Jail/Prison Hotel/Motel Industrial/Warehouse Facility Airport (Civil) Fitness Club Gambling Establishment Place of Worship Primary/Secondary School Restaurant / Bar / Nightclub Country Club / Golf Course Outdoor Athletic Facility Police/Fire/EMS facility College / University Stadium Convention Center Other Events (n) Events (%) Events with Layperson CPR (n) Events with Layperson CPR (%) Events with Layperson AED Use (n) Events with Layperson AED Use (%) 146 82 63 58 48 45 40 33 32 30 27 24 22 12 10 9 8 55 14.9% 8.6 7.7 7.1 5.8 5.5 4.9 4.0 3.9 3.7 3.6 2.9 3.0 1.6 1.3 1.2 1.1 7.4 66 57 52 18 24 34 30 13 15 22 14 20 17 3 8 5 7 36 44.3 69.0 82.5 31.0 50.0 75.6 75.0 39.4 46.9 73.3 55.0 83.3 77.3 25.0 80.0 55.6 87.5 65.4 17 31 35 0 16 22 22 10 4 16 2 11 6 0 6 4 5 17 13.9 38.0 55.6 0 33.3 48.9 55.0 30.3 12.5 53.3 5.0 45.8 27.3 0 60.0 44.4 62.5 30.9 ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 compared the median and percent of door-to-PCI times <90 minutes. The median door-to-PCI time was 46 minutes for African Americans and 70 minutes for Hispanics (p = 0.13). The median door-to-PCI time was 43 minutes for EMS patients and 83 minutes for walk-in patients (p < 0.0001). The percentage of patients with door-to-PCI time less than 90 minutes was 75% for Hispanics and 84% for African Americans (p = 0.48). The percentage less than 90 minutes was 94% for patients who arrived via EMS and was 60% for walk-in patients (p < 0.0001). Both measures of door-to-PCI time point to a significant difference in door-to-PCI time between EMS and walk-in patients but not between the two ethnic groups. The odds ratio of having a door-to-PCI time less than 90 minutes for Hispanics as compared to African Americans was 0.58 (95% CI: 0.16, 1.95; p = 0.48). After adjustment for the difference in patients who arrive via EMS and walk-in utilization, the adjusted odds ratio was 1.79 (95% CI: 0.36, 8.16; p = 0.67). The reverse relationship between the unadjusted and adjusted odds ratio indicates an interaction between ethnicity and patients mode of entry. Among patients who arrive via EMS, the odds ratio of having door-toPCI time less than 90 minutes for Hispanics as compared to African Americans was 0.69 (95% CI: 0.03, 7.29, p = 1.00) while the odds ratio was 4.80 (95% CI: 0.41, 126; p = 0.30) among the walk-in patients. Conclusion: These findings indicate that there is a significant difference between EMS and walk-in STEMI patients in door-to-PCI time but no significant difference between African American and Hispanic STEMI patients after adjusting for EMS utilization rates. 270 Identification of Delirium in Elderly Emergency Department Patients Maura Kennedy, Richard A. Enander, Richard E. Wolfe, Edward R. Marcantonio, Nathan I. Shapiro Beth Israel Deaconess Medical Center, Boston, MA Background: Delirium is common in the ED, associated with increased morbidity and mortality, and is often under-diagnosed by ED physicians. A better understanding of predictors of ED delirium would help to target case-finding, prevention, and treatment efforts. Objectives: The objective of this study was to identify independent risk factors for delirium in the ED utilizing patient demographic and clinical characteristics. Methods: We performed a prospective, observational study of elderly patients in our urban university ED. Inclusion criteria were: ED patient >= 65 years and ability of patient or surrogate to provide informed consent and cooperate with testing. A trained research assistant performed a structured mental status assessment including the MMSE and Delirium Symptom Interview. Delirium was determined using the Confusion Assessment Method (CAM). We collected data on patient demographics, comorbidities, medications, and ED course. Using the outcome of delirium from the CAM, we identified univariate correlates of delirium. Variables • www.aemj.org S147 with a p < 0.1 were included in a multivariate logistic regression model; backward selection was used to create a final model of significant predictor variables with a p <= 0.05. We allowed approximately 1 predictor per 10 delirious patients to avoid over-fitting the model. Model accuracy was assessed by the c-statistic and Hosmer Lemeshow tests. Results: There were 706 subjects enrolled with a complete delirium assessment, of whom 67 (9.5%) were delirious. Mean age was 77 (SD 8), 49% were male, 87% were Caucasian, and 11% were African American. The final multivariable predictor model consisted of age (OR 1.6; 95%CI 1.1–2.2 per decade above 65); history of dementia (4.3; 2.2–8.2), TIA or stroke (2.4; 1.1–5.2), or seizure disorder (3.8; 1.3–11.5); respiratory rate >20 in the ED (3.3; 1.6–6.9); and ED diagnosis of urinary tract infection (3.0; 1.4–6.8) or intracranial hemorrhage (5.9; 1.3–27.1). The c-statistic for our model was 0.80 and the Hosmer Lemeshow p-value was 0.17. Conclusion: Patients with preexisting neurological disease, evidence of respiratory distress, or presenting with a urinary tract infection or intracranial hemorrhage were most likely to present with ED delirium. This simple risk prediction model, easily performed in the emergency setting, demonstrated excellent discrimination and calibration. 271 Hospitalization Rates and Resource Utilization of Delirious ED Patients Maura Kennedy, Richard A. Enander, Richard E. Wolfe, Nathan I. Shapiro, Edward R Marcanatonio Beth Israel Deaconess Medical Center, Boston, MA Background: Delirium is common in elderly ED patients and under-diagnosed in the ED. Delirium in the hospital is associated with higher health care expenditures and mortality, but these associations have not been well-studied in ED delirium. Objectives: The purpose of this study is to compare resource utilization and mortality among delirious and non-delirious elderly ED patients. Methods: We performed a prospective, observational study of elderly patients in our urban university ED. Inclusion criteria were ED patients >= 65 years and ability of patient/surrogate to provide informed consent and cooperate with testing. A trained research assistant performed a structured mental status assessment after which delirium was determined using the Confusion Assessment Method. Patients were followed through their ED and hospital courses, and by telephone at 7 and 30 days. Data were collected on hospital and ICU admission, length of stay, hospital charges, and 30-day re-hospitalization and mortality. Proportions and Fisher’s exact tests are reported for nominal data. Medians, 25%–75% interquartile ranges (IQR), and Wilcoxon Rank tests are reported for non-normally distributed continuous data. Results: There were 706 subjects enrolled, of whom 67 (9.5%) were delirious. Delirious subjects were more likely to be admitted to the hospital than non-delirious patients (85% versus 63%, p < 0.001). Of the admitted S148 2012 SAEM ANNUAL MEETING ABSTRACTS patients, delirious subjects had a greater length of stay (median 4 days, IQR 2–6 days) compared with nondelirious patients (2 days, IQR 1–4 days; p < 0.001) and incurred greater hospital charges (median $17, 094, IQR $11,747–$31,825 vs. $13,317, IQR $8,789–$20,111; p = 0.004). There was a trend toward more ICU admissions among delirious patients (17% vs 9%, p = 0.06). At discharge, patients with ED delirium were more likely to be transferred to a new long term care or rehabilitation facility (49% vs 20%, p < 0.001). At 30 days, patients with ED delirium were more likely to have been re-admitted to the hospital (23% vs 11%, p = 0.006) or have died (8.8% vs 1.1%, p < 0.001). Conclusion: Delirium on presentation to the ED is associated with greater health care utilization and costs, and higher rates of discharge to a long term care facility and 30-day mortality. Further research is needed to determine if early diagnosis in the emergency setting may improve the outcome of patients presenting with delirium. 272 Physician and Prehospital Provider Unstructured Assessment of Delirium in the Elderly Adam N. Frisch, Brian P. Suffoletto, Thomas Miller, Christian Martin-Gill, Clifton W. Callaway UPMC, Pittsburgh, PA Background: An estimated 10% of emergency department (ED) patients 65 years of age and older have delirium, which is associated with short- and long-term risk of morbidity and mortality. Early recognition could result in improved outcomes, but the reliability of delirium recognition in the continuum of emergency care is unknown. Objectives: We tested whether delirium can be reliably detected during emergency care of elderly patients by measuring the agreement between prehospital providers, ED physicians, and trained research assistants using the Confusion Assessment Method for the ICU (CAM-ICU) to identify the presence of delirium. Our hypothesis was that both ED physicians and prehospital providers would have poor ability to detect elements of delirium in an unstructured setting. Methods: Prehospital providers and ED physicians completed identical questionnaires regarding their clinical encounter with a convenience sample of elderly (age >65 years) patients who presented via ambulance to two urban, teaching EDs over a three-month period. Respondents noted the presence or absence of (1) an acute change in mental status, (2) inattention, (3) disorganized thinking, and (4) altered level of consciousness (using the Richmond Agitation Sedation Scale). These four components comprise the operational definition of delirium. A research assistant trained in the CAM-ICU rated each component for the same patients using a standard procedure. We calculated inter-rater reliability (kappa) between prehospital providers, ED physicians, and research assistants for each component. Results: A total of 261 patients were enrolled. By CAMICU, 24 (9.2%) were positive for delirium. Physician and prehospital unstructured assessment would have catego- rized fewer patients as delirious (5.36%, 7.28%). For each component, there was poor to fair agreement (kappa 0.086–0.486) between observers (table). Conclusion: Prehospital providers, ED physicians and trained researchers using validated delirium instrument, have poor to fair agreement when assessing the components of delirium. Future studies should test whether more regimented bedside tools could improve agreement and recognition of delirium. Table - Abstract 272: Inter-rater Agreement for Delirium Components (Kappa Values) Delirium Component Acute Mental Status Change Agitation Score Inattention Disorganized Thinking 273 MD/EMS MD/CAM-ICU EMS/CAM-ICU 0.266 0.486 0.34 0.218 0.256 0.313 0.223 0.331 0.086 0.213 0.189 0.119 Mode of Arrival: The Effect of Age on EMS Use for Transportation to an Emergency Department Courtney Marie Cora Jones, Erin B. Philbrick, Manish N. Shah University of Rochester Medical Center, Rochester, NY Background: Previous studies have found older adults are more likely to use EMS than younger adults; however, these studies are limited because the effect of potentially important confounding variables, such as patient acuity and presence of comorbid medical conditions, were not evaluated. Objectives: This study aimed to assess the association between age and EMS use while controlling for potential confounders. We hypothesized that this association use would persist after controlling for confounders. Methods: A cross-sectional survey study was conducted at an academic medical center’s ED. An interview-based survey was administered and included questions regarding demographic and clinical characteristics, mode of ED arrival, health care use, and the perceived illness severity. Age was modeled as an ordinal variable (<60, 60–79, and ‡ 80 years). Bivariate analyses were used to identify potential confounders and effect measure modifiers and a multivariable logistic regression model was constructed. Odds ratios were calculated as measures of effect. Results: A total of 1092 subjects were enrolled and had usable data for all covariates, 465 (43%) of whom arrived via EMS. The median age of the sample was 60 years and 52% were female. There was a statistically significant linear trend in the proportion of subjects who arrived via EMS by age (p < 0.0001). Compared to adults aged less than 60 years, the unadjusted odds ratio associating age and EMS use was 1.41 (95% CI: 1.19, 1.67) for subjects 60–79 years and 1.98 (95% CI: 1.66, 2.36) for subjects ‡ 80 years. After adjustment for acuity, chief complaint, self-rated health, education, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S149 place of residence, comorbidities, history of depression, and perceived illness severity, age remained a statistically significant risk factor for with EMS use (1.36; 95% CI: 1.12, 1.65 for subjects 60–79 and 1.85; 95% CI: 1.25, 2.72 for subjects ‡ 80 years). Conclusion: In this sample of ED patients, the proportion of subjects arriving via EMS was found to increase linearly with age. In multivariable analysis, the relationship between increasing age and EMS use remained statistically significant and was not confounded by other variables. Additional research is needed to account for potential unmeasured confounders (e.g., insurance status and access to alternative transportation) and to elucidate reasons for the increased likelihood of EMS use among older adults. 274 Impact Of A New Senior Emergency Department On ED Recidivism, Rate Of Hospital Admission, And Hospital Length Of Stay Daniel Keyes, Ani Habicht, Bonita Singal, Mark Cowen St Joseph Mercy Ann Arbor/ University of Michigan EM Residency, Ann Arbor, MI Background: Senior (geriatric) EDs are opening with increasing frequency around the US, providing elders with more comprehensive screening for depression, polypharmacy, and other concerns, which may decrease the frequency of return visits (recidivism). Evaluation in the ED by social workers may also decrease admissions and allow earlier hospital discharge. Currently little information describes the effect on recidivism, rate of admission, or hospital length of stay for this important new phenomenon. Objectives: To investigate the effect of a Senior ED on recidivism, rate of hospital admission, and hospital length of stay, comparing a group before and after this intervention in a large suburban community hospital. Methods: We created a propensity variable for recidivism within 30 days for each of 66 DRG categories in the pre-Senior ED group, only. These DRG categories were then ranked by quintiles of propensity to return, and the rankings were applied to each patient in both cohorts. Multivariable analyses were used to balance any chance difference in risk between recidivism, admission rate, and LOS. The Cox proportional hazard model was used to test whether the intervention affected time to return within 30 days. Length of stay, propensity to return, and age were summarized using median and interquartile range (IQR). Triage level, sex, and insurance type were summarized using percentage. Results: A total of 8154 patients were included, 4252 in the pre-Senior ED group and 3902 in the Senior ED group. Subsequent exclusions were: 201 in the highest emergency severity index (ESI level 1), in 274 cases ESI was not determined, and one had an unrealistic recorded age of 130 years, leaving 7678, with 3936 in the pre-Senior ED group and 3742 in the Senior ED group. The mean (SD) age of the patients was 77.2 (8.3), ranging from 65 to 102, and 51% were female (see table). The risk of being admitted was lower in the Senior ED group (RR = 0.95, 95% CI: 0.92–0.98), but there was no significant difference on time to return within 180 days (HR = 1.08, 95% CI: 0.96–1.23) or time to return within 30 days (HR = 1.08, 95% CI: 0.95–1.23) or LOS (3.6 days for both, p = 0.85). Conclusion: A new senior ED resulted in a small decrease in rate of admission, but no significant effect on either recidivism or LOS was seen. Benefits of this ED may be apparent in other ways, including patient satisfaction, which will be valuable topics for future research. 275 Prospective Validation and Refinement of a Clinical Decision Rule for Chest Radiography in ED Patients with Chest Pain and Possible Acute Coronary Syndrome Joseph K. Poku1, Venkatesh R. Bellamkonda Athmaram1, Fernanda Bellolio1, Ronna L. Campbell1, David M. Nestler1, Ian G. Stiell2, Erik P. Hess1 1 Mayo Clinic, Rochester, MN; 2University of Ottawa, Ottawa, ON, Canada Background: We previously derived a clinical decision rule (CDR) for chest radiography (CXR) in patients with chest pain and possible acute coronary syndrome (ACS) consisting of the absence of three predictors: history of congestive heart failure, history of smoking, and abnormalities on lung auscultation. Objectives: To prospectively validate and refine a CDR for CXR in an independent patient population. Methods: We prospectively enrolled patients over 24 years of age with a primary complaint of chest pain S150 2012 SAEM ANNUAL MEETING ABSTRACTS and possible ACS from September 2009 to January 2010 at a tertiary care ED with 73,000 annual patient visits. Physicians completed standardized data collection forms before ordering chest radiographs and were thus blinded to CXR findings at the time of data collection. Two investigators, blinded to the predictor variables, independently classified CXRs as ‘‘normal,’’ ‘‘abnormal not requiring intervention,’’ and ‘‘abnormal requiring intervention’’ (e.g, heart failure, infiltrates) based on review of the radiology report and the medical record. Analyses included descriptive statistics, inter-rater reliability assessment (kappa), and recursive partitioning. Results: Of 1159 visits for possible ACS, mean age (SD) was 60.3 (15.6) and 51% were female. Twenty-four percent had a history of acute myocardial infarction, 10% congestive heart failure, and 11% atrial fibrillation. Seventy-one (6.1%, 95% CI 4.9–7.7) patients had a radiographic abnormality requiring intervention. The kappa statistic for CXR classification was 0.93 (95% CI 0.88– 0.97). The derived prediction rule (no history of congestive heart failure, no history of smoking, and no abnormalities on lung auscultation) was 67.6% sensitive (95% CI 56.1–77.3), 43.1% specific (95% CI 40.2–46.1), LR+ 1.19 (95% CI 1.00–1.41), LR- 0.75 (95% CI 0.53–1.06). The refined rule (no shortness of breath, no history of smoking, no abnormalities on lung auscultation, and age<55) was 93.0% sensitive (95% CI 84.6–97.0), 10.4% specific (95% CI 8.7–12.3), LR+ 1.04 (95% CI 0.97–1.11), LR- 0.68 (95% CI 0.29–1.61). Conclusion: The diagnostic accuracy of the refined CDR was substantially improved but is insufficient to recommend for use in clinical practice. Although the prevalence of CXR abnormalities in patients with chest pain and possible ACS is low, clinical data appear to inadequately predict abnormalities requiring intervention. 276 Cholesteryl Esters Associated with AcylCoA:cholesterol acyltransferase-2 Predict Coronary Artery Stenosis in Patients with Symptoms of Acute Coronary Syndrome Chadwick D. Miller, Michael J. Thomas, Brian Hiestand, Michael P. Samuel, Martha D. Wilson, Janet Sawyer, Lawrence L. Rudel Wake Forest Health Sciences, Winston-Salem, NC Background: After excluding MI in patients with symptoms of acute coronary syndrome (ACS), identify- ing the likelihood of coronary artery disease (CAD) could reduce the need for stress testing or coronary imaging. Acyl-CoA:cholesterol acyltransferase-2 (ACAT2) activity has been shown in monkey and murine models to correlate with atherosclerosis. Objectives: To determine if a novel cardiac biomarker consisting of plasma cholesteryl ester levels (CE) typically derived from the activity of ACAT2 is predictive of CAD in a clinical model. Methods: A single center prospective observational cohort design enrolled a convenience sample of subjects from a tertiary care center with symptoms of acute coronary syndrome undergoing coronary CT angiography or invasive angiography. Plasma samples were analyzed for CE composition with mass spectrometry. The primary endpoint was any CAD determined at angiography. Multivariable logistic regression analyses were used to estimate the relationship between the sum of the plasma concentrations from cholesteryl palmitoleate (16:1) and cholesteryl oleate (18:1) (defined as ACAT2-CE) and the presence of CAD. The added value of ACAT2-CE to the model was analyzed comparing the C-statistics and integrated discrimination improvement (IDI). Results: The study cohort was comprised of 113 participants enrolled over 24 months with a mean age 49 (±11.7) years, 59% with CAD at angiography. The median plasma concentration of ACAT2-CE was 938 lM (758, 1099) in patients with CAD and 824 lM (683, 998) in patients without CAD (p = 0.03) (Figure). When considered with age, sex, and the number of conventional CAD risk factors, ACAT2-CE were associated with a 6.5% increased odds of having CAD per 10 lM increase in concentration. The addition of ACAT2-CE significantly improved the C-statistic (0.89 vs 0.95, p = 0.0035) and IDI (0.15, p < 0.001) compared to the reduced model. In the subgroup of low-risk observation unit patients, the CE model had superior discrimination compared to the Diamond Forrester classification (IDI 0.403, p < 0.001). Conclusion: Plasma levels of ACAT2-CE, considered in a clinical model, have strong potential to predict a patient’s likelihood of having CAD. In turn, this could reduce the need for cardiac imaging after the exclusion of MI. Further study of ACAT2-CE as biomarkers in patients with suspected ACS is needed. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 277 The Role Of Bedside Carotid Ultrasonography In The Emergency Department To Risk Stratify Patients With Chest Pain Anita Datta1, Anjali Bharati1, Michelle Pearl1, Kenneth Perry1, Cristina Sison2, Sanjey Gupta1, Nidhi Garg1, Penelope Chun Lema1 1 New York Hospital Queens, Flushing, NY; 2 Feinstein Institute for Medical Research at the North Shore-LIJ Health System, Manhasset, NY Background: Outpatient studies have demonstrated a correlation between carotid intima-media thickness (CIMT) on ultrasound and coronary artery disease (CAD). There are no known published studies that investigate the role of CIMT in the ED using cardiac CT or percutaneous cardiac intervention (PCI) as a gold standard. Objectives: We hypothesized that CIMT can predict cardiovascular events and serve as a noninvasive tool in the ED. Methods: This was a prospective study of adult patients who presented to the ED and required evaluation for chest pain. The study location was an urban ED with a census of 120,000 annual visits and 24-hour cardiac catheterization. Patients who did not have CT or PCI or had carotid surgery were excluded from the study. Ultrasound CIMT measurements of right and left common carotid arteries were taken with a 10MHz linear transducer (Zonare, Mountain View, CA). Anterior, medial, and posterior views of the near and far wall were obtained (12 CIMT scores total). Images were analyzed by Carotid Analyzer 5 (Mailing Imaging Application LLC, Coralville, Iowa). Patients were classified into two groups based on the results from CT or PCI. A subject was classified as having significant CAD if there was over 70% occlusion or multi-vessel disease. Results: Ninety of 102 patients were included in the study; 55.7% were males. Mean age was 56.6 ± 13 years. There were 34 (37.8%) subjects with significant CAD and 56 (62.2%) with non-significant CAD. The mean of all 12 CIMT measurements was significantly higher in the CAD group than in the non-CAD group (0.60 ± 0.20 vs. 0.35 ± 0.23; p < 0.00001). A logistic regression analysis was carried out with significant CAD as the event of interest and the following explanatory variables in the model: sex, age group (‡55 yrs. vs. <55 yrs), DM, hypercholesterolemia, previous CAD, vascular disease, right mean IMTs, and left mean IMTs. The mean of all right IMT measurements and left IMT measurements correlated with significant CAD (p < 0.001). With each unit increase in CIMT value, subjects were 64 times more likely to have significant CAD (OR = 64.77, 95% CI 6.03, 696.06; P = 0.0006). For every 0.1 unit increment in the mean CIMT measurement, a subject • www.aemj.org S151 would be 1.52 times more likely to have significant CAD (64.77^0.1 = 1.52; 95% CI: 1.19,1.92). Conclusion: Increased thickness of the CIMT on ultrasound correlates with cardiovascular disease in the acute setting. 278 Are Echocardiography, Telemetry, Ambulatory Electrocardiography Monitoring, and Cardiac Enzymes in Emergency Department Patients Presenting with Syncope Useful Tests? Shamai A. Grossman1, Benjamin Sun2, David Chiu1, Nathan I. Shapiro1 1 Harvard Medical School, Beth Israel Deaconess Medical Center, Boston, MA; 2 Oregon Health and Science University, Portland, OR Background: Prior studies of admitted geriatric patients with syncope suggest diagnostic tests affect management <5% of the time; whether this is true among all ED patients with syncope remains unclear. Objectives: To determine the diagnostic yield of routine testing in-hospital or following ED discharge among patients presenting to an ED following syncope. Methods: A prospective, observational, cohort study of consecutive ED patients ‡18 years old presenting with syncope was conducted. The four most commonly utilized tests (echocardiography, telemetry, ambulatory electrocardiography monitoring, and cardiac markers) were studied. Interobserver agreement as to whether tests results determined the etiology of the syncope was measured using kappa (k) values. Results: Of 570 patients with syncope, 150 (26%) had echocardiography with 33 (6%) demonstrating a likely etiology of the syncopal event such as critical valvular disease or significantly depressed left ventricular function (k = 0.78). On hospitalization, 349 (61%) patients were placed on telemetry, 19 (3%) of these had worrisome dysrhythmias (k = 0.66). 317 (55%) patients had troponin levels drawn of whom 19 (3%) had positive results (k = 1); 56 (10%) patients were discharged with monitoring with significant findings in only 2 (0.4%) patients (k = 0.65). Overall, 73 (8%, 95% CI 7–10%) studies were diagnostic. Conclusion: Although routine testing is prevalent in ED patients with syncope, the diagnostic yield is relatively low. Nevertheless, some testing, particularly echocardiography, may yield critical findings in some cases. Current efforts to reduce the cost of medical care by eliminating non-diagnostic medical testing and increasing emphasis on practicing evidence-based medicine argue for more discriminate testing when evaluating syncope. (Originally submitted as a ‘‘late-breaker.’’) S152 279 2012 SAEM ANNUAL MEETING ABSTRACTS Needles In A Needlestack: ‘‘Prodromal’’ Symptoms of Unusual Fatigue and Insomnia Are Too Prevalent Among Adult Women Visiting the ED to be Useful in Diagnosing ACS Acutely Paris B. Lovett1, Yvonne N. Ezeala1, Rex G. Mathew1, Julia L. Moon2 1 Thomas Jefferson University, Philadelphia, PA; 2 Drexel University, School of Public Health, Philadelphia, PA Background: In 2003, McSweeney et al. reported surveys on ‘‘prodromal’’ symptoms recalled by women who had experienced myocardial infarctions (MI). Unusual fatigue was reported by 70.7% (severe 29.7%) and insomnia by 47.8% (severe 21.0%). These findings have led to risk management recommendations to consider these symptoms as predictive of acute coronary syndromes (ACS) among women visiting the ED. Objectives: To document the prevalence of these symptoms among all women visiting an ED. To analyze the potential effect of using these symptoms in the ED diagnostic process for ACS. Methods: A survey on fatigue and insomnia symptoms was administered to a convenience sample of all adult women visiting an urban academic ED (all arrival modes, acuity levels, all complaints). A sensitivity analysis was performed using published data and expert opinion for inputs. Results: We approached 548 women, with 379 enrollments. See table. The top box shows prevalences of prodromal symptoms among all adult female ED patients. The bottom box shows outputs from sensitivity analysis on the diagnostic effect of initiating an ACS workup for all female ED patients reporting prodromal symptoms. Conclusion: Prodromal symptoms of ACS are highly prevalent among all adult women visiting the ED in this study. This likely limits their utility in ED settings. While screening or admitting women with prodromal symptoms in the ED would probably increase sensitivity, that increase would be accompanied by a dramatic Table - Abstract 279: reduction in specificity. Such a reduction in specificity would translate to admitting, observing, or working up somewhere between 29% and 61% of all women visiting the ED, which is prohibitive in terms of personal costs, risks of hospitalization, and financial costs. While these symptoms may or may not have utility in other settings such as primary care, their prevalence, and the implied lack of specificity for ACS suggest they will not be clinically useful in the ED. 280 Length of Stay for Observation Unit Chest Pain Patients Tested with Coronary Computed Tomography Angiography Compared to Stress Testing Depends on Time of Emergency Department Presentation Simon A. Mahler, Brian C. Hiestand, James W. Hoekstra, David C. Goff, Chadwick D. Miller Wake Forest University Medical School, Winston-Salem, NC Background: Coronary computed tomography angiography (CCTA) has been identified as a way of reducing length of stay (LOS) relative to stress testing for emergency department (ED) and observation unit (OU) patients with chest pain. However, this relationship may differ based on the time of patient presentation to the ED. Objectives: To determine if the relationship between LOS and testing modality varies based on the time of ED patient presentation. Methods: We examined a cohort of low-risk chest pain patients evaluated in an ED-based OU using prospective and retrospective OU registry data elements. Cox proportional hazard modeling was performed to assess the effect of testing modality (stress testing vs. CCTA) on the LOS in the CDU. As CCTA is not available on weekends, only subjects presenting on weekdays were included. Cox models were stratified on time of patient presentation to the ED, based on four hour blocks beginning at midnight. The primary independent variable was first test modality, either stress imaging (exercise echo, dobutamine echo, stress MRI) or CCTA. Age, sex, and race were included as covariates. The proportional hazards assumption was tested using scaled Schoenfield residuals, and the models were graphically examined for outliers and overly influential covariate patterns. Test selection was a time varying covariate in the 8AM strata, and therefore the interaction with ln (LOS) was included as a correction term. After correction for multiple comparisons, an alpha of 0.01 was held to be significant. Results: Over the study period, 841 subjects (of 1,070 in the registry) presented on non-weekend days. The median LOS was 18.5 hours (IQR 12.4–23.3 hours), 57% were white, and 61% were female. The table shows the number of subjects in each time strata, the number tested, and the number undergoing stress testing vs. CCTA. After adjusting all models for age, race, and sex, the hazard ratio (HR) for LOS is as shown. Only those patients presenting between 8AM and noon noted a significant improvement in LOS with CCTA use (p < 0.0001). ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Conclusion: In OU management of low risk chest pain patients, in a setting of weekday/standard business hour CCTA availability, CCTA testing decreased LOS only in patients presenting to the ED during 8am– 11:59am time period. Table - Abstract 280: Time block Total n Tested n First test stress 00:00–03:59 04:00–07:59 08:00–11:59 12:00–15:59 16:00–19:59 20:00–23:59 81 74 197 223 152 113 77 67 186 206 142 110 57 34 84 149 99 73 HR (95% CI) 0.67 0.75 0.75 1.23 1.07 0.93 (0.38–1.16) (0.44–1.27) (0.68–0.84)* (0.90–1.69) (0.74–1.54) (0.49–1.39) * Time varying covariate, HR presented is for the interaction term of stress test*ln (LOS). 281 Correlation of Student OSCE Scores with Other Performance Metrics in an EM Clerkship: A Three Year Review Joshua Wallenstein, Douglas Ander, Connie Coralli Emory University, Atlanta, GA Background: A management-focused Objective Structured Clinical Examination (OSCE) became a component of our emergency medicine (EM) clerkship’s student performance evaluation in 2005. Since that time we have developed formal training protocols and standardization of our evaluators. There is limited literature exploring the validity of an OSCE as a measure of clinical skills in EM. Objectives: Determine the validity of a managementfocused EM OSCE as a measure of clinical skills by determining the correlation between OSCE scores and faculty assessment of student performance in the ED. Methods: Medical students in a fourth year EM clerkship were enrolled in the study. On the final day of the clerkship students participated in a five-station EM OSCE. Student performance on the OSCE was evaluated using a task-based evaluation system with 3–4 critical management tasks per case. Task performance was evaluated using a three-point system: performed correctly/timely (2), performed incorrectly/late (1), or not performed (0). Descriptive anchors were used for performance criteria. Communication skills were also graded on a three-point scale. Student performance in the ED was based on traditional faculty assessment using our core-competency evaluation instrument. A Pearson correlation coefficient was calculated for the relationship between OSCE score and ED performance score. Case item analysis included determination of difficulty and discrimination. Results: Between 2009–2011, 281 students completed the OSCE. Complete OSCE data were available for 242 students. A moderate positive correlation was found (r(237) = 0.400, p < 0.001), indicating a significant linear relationship between OSCE score and ED performance score. Difficulty index ranged from 21.9 to 97. Index of discrimination ranged from 34 to 118. • www.aemj.org S153 Conclusion: Our study of a management-focused EM OSCE revealed a moderate correlation between students’ OSCE scores and ED performance scores. This supports the use of the EM OSCE as a valid assessment of clinical skills. The presence of a moderate rather than a strong correlation may indicate the presence of confounding variables in both OSCE and ED performance evaluation. It may also indicate that the OSCE measures variables independent from traditional evaluations of clinical performance. Item analysis reveals specific case items that can be adjusted to refine the OSCE. 282 Rating The Emergency Medicine Core Competencies: Hawks, Doves, And The ACGME Jason T. Nomura, Hannah Bollinger, Debra Marco, James F. Reed, III Christiana Care Health System, Newark, DE Background: The ACGME has defined six core competencies (6CCs) of Patient Care, Medical Knowledge, Practice Based Learning, Interpersonal Communications, Professionalism, and System Based Practice. The ACGME also requires that trainees are evaluated on these 6CCs during their residency. Trainee evaluation in the 6CCs are frequently on a subjective rating scale. One of the recognized problems with a subjective scale is the rating stringency of the rater, commonly known as the Hawk-Dove effect. This has been seen in Standardized Clinical Exam scoring. Recent data have shown that score variance can be related to evaluator performance with a negative correlation. Higher-scoring physicians were more likely to be a stringent or Hawk type rater on the same evaluation. It is unclear if this pattern also occurs in the subjective ratings that are commonly used in assessments of the 6CCs. Objectives: Comparison of attending physician scores on the ACGME 6CCs with attending ratings of residents for a negative correlation or Hawk-Dove effect. Methods: Residents are routinely evaluated on the 6CCs with a 1–9 numerical rating scale as part of their training. The evaluation database was retrospectively reviewed. Residents anonymously scored attending physicians on the 6CCs with a cross-sectional survey that utilized the same rating scale, anchors, and prompts as the resident evaluations. Average scores for and by each attending were calculated and a Pearson Correlation calculated by core competency and overall. Results: In this IRB-approved study, a total of 43 attending physicians were scored on the 6CCs with 447 evaluations by residents. Attendings evaluated 162 residents with a total of 1,678 evaluations completed over a 5-year period. Attending mode score was 9 ranging from 2 to 9; resident scores had a mode of 8 with a range of 1 to 9. There was no correlation between the rated performance of the attendings overall or in each 6CCs and the scores they gave (p = 0.065–0.861). Conclusion: Hawk-Dove effects can be seen in some scoring systems and has the potential to affect trainee evaluation on the ACGME core competencies. However, a negative correlation to support a Hawk-Dove S154 2012 SAEM ANNUAL MEETING ABSTRACTS scoring pattern was not found in EM resident evaluations by attending physicians. This study is limited by being a single center study and utilizing grouped data to preserve resident anonymity. 283 Emergency Medicine Resident Self-Assessment of Competency During Training and Beyond Jeremy Voros1, Rita Cydulka2, Debra Perina3, John Moorhead4 1 Denver Health Emergency Medicine Residency, Denver, CO; 2MetroHealth Medical Center, Cleveland, OH; 3University of Virginia Health Sciences Center, Charlottesville, VA; 4 Oregon Health & Science University, Portland, OR Background: All ACGME-accredited residency programs are required to provide competency-based education and evaluation. Graduating residents must demonstrate competency in six key areas. Multiple studies have outlined strategies for evaluating competency, but data regarding residents’ self-assessments of these competencies as they progress through training and beyond is scarce. Objectives: Using data from longitudinal surveys by the American Board of Emergency Medicine, the primary objective of this study was to evaluate if resident self-assessments of performance in required competencies improve over the course of graduate medical training and in the years following. Additionally, resident self-assessment of competency in academic medicine was also analyzed. Methods: This is a secondary data analysis of data gathered from two rounds of the ABEM Longitudinal Study of Emergency Medicine Residents (1996–98 and 2001–03) and three rounds of the ABEM Longitudinal Study of Emergency Physicians (1999, 2004, 2009). In both surveys, physicians were asked to rate a list of 18 items in response to the question, ‘‘What is your current level of competence in each of the following aspects of work in EM?’’ The rated items were grouped according to the ACGME required competencies of Patient Care, Medical Knowledge, Practice-based Learning and Improvement, Interpersonal and Communication Skills, and System-based Practice. An additional category for academic medicine was also added. Results: Rankings improved in all categories during residency training. Rankings in three of the six categories improved from the weak end of the scale to the strong end of the scale. There is a consistent decline in rankings one year after graduation from residency. The greatest drop is in Medical Knowledge. Mean self-ranking in academic medicine competency is uniformly the lowest ranked category for each year. Conclusion: While self-assessment is of uncertain value as an objective assessment, these increasing rankings suggest that emergency medicine residency programs are successful at improving residents’ confidence in the required areas. Residents do not feel as confident about academic medicine as they do about the ACGME required competencies. The uniform decline in rankings the first year after residency is an area worthy of further inquiry. Screening Medical Student Rotators From Outside Institutions Improves Overall Rotation Performance Shaneen Doctor, Troy Madsen, Susan Stroud, Megan L. Fix University of Utah, Salt Lake City, UT 284 Background: Emergency medicine is a rapidly growing field. Many student rotations are limited in their ability to accommodate all students and must limit the number of students they allow per rotation. We hypothesize that pre-screening visiting student rotators will improve overall student performance. Objectives: To assess the effect of applicant screening on overall rotation grade and mean end of shift card scores. Methods: We initiated a medical student screening process for all visiting students applying to our 4-week elective EM rotation starting in 2008. This consisted of reviewing board scores and requiring a letter of intent. Students from our home institution were not screened. All end-of-shift evaluation cards and final rotation grades (honors, high pass, pass, fail) from 2004 to 2011 were analyzed. We identified two cohorts: home students (control) and visiting students. We compared pre-intervention (2004–2008) and postintervention (2008–2011) scores and grades. End of shift performance scores are recorded using a fivepoint scale that assesses indicators such as fund of knowledge, judgment, and follow-through to disposition. Mean ranks were compared and P-values were calculated using the Armitage test of trend and confirmed using t-tests. Results: We identified 162 visiting students (91 pre, 81 post) and 160 home students (90 pre, 80 post). 12 (13.2%) visiting students achieved honors pre-intervention while 31 (38.3%) achieved honors post-intervention (p = 0.000093). No significant difference was seen in home student grades: 28 (31.1%) received honors pre2008 and 17 (21.3%) received honors post-2008 Table - Abstract 283: Mean Rankings by Category (reported with Mean CI) EM1 PC MK PBL ICS SBP AM 2.81 (2.71–2.91) 3.19 (3.11–3.27) 2.92 (2.85–2.99) 3.56 (3.49–3.64) 2.47 (2.40–2.53) 2.20(2.14–2.26) EM2 3.19 3.58 3.41 3.72 2.83 2.48 (3.04–3.34) (3.45–3.70) (3.32–3.51) (3.62–3.83) (2.73–2.92) (2.39–2.57) EM3 3.44 3.90 3.62 3.88 3.21 2.76 (3.34–3.54) (3.82–3.99) (3.56–3.68) (3.81–3.94) (3.15–3.28) (2.69–2.83) PR1 3.37 3.33 3.43 3.79 3.01 2.47 (3.26–3.48) (3.23–3.44) (3.37–3.50) (3.72–3.88) (2.95–3.08) (2.40–2.54) PR5 PR10 3.44 (3.34–3.54) 3.56(3.46–3.66) 3.47 (3.39–3.55) 3.80 (3.72–3.88) 3.26 (3.19–3.33) 2.27 (2.19–2.36) 3.43 (3.27–3.59) 3.80(3.64–3.96) 3.47 (3.35–3.59) 3.79 (3.66–3.91) 3.32 (3.21–3.44) 2.37 (2.24–2.50) ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S155 (p = 0.17). Mean numerical grade comparison confirmed these results with pre and post visiting student means 4.13 and 4.27 (p = 0.00038) and home student means 4.19 and 4.13 (p = 0.17), respectively. Mean shift performance scores for visiting students was 3.99 pre and 4.18 post (p < 0.0001), while mean scores for home students was 4.00 and 4.06, respectively (p = 0.016). Conclusion: We found that implementation of a screening process for visiting medical students improved overall rotation scores and grades as compared to home students who did not receive screening. Screening rotating students may improve the overall quality of applicants and thereby the residency program. median score was 82.0. The average rbp was 0.226 (range 0.021–0.385). Conclusion: The National EM M4 Examination is a 50-question multiple choice high-stakes examination developed to help fill a void in the assessment of EM students. Initial data demonstrate acceptable mean and median scores as well as acceptable rpbs. Creation of a National Emergency Medicine Fourth-year Medical Student Examination Emily L. Senecal1, Corey Heitz2, Michael Beeson3 1 Massachusetts General Hospital - Harvard Medical School, Boston, MA; 2Carilion Clinic Virginia Tech Carilion School of Medicine, Roanoke, VA; 3Akron General Hospital, Akron, OH Background: There are many descriptions in the literature of computer-assisted instruction in medical education, but few studies that compare them to traditional teaching methods. Objectives: We sought to compare the suturing skills and confidence of students receiving video preparation before a suturing workshop versus a traditional instructional lecture. Methods: 88 first and second year medical students were randomized into two groups. The control group was given a lecture followed by 40 minutes of suturing time. The video group was provided with an online suturing video at home, no lecture, and given 40 minutes of suturing time during the workshop. Both groups were asked to rate their confidence before and after the workshop, and their belief in the workshop’s effectiveness. Each student was also videotaped suturing a pig’s foot after the workshop and graded on a previously validated 16-point suturing checklist. 83 videos were scored. Results: There was no significant difference between the test scores of the lecture group (M = 11.21, SD = 3.17, N = 42) and the video group (M = 11.27, SD = 2.53, N = 41) using the two-sample independent ttest for equal variances (t(81) = )0.09, p = 0.93). There was a statistically significant difference in the proportion of students scoring correctly for only one point: ‘‘Curvature of needle followed’’: 25/42 in the lecture group and 35/41 in the video group (chi = 6.92, df = 1, p = 0.008). Students in the video group were found to be 2.45 times more likely to have a neutral or favorable feeling of suturing confidence before the workshop (p = 0.067, CI 0.94–6.4) using a proportional odds model. No association was detected between group assignment and level of suturing confidence after the workshop (p = 0.475). There was also no association detected between group assignment and opinion of the suturing workshop (p = 0.681) using a logistic regression odds model. Among those students who indicated a lack of confidence before training, there was no detected association (p = 0.967) between group assignment and having an improved confidence using a logistic regression odds model. Conclusion: Students in the video group and students in the control group achieved similar levels of suturing skill and confidence, and equal belief in the workshop’s effectiveness. This study suggests that video instruction could be a reasonable substitute for lectures in procedural education. 285 Background: A National Board of Medical Examiners (NBME) subject examination does not exist for EM students. To fill this void, the Clerkship Directors in Emergency Medicine (CDEM) tasked a committee with development of an end-of-rotation examination geared towards fourth-year (M4) EM students, based on a national syllabus, and consisting of examination questions written according to published question writing guidelines. Objectives: To describe the development of the examination and provide data concerning its initial usage and performance. Methods: Exam Development: The CDEM Testing Committee systematically reviewed an existing EM, student-focused question database at www.saemtests.org. Question assessment included statistical performance analysis, compliance with published item writing guidelines, and topic inclusion within the published EM M4 syllabus. For syllabus topics without existing questions, committee members wrote new items. Committee members developed a system for secure online examination administration. Data Analysis: LXR 6.0 testing software was used to administer the examination. Data gathered included numbers of examinations completed and institutions participating, mean and median scores with standard deviation, and point biserial correlation (rpb) for each question. Results: Thirty-six questions meeting the stated criteria were selected for inclusion in the National EM M4 Examination. An additional fourteen questions were written by committee members to generate a 50-question examination. The National EM M4 Examination was released online on August 1, 2011. Three months into its availability, the examination had been completed 703 times by students from 22 participating clerkships. The mean score was 80.2 (SD 4.0) and the 286 A Novel Approach To ‘‘See One, Do One’’: Video Instruction For Suturing Workshops Amita Sudhir, Claire M. Plautz, William A. Woods University of Virginia, Charlottesville, VA S156 287 2012 SAEM ANNUAL MEETING ABSTRACTS Educational Technology Can Improve ECG Diagnosis of ST Elevation MI Among Medical Students Ali Pourmand, Steven Davis, Kabir Yadav, Hamid Shokoohi, Mary Tanski George Washington University, Washington, DC Background: Accurate interpretation of the ECG in the emergency department is not only clinically important but also critical to assess medical knowledge competency. With limitations to expansion of formal didactics, educational technology offers an innovative approach to improve the quality of medical education. Objectives: The aim of this study was to assess an online multimedia-based ECG training module evaluating ST elevation myocardial infarction (STEMI) identification among medical students. Methods: A convenience sample of fifty-two medical students on their EM rotations at an academic medical center with an EM residency program was evaluated in a before-after fashion during a 6-month period. One cardiologist and two ED attending physicians independently validated a standardized exam of ten ECGs: four were normal ECGs, three were classic STEMIs, and three were subtle STEMIs. The gold standard for diagnosis was confirmed acute coronary thrombus during cardiac catheterization. After evaluating the 10 ECGs, students completed a pre-intervention test wherein they were asked to identify patients who required emergent cardiac catheterization based on the presence or absence of ST segment elevation on ECG. Students then completed an online interactive multimedia module containing 13 minutes of STEMI training based on American Heart Association/American College of Cardiology guidelines on STEMI. Medical students were asked to complete a post-test of the 10 ECGs after watching online multimedia. Results: The participants included 52 medical students in their fourth year of training, of whom 27 (52%) were female. Overall, 36 (69%) had an improvement in their score, with a pre-test median of 5 [IQR 5–6] out of 10 correct ECG interpretations and a post-test median 7 [IQR 6–8]. Treating data as ordinal, Wilcoxon signed-rank test confirmed post-test was significantly different (p < 0.001). Spearman’s rank correlation and regression analysis confirmed that sex did not influence test scores. Conclusion: Educational technology using online multimedia modules improves medical students’ ability to interpret ECGs and can augment traditional teaching due to its ready availability and interactiveness. Future studies using standardized assessment tools are needed to validate improved competency in ECG interpretation in broader populations. 288 Better Than Expected: External Validation of the PECARN Head Injury Criteria in a Community Hospital Setting Payal Shah1, David Donaldson1, Shawn Munafo1, Teresa Thomas1, Robert Swor2, William Anderson1, Aveh Bastani1 1 Troy Beaumont Hospital, Troy, MI; 2William Beaumont Hospital, Royal Oak, MI Background: The Pediatric Emergency Care Applied Research Network (PECARN) head injury criterion provides an algorithm to emergency physicians for identifying patients at very low risk for clinically important traumatic brain injury (ciTBI). The PECARN investigators noted that by implementing their algorithm, head CTs could potentially be avoided for 25% of pre-verbal pediatric patients (age < 2). Objectives: Our objective was to quantify the number of pre-verbal pediatric head CTs performed at our community hospital that could have been avoided by utilizing the PECARN criteria. Methods: We conducted a standardized chart review of all children under the age of 2 who presented to our community hospital and received a head CT between Jan 1st, 2010 and Dec 31st, 2010. Following recommended guidelines for conducting a chart review, ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 we: 1) utilized four blinded chart reviewers, 2) provided specific training, 3) created a standardized data extraction tool, and 4) held periodic meetings to evaluate coding discrepancies. Our primary outcome measure was the number of patients who were PECARN negative and received a head CT at our institution. Our secondary outcome was to reevaluate the sensitivity and specificity of the PECARN criteria to detect ciTBI in our cohort. Data were analyzed using descriptive statistics and 95% confidence intervals were calculated around proportions using the modified Wald method. Results: A total of 138 patients under the age of 2 received a head CT at our institution during the study period. 23 patients were excluded from the final analysis because their head CTs were not for trauma. The prevalence of a ciTBI in our cohort was 2.6% (95% CI 0.6%– 7.7%) (Table). Of the three patients with ciTBI, all were identified utilizing the PECARN criteria. A total of 74 patients (64.3%, 95% CI 55.3%–72.5%) were identified as not requiring a head CT based on PECARN criteria. At our community hospital the PECARN criteria retained its high sensitivity of 100% (95% CI 31%–100%) and modest specificity of 66.1% (95% CI 56.4%–74.6%). Conclusion: Application of the PECARN criteria in the community hospital setting can significantly decrease the number of head CTs performed on pre-verbal pediatric patients while retaining its high sensitivity for ruling out ciTBI. Table - Abstract 288: (+) PECARN (-) PECARN Total 289 (+) ciTBI (-) ciTBI Total 3 0 3 38 74 112 41 74 115 Prevalence of Non-traumatic Incidental Findings Found on Pediatric Cranial CT scans Alexander Rogers1, Cormac O. Maher1, Jeff E. Schunk2, Kimberly S. Quayle3, Elizabeth S. Jacob4, Richard Lichenstein5, Elizabeth C. Powell6, Michelle Miskin2, Peter S. Dayan7, James F. Holmes8, Nathan Kuppermann8, the PECARN9 1 University of Michigan, Ann Arbor, MI; 2 University of Utah, Salt Lake City, UT; 3 Washington University School of Medicine, St. Louis, MO; 4Brown University, Providence, RI; 5 University of Maryland School of Mediine, Baltimore, MD; 6Northwestern University, Chicago, IL; 7Columbia University Medical Center, New York, NY; 8UC Davis School of • www.aemj.org S157 Medicine, Sacramento, CA; MD 9 HRSA, Rockville, Background: Blunt head trauma is a leading cause of emergency department (ED) visits in children. Cranial CT scans are often used to detect intracranial injuries, but are also sensitive for unexpected non-traumatic findings that may be inconsequential or clinically important. Objectives: To describe the prevalence and urgency of incidental findings on CT in children evaluated in a large, multi-center observational study of children with blunt head trauma. Methods: This was a planned secondary analysis of a study of children with blunt head trauma. Radiologist CT reports were reviewed. Reports were abstracted and categorized into three groups of clinical urgency based on an a priori list of diagnoses: Category 1) requires prompt evaluation/intervention, 2) requires timely outpatient follow-up, 3) no specific follow-up needed. Consensus was achieved on diagnosis and urgency by a pediatric emergency medicine physician and pediatric neurosurgeon. Results: 43,904 patients were enrolled in the parent study, of whom 16,137 (37%) received a cranial CT scan in the ED. We identified incidental, non-traumatic findings on 830/16,137 (5%) CT scans. Of these, 125/830 (15%) were reclassified as normal variants. Of the remaining 705, the most common non-traumatic findings were: ventricular abnormality (12%), arachnoid cyst (10%), and increased extra-axial fluid (10%). Four percent were category 1, with hydrocephalus most common. Category 3 represented 70%, and category 2 represented 26%. Conclusion: Incidental non-traumatic findings on cranial CT are present in 5% of children evaluated in the ED for blunt head trauma. Most do not require specific intervention. However, 30% (category 1 + 2) warranted either prompt or timely subspecialty consultation. ED providers should be alert to the possibility of incidental findings on cranial CT obtained following trauma, and engage mechanisms to ensure appropriate follow-up. Funded by HRSA/MCHB (R40MC02461-01-00). 290 Evidence of Axonal Injury for Children with Mild Traumatic Brain Injuries Lynn Babcock1, Weihong Yuan1, Nicole McClanahan1, Yingying Wang1, Jeffrey Bazarian2, Shari Wade1 1 Cincinnati Children’s Hospital Medical Center, Cincinnati, OH; 2University of Rochester, Rochester, NY Background: Diffuse axonal injury (DAI) is hypothesized to be the underlying neuropathology in mild traumatic brain injury (mTBI). Diffusion tensor imaging Table - Abstract 289: Findings by Urgency Category Category 1 Hydrocephalus Tumor/Mass AVM Cysticercosis N = 25 12 11 1 1 (46%) (42%) (4%) (4%) Category 2 Arachnoid Cyst Extra-Axial Fluid Chiari Other N = 185 71 67 16 31 (38%) (36%) (9%) (17%) Category 3 Ventricular abnormality Sinus Cyst Skull Abnormality Other N = 495 83 (17%) 65 (13%) 53 (11%) 294 (59%) S158 2012 SAEM ANNUAL MEETING ABSTRACTS (DTI) measures disruption of axonal integrity on the basis of anisotropic diffusion properties. Findings on DTI may relate to the injury, as well as the severity of postconcussion syndrome (PCS) following mTBI. Objectives: To examine acute anisotropic diffusion properties based on DTI in youth with mTBI relative to orthopedic controls and to examine associations between white matter (WM) integrity and PCS symptoms. Methods: Interim analysis of a prospective casecontrol cohort involving 12 youth ages 11–16 years with mTBI and 10 orthopedic controls requiring extremity radiographs. Data collected in ED included demographics, clinical information, and PCS symptoms measured by the postconcussion symptom scale. Within 72 hours of injury, symptoms were re-assessed and a 61-direction, diffusion weighted, spin-echo imaging scan was performed on a 3T Philips scanner. DTI images were analyzed using tract-based spatial statistics. Fractional anisotropy (FA), mean diffusivity (MD), axial diffusivity (AD), and radial diffusivity were measured. Results: There were no group demographic differences between mTBI cases and controls. Presenting symptoms within the mTBI group included GCS = 15 83%, loss of consciousness 33%, amnesia 33%, post-traumatic seizure 8%, headache 83%, vomiting 33%, dizziness 42%, and confusion 42%. PCS symptoms were greater in mTBI cases than in the controls at ED visit (30.1 ± 17.0 vs. 15.5 ± 16.8, p < 0.06) and at the time of scan (19.1 ± 12.9 vs. 5.7 ± 6.5, p < 0.01). The mTBI group displayed decreased FA in cerebellum and increased MD and AD in the cerebral WM relative to controls (uncorrected p < 0.05). Increased FA in cerebral WM was also observed in mTBI patients but the group difference was not significant. PCS symptoms at the time of the scan were positively correlated with FA and inversely correlated with RD in extensive cerebral WM areas (p < 0.05, uncorrected). In addition, PCS symptoms in mTBI patients were also found to be inversely correlated with MD, AD, and RD in cerebellum (p < 0.05). Conclusion: DTI detected axonal damage in youth with mTBI which correlated with PCS symptoms. DTI performed acutely after injury may augment detection of injury and help prediction of those with worse outcomes. Objectives: The aim of this study was to evaluate parental knowledge of concussion in young children who participate in recreational tackle football. Methods: Parents/legal guardians of children aged 5– 15 years enrolled in recreational tackle football were asked to complete an anonymous questionnaire based on the CDC’s Heads Up: Concussion In Youth Sports quiz. Parents were asked about their level of agreement in regard to statements that represent definition, symptoms, and treatment of concussion. Results: A total of 310 out of 369 parents voluntarily completed the questionnaire (84% response rate). Parent and child demographics are listed in Table 1. Ninety four percent of parents believed their child had never suffered a concussion. However, when asked to agree or disagree with statements addressing various aspects of concussion, only 13% (n = 41) could correctly identify all seven statements. Most did not identify that a concussion is considered a mild traumatic brain injury and can be achieved from something other than a direct blow to the head. Race, sex, and zip code had no significant association with correctly answering statements. Education (0.24; p < 0.01) and number of years the child played (0.11; p < 0.05) had a small effect. Fifty-three percent of parents reported someone had discussed the definition of concussion with them and 58% the symptoms of concussion. See Table 2 for source of information to parents. No parent was able to classify all symptoms listed as correctly related or not related to concussion. However, identification of correct concussion definitions correlated with identification of correct symptoms (0.25; p < 0.05). Conclusion: While most parents had received some education regarding concussion from a health care provider, important misconceptions remain among parents of young athletes regarding the definition, symptoms, and treatment of concussion. This study highlights the need for health care providers to increase educational efforts among parents of young athletes in regard to concussion. Table 1 - Abstract 291: Demographics Parent Demographics Race 291 Knowledge Assessment of Sport-related Concussion among Parents of Children Aged 5–15 years Enrolled in Recreational Tackle Football Carol Mannings, Colleen Kalynych, Madeline Joseph, Carmen Smotherman, Dale Kraemer University of Florida, Jacksonville, FL Background: Sports-related concussion among professional, collegiate, and more recently high school athletes has received much attention from the media and medical community. To our knowledge, there is a paucity of research in regard to sports-related concussion in younger athletes. Child Demographics N (%) White Black of African American Other Education Level 133 (43) 148 (48) High school or less 68 (22) Some college College graduate 117 (38) 124 (40) 25 (8) N (%) Age (years) N (%) 5–7 8–10 45 (15) 156 (50) 11–15 Years child has played This is the first year 1–2 3–4 5 or more 109 (35) N (%) 103 (33) 85 (28) 75 (24) 46 (15) ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Table 2 - Abstract 291: Concussion-related information was discussed with me by someone Definition N = 165 (53%) Symptoms (58%) N (%) N = 179 38 (23) 93 (56) List of symptoms, only some of which are consistent with concussion Headache, nausea, difficulty remembering things, ringing in the ears, concentration difficulty, increased irritability or emotional outburst, disorientation, sleeping difficulty, difficulty urinating, weakness in arms or legs, hearing voices, numbness in the arms or legs 292 Coach Doctor/health care provider Athletic trainer Friends/relative 293 40 (22) 95 (53) 27 (15) 28 (16) Young Children Have Difficulty Using the Wong Baker Pain Faces Scale in an ED Setting Gregory Garra, Anna Domingo, Henry C. Thode Jr, Adam J. Singer Stony Brook University, Stony Brook, NY Background: Pain is a multidimensional experience. Fear and anxiety may bias pain reporting and interfere with measuring pain. Few studies on pain severity reporting in children have evaluated discriminate validity. Objectives: The objective of our study was to determine discriminate validity of the Wong Baker FACES Pain Rating Scale (WBS). We hypothesized that children (3–6 years) would be able to appropriately use the WBS for reporting pain. Methods: Study Design: Prospective observational. Setting: University-based, suburban pediatric ED. Subjects: Patients age 3–6 years. Measures: Research assistants recorded demographic variables along with presence of pain (yes/no) and pain severity rating using the six-category WBS. Patients also completed a 26item Likert questionnaire assessing fears in medical settings: the Child Medical Fear Scale (CMFS), minimum score 26, maximum 78. Analysis: Descriptive statistics. Pearson’s correlation was used to measure agreement between dichotomous and ordinal/categorical data. Spearman’s correlations were used to measure agreement between WBS and CMFS. Results: One hundred and thirty four children were enrolled; 118 with pain, 16 without. There was an even distribution of WBS ratings. Five children with no pain utilized face categories reflective of pain. Likewise, 23 children (20%) with reported pain provided assessments using the ‘‘no pain’’ face. The correlation between pain presence and WBS reporting was poor (r = 0.38; 95%CI 0.23 to 0.52). Only 50% were able to complete the CMFS. The median CMFS score was 41 (IQR 34–48). Correlation between pain presence and CMFS score was moderate (r = 0.54; 95%CI 0.35 to 0.69) although correlation between CMFS and WBS was poor (q = 0.27). Conclusion: Young children appear to have difficulty reporting pain on the WBS. However, young children www.aemj.org S159 do not appear to mistake fear for pain presence or misuse face categories for fear severity. N (%) Coach Doctor/health care provider Athletic trainer Friends/relative 23 (14) 18 (11) • Timeliness and Effectiveness of Intranasal Fentanyl Administration for Children Ryan J. Scheper, Amy L. Drendel, Marc H. Gorelick, Martha W. Stevens, Steven J. Weisman, Raymond J. Hoffmann, Kim Gejewski Medical College of Wisconsin, Milwaukee, WI Background: Inadequate treatment of children with painful conditions is a significant problem in the emergency department (ED). Intranasal (IN) fentanyl is a painless, rapid means to administer analgesia for moderate to severe pain that may be particularly useful for the pediatric patient. Objectives: Determine if IN fentanyl 1) improves the timeliness of analgesic administration for children with fractures and 2) improves the treatment of pain for children compared to usual treatment in the ED prior to its utilization. Methods: A retrospective observational study, before and after the introduction of IN fentanyl on September 1, 2009, was performed. All children ages 0–18 years with a fracture and an Emergency Severity Index triage level 2 were identified. A random sample of children evaluated during a 15-month period prior to initiation of IN fentanyl (Group A) was compared to a consecutive sample of children administered IN fentanyl during the first 15 months of IN fentanyl use (Group B). A chart review identified demographic information, ED analgesic used, time to analgesic, and pain scores. Mann-Whitney and chi-square tests were used to test for significant differences. Results: 255 children (Group A) prior to, and 121 children (Group B) after initiation of IN fentanyl were enrolled. Most children in Group A received morphine (87.9%), oxycodone with acetaminophen (5.2%), or ibuprofen (3.6%). There was no difference in age, race/ ethnicity, sex, or initial pain score between the two groups. Median time to analgesic administration was 28.2 minutes for Group A, compared to 15.0 minutes for Group B (p < 0.0001). The median decrease in pain score on a 10-point scale was 2 points for Group A, compared to 3 points for Group B (p = 0.04). One hour after receiving an analgesic, only 60.83% of children in Group A achieved a clinically relevant (2-point) improvement in pain score, compared to 84.93% in Group B (p = 0.0006). Conclusion: IN fentanyl was associated with a more timely analgesic administration and more effective pain relief for children with fractures compared to usual care in the ED. 294 National Study of Emergency DepartmentAssociated Deaths Elaine Reno, Adit Ginde University of Colorado, Denver, CO Background: The ED centers on the stabilization and resuscitation of the acutely ill and injured patients. S160 2012 SAEM ANNUAL MEETING ABSTRACTS Deaths during and immediately following ED visits are becoming increasingly common, and emergency physicians must be adept at identifying patients at risk for death. Objectives: To characterize the incidence of ED-associated deaths nationally and to compare characteristics of patients who die during or immediately following ED visits (within 3 days of admission), to those that are admitted to the hospital and survive to discharge. Methods: We analyzed adult participants of the 2006– 2009 National Hospital Ambulatory Medical Care Survey, an annual, nationally representative sample of all U.S. ED visits. We grouped deaths as those who were dead on arrival, died during the ED visit, and admitted to the hospital and died prior to hospital discharge. We analyzed the survey-weighted data to generate nationally representative estimates. Results: Of the 124 million annual ED visits, 110,000 (95%CI, 78,000–140,000) patients/year were dead on arrival, 140,000 (95%CI, 110,000–180,000) patients/year died in the ED, and 350,000 (95%CI, 300,000–420,000) patients/year died during hospitalization after an ED visit. Of the 130,000 patients/year who died within 3 days of hospital admission, the majority (120,000) were admitted to a critical care area. Patients who died in the ED, compared to those who died during hospital stay, were more likely to have been hypothermic (28% vs. 17%), bradycardic (29% vs. 6%), and hypotensive (44% vs. 11%) on ED vitals signs. Of note, for patients who died during their ED stays, 71% arrived normothermic (36–37.9C), 20% had a normal heart rate (60–89 beats/minute), 37% had a normal range sBP (90–159 mmHg), 39% had a normal respiratory rate (12–20/minute), and 45% had a normal oxygen saturation (‡93%). Of note, only 73% of patients who died during their ED visit were triaged as category 1 (immediate). Conclusion: Despite the large number of patients, emergency physicians identified most patients that died within 3 days of admission and triaged them to ICU care. Many patients who died during the ED visit had lower triage acuity and were without severe vital sign abnormality at presentation; this finding was magnified in those that were admitted and died during hospitalization. 295 Post-intubation Care In Mechanically Ventilated Patients Boarding In The Emergency Department Rahul Bhat1, Munish Goyal1, Anu Bhooshan1, Bill Frohna1, Jeff Dubin1, Daniel Moy1, Vikaas Kataria1, Linn Lung1, Mihrye Mete2, David Gaieski3 1 Georgetown University Hospital/Washington Hospital Center, Washington, DC; 2Medstar Health Research Institute, Washington, DC; 3 University of Pennsylvania, Philadelphia, PA Background: Emergency department (ED) crowding has led to increased length of stay for intubated patients, leaving early post-intubation care to the emergency physician. The quality of post-intubation care in the ED has not been extensively studied. We sought to determine how frequently common post-intubation interventions are performed while patients board in the ED. Objectives: We hypothesized that few intubated patients receive all interventions while boarding in the ED. Methods: This is a retrospective chart review of all patients intubated in the Washington Hospital Center ED between 11/14/09 to 6/1/11. Patients were excluded if they were in the ED for <2 hours post-intubation, had emergent surgery within six hours post-intubation, were managed by the trauma team, had incomplete data, or had their status changed to ‘‘do not resuscitate’’(DNR) during their admission. Trained research assistants blinded to the objectives reviewed each chart to determine if the following interventions were performed in the ED post-intubation: appropriate initial tidal volume (6–10 cc/kg ideal body weight), sedation given <30 minutes of intubation, oro/nasogastric tube (OGT) placement, chest x-ray (CXR), arterial blood gas (ABG) sampling, and use of continuous endtidal capnography (ETCO2). Additionally, ventilator duration, ICU length of stay, mortality, and development of ventilator-associated pneumonia (VAP) were recorded. Results: 622 charts were reviewed, with 453 charts excluded (158 in the ED for <2 hours post-intubation, 143 trauma patients, 112 DNR, 40 other reasons), leaving 169 patients in our cohort. Of the six post-intubation interventions, CXR was obtained most frequently (96.4%), followed by OGT placement (84.0%), sedation (83.4%), ABG sampling (76.9%), appropriate tidal volume (71.0%), and least frequently ETCO2 monitoring (8.3%). The percentage of patients receiving all interventions was 2.4%; 5 of 6 interventions was 42.0%. Mean duration of ventilation and ICU length of stay were 3.4 days, and 5.2 days respectively with a VAP rate of 5.9% and a mortality rate of 8.9%. Conclusion: In this single center study, few patients received all six measured post-intubation interventions while boarding in the ED. 296 Assessing Vitamin D Status In Sepsis And Association With Systemic Inflammation Justin D. Salciccioli1, Adit A. Ginde2, Amanda Graver1, Tyler Giberson1, Cristal Cristia1, Michael N. Cochhi1, Michael W. Donnino1 1 BIDMC Center for Resuscitation Science, Boston, MA; 2University of Colorado School of Medicine, Denver, CO Background: Recent evidence has suggested a potential role of vitamin D in modulating innate immune function. However, the relationship between vitamin D and systemic inflammation has yet to be assessed in human sepsis. Objectives: To determine vitamin D status in critically ill patients with sepsis and the association between 25(OH)D level and systemic inflammation. Methods: We performed a prospective study of patients with sepsis at an urban tertiary care hospital ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S161 markers of inflammation. Whether vitamin D supplementation can attenuate systemic inflammation in critically ill patients with sepsis should be assessed in future investigations 297 during the period from 10/2010 to 05/2011. Inclusion criteria: 1. Adult (>18 years); 2. Presence of two or more SIRS criteria; 3. Admission to ICU with sepsis. Patients with a life-expectancy <24 hrs from arrival were excluded. Patient demographics, co-morbid conditions, vital signs and laboratory data, and in-hospital mortality were recorded. Serum 25(OH)D levels and inflammatory cytokines (TNF-a and IL-6) were measured with previously described assays. Patients were categorized by 25(OH)D level as follows: Deficient: <20 ng/mL; Insufficient 20–29 ng/mL; Normal: >29 ng/mL. We used simple descriptive statistics to describe the study population and multiple linear regression (adjustments for age and baseline liver dysfunction) to assess the linear association between 25(OH)D level and markers of inflammation. Results: 39 patients were enrolled. The median age was 68 years (IQR: 56–79) and 44% were female. The median 25(OH)D level was 25.2 (IQR: 21.9–32.7). 15% of patients expired in-hospital. 23/39 (59%) of patients were either deficient or insufficient (Figure 1). 2/2 (100%) of patients with baseline liver dysfunction were 25(OH)D deficient and 5/6 (83%) of deaths were patients who had insufficient levels of 25(OH)D. There was an inverse association between 25(OH)D level and TNF-a (p = 0.03; Figure 2) and IL-6 (p = 0.04). Conclusion: In this small cohort of critically ill adults with sepsis, there was a high prevalence of vitamin D deficiency or insufficiency. The majority (83%) of who expired in-hospital were vitamin D insufficient. There is an inverse association between 25(OH)D levels and Antipyretic Use Does Not Increase Mortality in Emergency Department Patients with Severe Sepsis Nicholas M. Mohr1, Brian M. Fuller2, Craig A. McCammon3, Rebecca Bavolek2, Kevin Cullison4, Matthew Dettmer2, Jacob Gadbaw5, Elizabeth C. Hassebroek1, Sarah Kennedy2, Nicholas Rathert2, Christine Taylor5 1 University of Iowa Carver College of Medicine, Iowa City, IA; 2Washington University School of Medicine, St. Louis, MO; 3Barnes-Jewish Hospital, St. Louis, MO; 4Saint Louis University School of Medicine, St. Louis, MO; 5 Washington University in St. Louis, St. Louis, MO Background: Fever is common in the emergency department (ED), and 90% of those diagnosed with severe sepsis present with fever. Despite data suggesting that fever plays an important role in immunity, human data conflict on the effect of antipyretics on clinical outcomes in critically ill adults. Objectives: To determine the effect of ED antipyretic administration on 28-day in-hospital mortality in patients with severe sepsis. Methods: Single-center, retrospective observational cohort study of 171 febrile severe sepsis patients presenting to an urban academic 90,000-visit ED between June 2005 and June 2010. All ED patients meeting the following criteria were included: age ‡ 18, temperature ‡ 38.3C, suspected infection, and either systolic blood pressure £ 90 mmHg after a 30 mL/kg fluid bolus or lactate of ‡ 4. Patients were excluded for a history of cirrhosis or acetaminophen allergy. Antipyretics were defined as acetaminophen, ibuprofen, or ketorolac. Results: One hundred-thirty five (78.9%) patients were treated with an antipyretic medication (89.4% acetaminophen). Intubated patients were less likely to receive antipyretic therapy (51.9% vs. 84.0%, p < 0.01), but the groups were otherwise well matched. Patients requiring ED intubation (n = 27) had much higher in-hospital mortality (51.9% vs. 7.6%, p < 0.01). Patients given an antipyretic in the ED had lower mortality (11.9% vs. 25.0%, p < 0.05). When multivariable logistic regression was used to account for APACHE-II, intubation status, and fever magnitude, antipyretic therapy was not associated with mortality (adjusted OR 0.97, 0.31–3.06, p = 0.96). Conclusion: Although patients treated with antipyretic therapy had lower 28-day in-hospital mortality, antipyretic therapy was not independently associated with mortality in multivariable regression analysis. These findings are hypothesis-generating for future clinical trials, as the role of fever control has been largely unexplored in severe sepsis (Grant UL1 RR024992, NIHNCRR). S162 298 2012 SAEM ANNUAL MEETING ABSTRACTS Risk Factors for Unplanned Transfer to Intensive Care Within 24 Hours of Admission from the Emergency Department in an Integrated Health Care System M. Kit Delgado1, Vincent Liu2, Jesse M. Pines3, Patricia Kipnis2, Gabriel J. Escobar2 1 Stanford University School of Medicine, Stanford, CA; 2Kaiser Permanente, Division of Research, Oakland, CA; 3George Washington University, Washington, DC Background: ED patients admitted to hospital wards who are subsequently transferred to the intensive care unit (ICU) within 24 hours have higher mortality than direct ICU admissions. Objectives: Describe risk factors for unplanned transfer to the ICU within 24 hours of ward arrival from the ED. Methods: Retrospective cohort analysis of all ED nonICU admissions (N = 178,315) to 14 U.S. community hospitals from 2007–09. We tabulated patient demographics, clinical characteristics, and hospital volume by the outcome of unplanned ICU transfer. We present factors that were independently associated with unplanned ICU transfer within 24 hours after adjusting for patient and hospital differences in a multilevel mixed-effects logistic regression model. Results: Of all ED non-ICU admissions, 4,252 (2.4%) were transferred to the ICU within 24 hours. After adjusting for patient and hospital differences, the top five admitting diagnoses associated with unplanned transfer were: sepsis (odds ratio [OR] 2.6; 95% CI 2.1– 3.1), catastrophic conditions (OR 2.3; 95% 1.9–2.8), pneumonia/acute respiratory infections (OR 1.6; 95% CI 1.4–1.8), acute myocardial infarction (AMI) (OR 1.6; 95% CI 1.3–1.8), and chronic obstructive pulmonary disease (COPD) (OR 1.5; 95% CI 1.3–1.7). Other factors associated with unplanned transfer included: male sex, Comorbidity Points Score (COPS) >145, Laboratory Acute Physiology Score (LAPS) >7, and arriving on the ward between 11 PM-7 AM. Decreased risk of unplanned transfer was found with admission to monitored transitional care units vs. non-monitored wards (OR 0.86; 95% CI 0.80–0.96) and admission to a high-volume vs. low-volume hospital (OR 0.73; 95% CI 0.59–0.89). Conclusion: ED patients admitted with respiratory conditions, sepsis, AMI, multiple comorbidities, and abnormal lab results are at higher risk for unplanned ICU transfer and may benefit from better inpatient triage from the ED, earlier intervention to prevent acute decompensation, or closer monitoring. More research is needed to determine how intermediate care units, hospital volume, time of day, and sex affect risk of unplanned ICU transfer. 299 Effect of Weight-Based Volume Loading on the Inferior Vena Cava in Fasting Subjects: A Randomized, Prospective Double-Blinded Trial Margaret R. Lewis1, Anthony J. Weekes1, Zachary P. Kahler1, Donald E. Stader1, Dale P. Quirke2, H. James Norton1, Courtney Almond1, Dawn Middleton1, Vivek S. Tayal1 1 Carolinas Medical Center, Charlotte, NC; 2Bay Area Emergency Physicians, Clearwater, FL Background: Inferior vena cava ultrasound assessment (IVC-US) has been proposed as a noninvasive method of assessing volume status. Current literature is divided on its ability to do so. Objectives: Our primary hypothesis was that, in fasting asymptomatic subjects, larger fluid boluses would lead to proportional IVC-US changes. Our secondary endpoint was to determine inter-subject variation in IVCUS measurements. Methods: The authors performed a prospective randomized double-blinded trial using fasting volunteers to determine the inter-subject variation of baseline IVCUS and changes associated with different acute weightbased volume loads. Subjects with no history of cardiac disease or hypertension fasted for 12 hours and were then randomly assigned to receive a normal saline bolus of 2 ml/kg, 10 ml/kg or 30 ml/kg over 30 minutes. IVC-US was performed before and after each bolus. Results: Forty-two fasting subjects were enrolled. Mean (±SD) (in cm) baseline maximum IVC diameter (IVC max) of the 42 subjects was 2.07 ± 0.57: range was 0.85 to 4.53 cm. Mean baseline minimum IVC diameter (IVC min) was 1.4 ± 0.64. Mean baseline caval index was 0.33 ± 0.15. The overall mean (±SD) (95% confidence interval [CI]) findings for pre- and post-fluid differences were: IVC max 0.18 ± 0.40 (CI 0.06, 0.31); IVC min 0.31 ± 0.46 (CI 0.17, 0.46), and caval index )0.09 ± 0.14 (CI )0.14, )0.05) and all were statistically significant. The groups receiving 10 ml/kg and 30 ml/kg had statistically significant changes in caval index; ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 however the 30 ml/kg group had no significant change in mean IVC diameter. One-way ANOVA differences between the means of all groups were not statistically different. Conclusion: Overall, there were statistically significant differences in mean IVC-US measurements before and after fluid loading, but not between groups. Fasting asymptomatic subjects had a wide inter-subject variation in both baseline IVC-US measurements and fluid-related changes. The wide differences within our 30 ml/kg group may limit conclusions regarding proportionality. • www.aemj.org S163 come under increasing pressure to improve scores in order to reduce potential financial losses under the program. Our data provide early information on the types of hospitals with the greatest opportunity for improvement. An Early Look at Performance Variation on Emergency Care Measures included in Medicare’s Hospital Inpatient Value-Based Purchasing Program Megan McHugh, Jennifer Neimeyer, Rahul K. Khare, Emilie Powell Northwestern University, Chicago, IL Do Improvements In Emergency Department Operations Influence The Patient Experience? Impacts Of Large Scale Implementation Of Overcapacity Protocols On Perceptions Of Crowding And Overall Ratings Of Care Timothy Cooke1, Eddy Lang2, Christian Schmid3, Grant Innes2, Nancy Guebert3, Brian Holroyd4, Brian Rowe4, Andrew McRae2, John Cowell1 1 Health Quality Council of Alberta, Calgary, AB, Canada; 2University of Calgary, Calgary, AB, Canada; 3Alberta Health Services, Calgary, AB, Canada; 4University of Alberta, Edmonton, AB, Canada Background: Medicare’s Hospital Inpatient ValueBased Purchasing (VBP) program is the first national, mandatory effort to reward and penalize hospitals for their performance. In 2012, 1% of Medicare payments to hospitals will be withheld and distributed back based on achievement or improvement on twelve process measures and eight patient satisfaction measures. Four process measures are related to care provided in the ED: fibrinolytic therapy within 30 minutes, percutaneous intervention within 90 minutes, blood cultures in ED before initial antibiotic, and initial antibiotic selection for community acquired pneumonia. Objectives: To identify variations in performance on ED measures included in the VBP program. These early data offer a first look at the types of hospitals where improvement efforts may be best directed. Methods: This was an exploratory, descriptive analysis. We obtained 2008–2010 performance data from Hospital Compare for the population of hospitals that met the criteria for the VBP program (i.e., paid under the Inpatient Prospective Payment System with adequate sample sizes). Data were merged with the 2009 American Hospital Association Annual Survey to obtain hospital characteristics. Two researchers independently calculated performance scores for each hospital following the program guidelines published in the Federal Register, and compared with mean performance across all hospitals. T-tests and ANOVA were used to detect differences in performance across by hospital bed size, number of ED visits, teaching status, ownership, system affiliation and region. Results: 3,030 hospitals qualified for the VBP program. There were significant differences in performance on ED measures by ownership (P < 0.0001) and region (P = 0.0002). Scores on ED process measures were highest at for-profit hospitals (27% above average) and hospitals in the south (5% above average), and lowest at public hospitals (16% below average) and hospitals in the northeast (8% below average). Conclusion: There was considerable variation in performance on the ED measures included in the VBP program by hospital ownership and region. ED directors may Background: In December 2010, the province-wide health authority launched a series of measures, known as the Overcapacity Protocols (OCP), to address crowding in Alberta EDs. Its effect on the patient experience was evaluated as changes in throughput and output can compromise components of care ranging from perceptions of crowding to communication and privacy. Objectives: To determine if OCP and its resulting effect on ED crowding influenced how patients perceived their care. We hypothesized that OCP would result in improved perceptions of ED crowding. Methods: Design/Setting - An independent agency mandated by the government collected and analyzed ED patient experience data using a comprehensive, validated multidimensional instrument and a random periodic sampling methodology of all ED patients. A prospective pre-post experimental study design was employed in the eight community and tertiary care hospitals most affected by crowding. Two 5.5 month study periods were evaluated (Pre: 28/06- 12/12/2010; Post: 13/12/2010–29/ 05/2011). Outcomes - The primary outcome was patient perception of wait times and crowding reported as a composite mean score (0–100) from six survey items with higher scores representing better ratings. The overall rating of care by ED patients (composite score) and other dimensions of care were collected as secondary outcomes. All outcomes were compared using chi-square and two-tailed Student’s t-tests. Results: A total of 3774 surveys were completed in both the pre-OCP and post-OCP study periods representing a response rate of 45%. The composite for perceived wait times and crowding improved from 61.7 to 65.3 in Calgary and from 59.8 to 64.4 in Edmonton (P < 0.001) with some variation among specific sites. The overall composite score for care was 75.6 of Calgary patients and 74.1 of Edmonton patients, respectively in the pre-OCP phase and did not change in the post-period (77.3 and 76.1 for both cities in the postOCP phase comparison; P = NS). The global rating also remained unchanged following the intervention for both Calgary and Edmonton zones. Other dimensions of care remained unchanged. (See tables.) 300 301 S164 2012 SAEM ANNUAL MEETING ABSTRACTS Table 1 - Abstract 301: Specific items: patient experience of wait times Specific question item Proportion who found the waiting room extremely or very crowded Proportion who could not find a comfortable place to sit in the waiting room Proportion who considered leaving Proportion who reported waited more than 15 minutes to speak with the triage nurse Proportion who reported waited longer than 2 hours to see a physician Proportion whose emergency department visit was reported as longer than 4 hours Weighted N Pre-OCP Calgary (%) Post -OCP Calgary (%) Pre-OCP Edmonton (%) Pre-OCP Edmonton (%) p-value 3696 21.7 16.1 36.8 28.8 <0.001 3215 21.4 21.0 27.8 24.0 .04 4268 24.0 17.6 25.5 21.8 .02 3738 36.8 37.5 NS 34.8 31.1 .05 4070 41.6 29.5 <0.001 40.4 33.8 <0.001 4103 62.2 57.0 .055 60.2 54.1 0.001 p-value .015 NS .003 Table 2 - Abstract 301: Patient Experience Domain - Composite Scores Domain Staff care composite Wait and crowding composite Wait time communication composite Discharge information and concerns Respect composite without LWBS Medication communication composite Facility cleanliness composite Privacy composite Pain management composite N Pre-OCP Calgary Post-OCP Calgary Sig (2-tailed) Pre-OCP Edmonton Post-OCP Edmonton Sig (2-tailed) 3761 3726 3572 2774 3767 1275 3628 3722 2299 79.3 61.7 52.9 51.9 83.5 80.3 82.5 82.7 63.0 80.5 65.3 52.7 55.5 85.8 76.1 82.6 84.0 65.2 NS <0.001 NS NS 0.018 NS NS NS NS 77.4 59.8 43.8 52.6 84.7 77.4 76.0 80.3 59.2 78.1 64.4 46.2 53.8 85.3 76.4 78.5 84.3 63.1 NS <0.001 NS NS NS NS .008 <0.001 NS Conclusion: Improving ED crowding results in significant, albeit modest, improvements in patients’ perceptions of crowding in busy urban hospital EDs. Site level variability in these effects should be explored further. 302 Cost Savings Associated with Use of Observation Units as Compared to 1-day Inpatient Admission in Massachusetts Philip Anderson, Leah S. Honigman, Marie-France Petchy, Laura Burke, Shiva Gautam, Peter Smulowitz Beth Israel Deaconess Medical Center/Harvard Medical School, Boston, MA Background: ED observation units are increasingly utilized nationwide and allow continued medical care for patients who are unable to be safely discharged but who may not require a full hospital admission. In addition to providing effective care, observation units have been tied to decreased ED crowding and lower rates of diversion. Additionally, ED observation units have the potential economic benefit of avoiding the high cost of hospital admissions. Interventions to reduce admission are vital to cost-savings in health care; however, the cost savings associated with preferential use of observation units has not been specifically described. Objectives: To determine the cost savings associated with using ED observation stays as compared to a 1-day inpatient admission in the state of Massachusetts. Methods: Using the Massachusetts Division of Health Care Finance and Policy Acute Hospital Case Mix Databases, we analyzed patient visits in Fiscal Year 2008 from 65 hospitals for ED observation stays and 1-day length-of-stay (LOS) inpatient admissions from the ED as well as the total charges for each category of visit. Within each hospital, we first calculated the average charge for observation visits. The cost savings was then calculated as the difference between the calculated charges for 1-day LOS and the charges that would have been accrued if these additional admissions had instead been ED observation status. We determined total cost savings, mean cost savings per hospital, and average savings per patient with a 95%CI constructed around the difference in savings. Results: A total of 103,150 ED observation stays and 67,352 1-day LOS inpatient admissions were examined. The mean cost savings per hospital in Massachusetts was $354,372 (95%CI, -$194,015 to $902,759), and the total estimated savings for all 65 hospitals was $23 million (p = 0.07, paired t-test). This translated to an average savings per patient of $289 (95%CI -$54 to $633) for an observation stay as opposed to 1-day admission. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Conclusion: Despite the range of potential savings between hospitals in Massachusetts, significant cost savings might be realized if patients who only require a short-stay admission are instead placed in ED observation units. These results highlight the opportunity for EDs to utilize observation units to their full capacity to reduce overall health spending. 303 National Study of Non-Urgent Emergency Department Visits Leah S. Honigman1, Jennifer L. Wiler2, Sean Rooks2, Adit A. Ginde2 1 Beth Israel Deaconess Medical Center/Harvard Medical School, Boston, MA; 2University of Colorado School of Medicine, Denver, CO Background: Policymakers have suggested reducing ‘‘unnecessary’’ ED visits as a way to generate significant cost savings for the U.S. health care system. Although what constitutes an unnecessary ED visit is ill-defined, visits classified as non-urgent at triage by the emergency severity index (ESI) classification system are often considered unnecessary. Objectives: To compare resource utilization of ED visits characterized as non-urgent at triage to immediate, emergent, or urgent (IEU) visits. Methods: We performed a retrospective, cross-sectional analysis of the 2006–2009 National Hospital Ambulatory Medical Care Survey. Visits were categorized by the assigned ESI five-level triage acuity score representing immediate, emergent, urgent, semi-urgent, or non-urgent. Within each triage categorization, clinical and hospital characteristics were analyzed. Results: In 2006–2009, 10.1% (95%CI 9.2–11.2) of U.S. ED visits were categorized as non-urgent. Most (87.8%, 95%CI 86.3–89.2) non-urgent visits had some diagnostic testing or treatment in the ED. Although more common in IEU visits (52.9%, 95%CI 51.6–54.2), imaging was also common in non-urgent visits (29.8%, 95%CI 27.8–31.8), with 7.3% (95%CI 6.2–8.6) of non-urgent visits requiring cross-sectional imaging. Similarly, procedures were performed more frequently in IEU (56.3%, 95%CI 53.5–59.0) compared to non-urgent (34.1%, 95%CI 31.8–36.4), while medication administration was similar between the two groups (80.6%, 95%CI 79.5– 81.7 vs. 76.3%, 95% CI 74.7–77.8, respectively). The rate of hospital admission was 4.0% (95%CI 3.3–4.8) for non-urgent visits compared to 19.8% (95%CI 18.4–21.3) for IEU visits. Visits requiring critical care, operating room, or catheterization lab intervention was 0.5% (95%CI 0.3–0.6) for non-urgent vs. 3.4% (95%CI 3.1– 3.8) for IEU. Conclusion: Most non-urgent ED patients required some diagnostic or therapeutic interventions during their visits. Relatively low, but not insignificant, proportions of non-urgent ED patients had advanced imaging or were admitted to the hospital, sometimes to critical care settings. These data call into question non-urgent ED visits being categorized as ‘‘unnecessary,’’ particularly in the setting of limited access to primary care for acute illness or injury. • 304 www.aemj.org S165 ‘If My ER Doesn’t Close, But Someone Else’s Does, Am I At Higher Risk Of Dying?’ An Analysis Of Cardiac Patients And Access To Care Renee Hsia1, Tanja Srebotnjak2, Judy Maselli1 1 University of California San Francisco, San Francisco, CA; 2Ecologic Institute, San Mateo, CA Background: Between 1990–2009, the number of nonrural emergency departments (EDs) dropped by 27%, contributing to the national crisis of supply and demand for ED services. Negative health outcomes have been documented in patients who experience ED closure directly perhaps because they travel farther distances to reach emergency care. Objectives: We seek to determine if patients who experience ED closure indirectly - those who have a closure in their vicinity but do not need to travel farther - have higher inpatient mortality rates due to AMI than those patients without closures in their communities. Methods: We performed a retrospective study using all admissions from non-federal hospitals in California for acute myocardial infarction (AMI) between 1999–2008. We compared in-patient mortality from AMI for patients who lived in a community with either 2.5 miles or 5 miles of a closure but did not need to travel farther to the nearest ED with those who did not. We used patient-level data from the California Office of Statewide Health and Planning Development (OSHPD) database patient discharge data, and locations of patient residence and hospitals were geo-coded to determine any changes in distance to the nearest ED. We applied a generalized linear mixed effects model framework to estimate a patient’s likelihood to die in the hospital of AMI as a function of being affected by a neighborhood closure event. Results: Between 1998–2008, there were 352,064 patients with AMI and 29 ED closures in California. Of the AMI patients, 5868 (1.7%) experienced an ED closure within a 2.5-mile radius and had 15% higher odds (OR 1.15, 95% CI 1.04, 1.96) of mortality compared with those who did not. Results of the 5-mile radii analyses were similar (OR 1.13, 95% CI 1.03, 1.23). Conclusion: Our findings suggest that patients indirectly experiencing ED closure do face poor health outcomes. These results should be factored into closure decisions and regionalization plans. (Originally submitted as a ‘‘late-breaker.’’) 305 Emergency Department Visit Rates after Common Inpatient Procedures for Medicare Beneficiaries Keith E. Kocher, Justin B. Dimick University of Michigan, Ann Arbor, MI Background: Fragmentation of care has been recognized as a problem in the US health care system. However, little is known about ED utilization after hospitalization, a potential marker of poor outpatient care coordination after discharge, particularly for common inpatient-based procedures. S166 2012 SAEM ANNUAL MEETING ABSTRACTS Objectives: To determine the frequency and variability in ED visits after common inpatient procedures, how often they result in readmission, and related payments. Methods: Using national Medicare data for 2005–2007, we examined ED visits within 30 days of hospital discharge after six common inpatient procedures: percutaneous coronary intervention, coronary artery bypass grafting (CABG), elective abdominal aortic aneurysm repair, back surgery, hip fracture repair, and colectomy. We categorized hospitals into risk-adjusted quintiles based on the frequency of ED visits after the index hospitalization. We report visits by primary diagnosis ICD-9 codes and rates of readmission. We also assessed payments related to these ED visits. Results: Overall, the highest quintile of hospitals had 30-day ED visit rates that ranged from a low of 17.8% with an associated 7.3% readmission rate (back surgery) to a high of 27.8% with an associated 13.6% readmission rate (CABG). The most variability was more than 3-fold and found among patients undergoing colectomy in which the worst-performing hospitals saw 24.1% of their patients experienced an ED visit within 30 days while the best-performing hospitals saw 7.4%. Average total payments for the 30-day window from initial discharge across all surgical cohorts varied from $18,912 for patients discharged without subsequent ED visit; $20,061for those experiencing an ED visit(s); $38,762 for those readmitted through the ED; and $33,632 for those readmitted from another source. If all patients who did not require readmission also did not incur an ED visit within the 30-day window, this would represent a potential cost savings of $125 million. Conclusion: Among elderly Medicare recipients there was significant variability between hospitals for 30-day ED visits after six common inpatient procedures. The ED visit may be a marker of poor care coordination in the immediate discharge period. This presents an opportunity to improve post-procedure outpatient care coordination which may save costs related to preventable ED visits and subsequent readmissions. 306 Does Pharmacist Review of Medication Orders Delay Medication Administration in the Emergency Department? James P. Killeen, Theodore C. Chan, Gary M. Vilke, Sally Rafie, Ronald Dunlay, Edward M. Castillo University of California, San Diego, San Diego, CA Background: Despite the recommendation of The Joint Commission (TJC), prospective pharmacist review of medications administered in the ED remains controversial. Proponents believe pharmacist review reduces medication errors, whereas others are concerned about potential delays in medication delivery. Objectives: We sought to assess the effect of pharmacist medication review on ED patient care, in particular time from physician order to medication administration for the patient (order-to-med time). Methods: We conducted a multi-center, before-after study in two EDs (urban academic teaching hospital and suburban community hospital, combined census of 61,000) after implementation of the electronic prospective pharmacy review system (PRS). The system allowed a pharmacist to review all ED medication orders electronically at the time of physician order and either approve or alter the order. We studied a 5-month time period before implementation of the system (pre-PRS, 7/1/10-11/30/11) and after implementation (post-PRS, 7/ 1/11-11/30/11). We collected data on all ED medication orders including dose, route, class, pharmacist review action, time of physician order, and time of medication administration. Differences in order-to-medication between the pre- and post-PRS study periods were compared using a Mann-Whitney U test. Median times and interquartile ranges (25, 75) are reported. Results: In the two study EDs, there were 17,428 patients seen in the ED resulting in 58,097 medication orders during the pre-PRS period, and 18,862 ED patients resulting in 65,968 medication orders in the post-PRS period. Overall median order-to-med times were minimally, but statistically longer for the post-PRS (19 minutes; IQR 9, 35) compared with pre-PRS periods (18 minutes; IQR 8, 35). When stratified by route, a similar pattern in was seen for PO (17 [10, 31] vs 16 [7, 29]) and IM/SQ (18 [10, 33] vs 16 [8, 33]) routes, but no difference was seen for IV routes (19 [9, 37] vs 19 [9, 39], p = 0.127) post- and pre-PRS respectively. Topical medications demonstrated a larger magnitude difference of 23 [10, 52] vs 13 [4, 28] minutes, post- vs pre-PRS respectively. Conclusion: In this multicenter ED study, an electronic prospective pharmacy review system for medications was associated with a minimal, but significant increase in time from physician order to medication delivery. 307 Empiric Antibiotic Prescribing Practices after the Introduction of a Computerized Order Entry System in an Adult Emergency Department Sheeja Thomas, Ian Schwartz, Jeffrey Topal, Vinnita Sinha, Caroline Pace, Charles R. Wira III Yale University School of Medicine, New Haven, CT Background: Antibiotics for pneumonia are frequently prescribed in the ED. Selection does not always comply with existing guidelines. Incorporating computerized decision support systems in the clinical arena may bolster compliance to prescribing guidelines. Objectives: The objective of this study was to assess whether compliance to institutional empiric antibiotic guidelines for pneumonia could be improved through implementation of a computerized order entry system. Methods: This is a prospective and consecutive preand post-interventional study evaluating empiric antibiotic prescribing patterns in a high-volume, academic ED after the installation of an electronic order entry screen for non-critically ill ED patients admitted with pneumonia. Appropriateness of regimens was based on adherence to Infectious Disease Society of America (IDSA) and hospital guidelines. Established a priori, the pre-intervention (PRE) enrollment period was from ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 September 1, 2009 to October 21, 2009. The intervention was launched on September 21, 2010 with a subsequent educational and transitional period. The post-intervention (POST) enrollment period was from February 1, 2011 to March 6, 2011. Results: 205 patients were identified (PRE, n = 100; POST, n = 105). Demographic variables in the PRE and POST groups respectively were age of 70.1 + 20.2 vs. 67.5 + 18.8 (P = 0.35), with 65% and 44.7% being female (P = 0.005). In the PRE group, 59% (59 of 100) received appropriate antibiotics by ED providers. In the POST group, 64.8% (68 of 105) received appropriate antibiotics (P = 0.47). Subgroup analysis revealed a significant increase in ED provider compliance with hospitalacquired pneumonia guidelines from 5.1% (2/39) to 31.8% (14/44) (P = 0.002) after the implementation of the computerized order entry screens. Conclusion: The institution of a computerized physician order entry system did not change overall empiric antibiotic prescribing compliance for patients admitted from the ED with pneumonia. But, there was significantly higher compliance in the POST group patients who were suspected of having hospital-acquired pneumonia. Implementation of a computerized order entry system is feasible, and may improve provider compliance with more complicated treatment guidelines. 308 Factors Associated with Left Before Treatment Complete Rates Differ Depending on ED Patient Volume Daniel A. Handel1, John R. Woods2, James Augustine3, Charles M. Shufflebarger2 1 Oregon Health & Science University School of Medicine, Portland, OR; 2Indiana University School of Medicine, Indianapolis, IN; 3EMP, Cincinnati, OH Background: The proportion of patients leaving the emergency department (ED) before treatment is • www.aemj.org S167 complete (LBTC) is known to vary as a function of ED patient volume; high-volume EDs have generally higher LBTC rates. Objectives: The objective of this study was to look at factors associated with higher rates of LBTC across a large, national sample. We hypothesized that explanatory variables associated with LBTC rate would differ depending on ED patient volume. Methods: EDs (n = 161) from the 2010 national ED Benchmarking Alliance database were grouped by annual patient volume. Within each volume category, separate linear regressions were performed to identify factors that help explain differences in LBTC rates. Results: ED metrics that were significantly associated with LBTCs varied across ED patient-volume categories (Table). For EDs seeing less than 20K patients annually, the percentage of EMS arrivals admitted to the hospital and ED square footage were both weakly associated with LBTCs (p = 0.09). For EDs seeing at least 20K–39K patients, median ED length of stay (LOS), percent of patients admitted to hospital through the ED, percent of EMS arrivals admitted to hospital, and percent of pediatric patients were all positively associated, while percent of patients admitted to the hospital was negatively associated with LBTCs. For EDs seeing 40K–59K, median LOS and percent of x-rays performed were positively associated, while percent of EKGs performed was negatively associated with LBTCs. For EDs seeing 60K–79K, percent of patients admitted to the hospital through the ED was negatively associated and percent of EKGs performed was positively associated with LBTCs. For EDs with volume greater than 80K, none of the selected variables were associated with LBTC. Conclusion: ED factors that help explain high LBTC rates differ depending on the size of an ED. Interventions attempting to improve LBTC rates by modifying ED structure or process will need to consider baseline ED volume as a potential moderating influence. Table - Abstract 308: Coefficients from regression analyses to predict variation in LBTC rates. Separate regressions were performed for categories of EDs grouped by annual patient volume <20K (n = 25) Predictor Variable Constant Median LOS LOS (Patients released) LOS (Patients admitted) Hospital admits thru the ED (%) Patients admitted to hospital (%) Patients transferred (%) High CPT acuity (%) EMS arrivals (%) EMS arrivals (admitted %) Pediatric patients (%) EKGs per 100 patients X)rays per 100 patients ED square footage Variance explained (R2) Coeff 0.01238 )0.00009 0.00029 )0.00004 )0.01318 )0.05402 0.04529 )0.01008 0.04739 0.05882 )0.01061 )0.00042 )0.00009 )0.00000 0.68 20–39K (n = 79) p 0.09 0.09 Coeff )0.05078 0.00023 )0.00008 0.00002 0.02178 )0.08750 0.12118 0.01701 0.02291 0.02192 0.02829 0.00003 0.00014 )0.00000 0.68 40–59K (n = 47) p 0.0002 0.05 0.0005 <0.05 <.05 Coeff )0.03475 0.00028 0.00001 0.00003 0.01451 )0.06167 )0.21078 0.00740 )0.08348 )0.00522 )0.02314 )0.00084 0.00064 0.00000 0.49 60–79K (n = 27) P .05 .04 .03 Coeff )0.00062 )0.00004 0.00027 0.00004 )0.09562 )0.09336 )0.66502 0.02750 0.05302 0.01902 0.05935 0.00141 )0.00043 0.00000 0.80 >80K (n = 17) p 0.03 0.03 Coeff )0.01270 )0.00057 0.00083 0.00015 )0.03242 0.18719 )0.83174 0.01789 )0.37387 )0.10471 0.01148 0.00230 )0.00067 0.00000 ) p S168 309 2012 SAEM ANNUAL MEETING ABSTRACTS Environmental Contamination in the Emergency Department - A Non-toxic Alternative to Surface Cleaning Stephanie Benjamin, Ashlee Edgell, Allie Thompson, Neil Batra, Victor Blanco, Christopher Lindsell, Andra L. Blomkalns University of Cincinnati, Cincinnati, OH Background: Work environment surface contamination plays a central role in the transmission of many health care acquired pathogens such as MRSA and VRE. Current federal guidelines for hospitals recommend the use of quaternary ammonium compounds (QUATs) for surface cleaning. These agents have limitations in the pathogens for which they are effective, toxicity, corrosiveness, and time needed for effectiveness. No data exist comparing these sterilants in a working ED setting where varied surfaces need to be cleaned quickly and effectively. Objectives: Our study sought to compare bacterial growth of samples taken from surfaces after use of a common approved QUAT compound and a virtually non-toxic, commercially available solution containing elemental silver (0.02%), hydrogen peroxide (15%), and peroxyacetic acid (20%) (SHP) in a working ED. We hypothesized that, based on controlled laboratory data available, SHP compound would be more effective on surfaces in an active urban ED. Methods: We cleaned and then sampled three types of surfaces in the ED (suture cart, wooden railing, and the floor) during midday hours one minute after application of tap water, QUAT, and SHP and then again at 24 hours without additional cleaning. Conventional environmental surface surveillance RODAC media plates were used for growth assessment. Images of bacterial growth were quantified at 24 and 48 hours. Standard cleaning procedures by hospital staff were maintained per usual. Results: SHP was superior to control and QUAT one minute after application on all three surfaces. QUAT and water had 10x and 40x more bacterial growth than the surface cleaned with SHP, respectively. 24 hours later, the SHP area produced fewer colonies sampled from the wooden railing: 4x more bacteria for QUAT, and 5x for water when compared to SHP. 24h cultures from the cart and floor had confluent growth and could not be quantified. Conclusion: SHP outperforms QUAT in sterilizing surfaces after one minute application. SHP may be a superior agent as a non-toxic, non-corrosive, and effective agent for surfaces in the demanding ED setting. Further studies should examine sporidical and virucidal properties in a similar environment. 310 Emergency Department Extended Waiting Room Times Remain a Barrier to the Opportunity for High Patient Satisfaction Scores John B. Addison, Scott P. Krall Christus Spohn Memorial Hospital, Corpus Christi, TX Background: Hospitals subject to the Inpatient Prospective Payment System (IPPS) submit HCAHPS data to CMS for IPPS annual payment update. IPPS hospitals that fail to publicly report may have their annual payment update reduced. Improving and maintaining patient satisfaction has become a prominent focus for hospital administration. Objectives: Evaluate the effect on patient satisfaction of increasing waiting room times and physician evaluation times. Methods: Emergency department flow metrics were collected on a daily basis as well as average daily patient satisfaction scores. The data were from July 2010 through February 2011, in a 44,000 census urban hospital. The data were divided into equal intervals. The arrival to room time was divided by 15 minute intervals up to 135 minutes with the last group being greater than 136 minutes. The physician evaluation times were divided into 20 minute intervals, up to 110, the last group greater than 111 with 46 days in the group. Data were analyzed using means and standard deviations, and well as ANOVA for comparison between groups. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 • www.aemj.org S169 Table 1 - Abstract 310: Effect Waiting Room Time minutes 0–15 16–30 31–45 46–60 61–75 76–90 91–105 106–120 121–135 Overall Visit Satisfaction Provider Satisfaction 88.4 94 83 84 86 84 79 77 84 88 80 82 84 88 80 84 75 74 Results: The overall satisfaction score for the outpatient emergency visit was higher when the patient was in a room within 15 minutes of arrival (88.4, std deviation 5.9), analysis of variation between the groups had a p = 0.13, for the means of each interval (see table 1). The total satisfaction with the visit as well as satisfaction with the provider dropped when the evaluation extended over 110 minutes, but was not statistically significant on ANOVA analysis (see table 2 for means). Conclusion: Once a patient’s time in the waiting room extends beyond 15 minutes, you have lost a significant opportunity for patient satisfaction; once they have been in the waiting room for over 120 minutes, you are also much more likely to receive a poor score. Physician evaluation time scores are much more consistent but as longer evaluation times occurred beyond total of 110 minutes we started to see a trend downward in the satisfaction score. Table 2 - Abstract 310: Effect of Physician Evaluation Time Minutes Overal Visit Satisfaction Satisfaction with Provider 311 51–70 71–90 91–110 >110 82 84 83 85 83 83 78 79 Pain Management Practices Vary Significantly at Three Emergency Departments In A Single Health Care System Melissa L. McCarthy1, Ru Ding1, Scott L. Zeger2, Melinda J. Ortmann2, Donald M. Steinwachs2, Julius C. Pham2, Rodica Retezar2, Walter Atha2, Edward S. Bessman2, Sara C. Bessman2, Gabor D. Kelen2, Nancy K. Roderer2 1 George Washington University, Washington, DC; 2Johns Hopkins University, Baltimore, MD Background: Previous studies have noted variation in ED analgesic rates by patient factors such as age and race. Objectives: To compare the pain management practices among three EDs in the same health care system by patient, clinical, provider, and site characteristics. We expected variation in pain management practices across the sites and providers. Methods: We conducted a randomized controlled trial (RCT) to evaluate the effect of information prescriptions on medication adherence. This study focused on a subgroup of the subjects who reported moderate (4–7) to severe pain (8–10) at triage. Subjects were enrolled at three EDs affiliated with one health care system. Site A is an academic, tertiary care hospital, site B is a community teaching affiliate hospital, and site C is a community hospital. To be eligible for the RCT, adult ED patients had to be discharged with a prescription medication. Patients were excluded if they were non-English speaking, disoriented, or previously enrolled. Research assistants prospectively enrolled subjects from November 15, 2010–September 9, 2011 (rotating among the three EDs) during the hours of 10AM–10PM seven days per week. Of the 3,940 enrollees, 2,603 subjects reported moderate (35%) to severe pain (65%). The main outcomes are pain medication received in the ED or prescribed at discharge (Rx). Pain medication rates were modeled as a function of patient, clinical, provider, and site factors using a hierarchical logistic regression model that accounted for clustering of patients treated by the same provider and clustering of providers working in the same ED. Results: In all three EDs, pain medication rates (both in ED and Rx) varied significantly by clinical factors including location of pain, discharge diagnosis, pain level, and acuity. We observed little to no variation in pain medication rates by patient factors such as age, sex, race, insurance, or prior ED visits. The table displays key pain management practices by site and provider. After adjusting for patient and clinical characteristics, significant differences in pain medication rates remained by provider and site (see figure). Conclusion: Within this health system, the approach to pain management by both providers and sites is not standardized. Investigation of the potential effect of this variability on patient outcomes is warranted. Table - Abstract 311: Pain Management Practices by Site And Provider. Pain Management Practice ‡ 1 pain assessment by treating nurse Discharge pain score documented by nurse Received pain medication in ED ED pain medication rate by providers Pain medication prescribed at discharge Pain medication prescription rate by providers Overall N = 2603 Site A N = 879 Site B N = 978 Site C N = 746 77% 55% 89% 98% 38% 35% 69% 1% 52% 72% 28% 61% 41–98% 12–33% 44–86% 76% 78% 72% 59–92% 53–91% 37–93% 75% S170 312 2012 SAEM ANNUAL MEETING ABSTRACTS Decreases in Provider Productivity After Implementation of Computer Physician Order Entry Neil Roy, Thomas Damiano, Heather F. Farley, Richard Bounds, Debra Marco, James F. Reed III, Jason T. Nomura Christiana Care Health Systems, Newark, DE Background: Reported patient safety, care integration, and performance improvement of health information technology have resulted in increasing adoption of computer provider order entry (CPOE). However, there are frequent anecdotes of decreased productivity due to workflow changes. Objectives: This study compared the changes in provider productivity measures before, during, and after CPOE implementation. Methods: This was a retrospective cross-sectional study at an academic tertiary care system with an ED volume of >160,000 visits/year. Average of measures of productivity including patients per hour (pt/h), charges per hour (charge/h), and relative value units (RVUs) per hour (RVU/h) were calculated for all providers and visits during the study period. Productivity was compared for three time periods: 6 months prior to CPOE implementation, a transitional period of 4 months immediately following CPOE implementation that included refinements and customizations, and 6 months posttransition. Data were analyzed for significant variances by Kruskal-Wallis nonparametric testing. Significant groups then had time period pairs compared using Student Neuman-Keuls analysis with Wald type 95% CI. Results: All measures showed significant differences, p < 0.01. Average pts/h decreased post-CPOE and did not recover post transitional period, 1.92 ± 0.13 vs 1.75 ± 0.11, p < 0.05. RVU/h also decreased post-CPOE and did not recover post transitional period, 5.23 ± 0.37 vs 4.79 ± 0.32 and 4.82 ± 0.33, p < 0.05. Charges/h also decreased after CPOE implementation and did not recover after system optimization. There was a sustained significant decrease in charges/h of 4.5% ± 6.5% post CPOE and 3.6% ± 6.4% post optimization, p < 0.05. Sub-group analysis for each provider group was also evaluated and showed variability for different providers. Conclusion: There was a significant decrease in all productivity metrics four months after the implementation of CPOE. The system did undergo optimization initiated by providers with customization for ease and speed of use. However, productivity measurements did not recover after these changes were implemented. These data show that with the implementation of a CPOE system there is a decrease in productivity that continues even after a transition period and system customization. 313 Computer Order Entry Systems in the Emergency Department Significantly Reduce the Time to Medication Delivery for High Acuity Patients Shahbaz Syed, Dongmei Wang, Debbie Goulard, Tom Rich, Grant Innes, Eddy Lang University of Calgary, Calgary, AB, Canada Background: Computerized physician order entry (CPOE) systems are designed to increase safety and quality of care; however, their effect on emergency department (ED) efficiency has been evaluated in limited settings. Objectives: Our objective was to assess the effect of CPOE on process times for medication delivery, laboratory testing, and diagnostic imaging after region-wide multi-hospital ED-CPOE implementation. Methods: Setting: This administrative database study was conducted in Calgary, Alberta, a city of 1.2 million people served by three acute care hospitals. Intervention: In March, 2010, our three tertiary care emergency ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 departments implemented a common ED CPOE system. Patients: Eligible patients consisted of a stratified random sample of CTAS 2 emergent (40%) and CTAS 3 urgent (60%) patients seen 30 days preceding CPOE implementation (Control group), 30 days immediately after CPOE implementation (early CPOE group), and 5– 6 months after CPOE implementation (late CPOE group). Overall, nine patient groups of 100 patients were analyzed, including control, early, and late groups from all three hospitals. Outcomes: Primary outcomes were time from physician assessment to imaging and lab completion. An ANOVA and t-test were employed for statistical analysis. Results: After CPOE implementation, time to (TT) first medication decreased in the early and late CPOE groups (from 102.6 to 62.8 min and 65.7 minutes respectively, p < 0.001). Time to first lab result increased from 73.8 min (control) to 76.4 and 85.3 minutes in the early and late groups respectively (p < 0.001). TT first x-ray also increased from 68.1 min to 80.4 and 84.8 minutes, respectively (p < 0.001). Conclusion: Regional implementation of CPOE afforded important efficiencies in time to medication delivery for high acuity ED patients. Increased times observed for laboratory and radiology results may reflect system issues outside of the emergency department and as a result of potential confounding may not be a reflection of CPOE effect. 314 Emergency Medicine Procedures: Examination of Trends in Procedures Performed by Emergency Medicine Residents Kristin Swor Wolf, Suzanne Dooley-Hash University of Michigan, Ann Arbor, MI Background: Procedural competency is a key component of emergency medicine residency training. Residents are required to log procedures to document quantity of procedures and identify potential weaknesses in their training. As emergency medicine evolves, it is likely that the type and number of procedures change over time. Also, exposure to certain rare procedures in residency is not guaranteed. Objectives: We seek to delineate trends in type and volume of core EM procedures over a decade of emergency medicine residents graduating from an accredited four-year training program. Methods: Deidentified procedure logs from 2003–2011 were analyzed to assess trends in type and quantity of procedures. Procedure logs were self-reported by individual residents on a continuous basis during training onto a computer program. Average numbers of procedures per resident in each graduating class were noted. Statistical analysis was performed using SPSS and includes a simple linear regression to evaluate for significant changes in number of procedures over time and an independent samples two-tailed t-test of procedures performed before and after the required resident duty hours change. Results: A total of 112 procedure logs were analyzed and the frequency of 29 different procedures was • www.aemj.org S171 evaluated. A significant increase was seen in one procedure, the venous cutdown. Significant decreases were seen in 12 procedures including key procedures such as central venous catheters, tube thoracostomy, and procedural sedation. The frequency of five high-stakes/ resuscitative procedures, including thoracotomy and cricothyroidotomy, remained steady but very low (<4 per resident over 4 years). Of the remaining 11 procedures, 8 showed a trend toward decreased frequency, while only 5 increased. Conclusion: Over the past 9 years, EM residents in our program have recorded significantly fewer opportunities to perform most procedures. Certain procedures in our emergency medicine training program have remained stable but uncommon over the course of nearly a decade. To ensure competency in uncommon procedures, innovative ways to expose residents to these potentially life saving skills must be considered. These may include practice on high-fidelity simulators, increased exposure to procedures on patients during residency (possibly on off-service rotations), or practice in cadaver and animal labs. 315 A Brief Educational Intervention Effectively Trains Senior Medical Students to Perform Ultrasound-Guided Peripheral Intravenous Access in the Simulation Setting Scott Kurpiel, William Manson, Douglas Ander, Daniel Wood Emory University School of Medicine, Atlanta, GA Background: Studies have shown ultrasound-guided peripheral intravenous access (USGPIV) to be rapid, safe, and effective in patients with difficult IV access. Little data exist on training medical students in USGPIV. Objectives: To study the effectiveness of a unique educational intervention using didactic and hands-on training in USGPIV. We hypothesized that senior medical students would improve performance and confidence with USGPIV after the simulation training. Methods: Fourth year medical students were enrolled in an experimental, prospective, before and after study conducted at a university medical school simulation center. Baseline skills in participant’s USGPIV on simulation vascular phantoms were graded by ultrasound expert faculty using standardized checklists. The primary outcome was time to cannulation, and secondary outcomes were ability to successfully cannulate, number of needle attempts, and needle-tip visualization. Subjects then observed a 15-minute presentation on correct performance of USGPIV followed by a 30-minute hands-on practical session using the vascular simulators with a 1:4 to 1:6 ultrasound instructor to student ratio. An expert blinded to the participant’s initial performance graded post-educational intervention USGPIV ability. Pre- and post-intervention surveys were obtained to evaluate USGPIV confidence, previous experience with ultrasound, peripheral IV access, USGPIV, and satisfaction with the educational format. S172 2012 SAEM ANNUAL MEETING ABSTRACTS Results: 37 subjects (54% men; 26.2 mean age) were enrolled. Time spent during cannulation improved from 70.1 seconds to 33.2 seconds (p = 0.048) with 100% success in cannulation post-intervention (89% pre-intervention). Subjects required fewer needle attempts (p = 0.004) after the educational intervention. Participants reported improved confidence in their ability to perform USGPIV after education from 1.92 to 4.32 (p < 0.001) and a satisfaction of 4.92 with the teaching intervention on a five-point Likert scale. There was no statistically significant difference in needle-tip visualization before and after the intervention. Conclusion: A brief educational intervention improves time to cannulation, number of attempts, and confidence in senior medical students learning USGPIV. Skill retention and translation to patient care remain to be assessed. >25 scans had improved IIS in the FAST and AAA (p < 0.08 for RUQ; p < 0.03 for FAST; p < 0.001 for AAA). IIS trended with PGY for each competency, and significantly for RUQ (p < 0.03, p < 0.27 and p < 0.39, for RUQ, FAST, and AAA). In general, confidence in diagnosis was directly related to accurate image interpretation, and incorrect interpretation was poorly associated with high level of confidence. Conclusion: Competency in interpretation of RUQ and AAA images improved with increased number of scans; the 25-scan threshold did not hold for RUQ. Conversely, PGY year was significant for RUQ alone. This difference may be explained by increased clinical knowledge or other modalities such as didactics or simulation. A numbers only-based guideline for competency assessment in EM US applications is not adequate. 317 316 Factors Predicting EM Resident Ultrasound Competency: Number of Scans or PostGraduate Year? Angela Cirilli1, Alvin Lomibao1, Andrew Loftus1, David Francis2, Jordana Kaban3, Sarah Sartain3, Christine Haines1, Ann Prokofieva1, Mathew Nelson1, Christopher Raio1 1 North Shore-LIJ Health System, Manhasset, NY; 2Stanford University Hospital, Palo Alto, CA; 3University of Missouri-Kansas City School of Medicine, Truman Medical Center, Kansas City, MO Background: Although the ACGME requires EM residents to be competent in emergency ultrasound (US) applications, a standardized curriculum in emergency US, an educational methodology, and a validated competency assessment tool do not exist. Many EM residency programs define a resident US-’’competent’’ upon completing a pre-determined number of scans, commonly 25. Objectives: To assess the numbers-based guideline for competency in three US applications. Methods: This prospective, multi-site study examined EM residents’ ability to interpret right upper quadrant (RUQ), FAST, and aortic US to diagnose aneurysms (AAA). After viewing 20, 1-minute video clips each of RUQ, FAST, and aorta exams, residents were asked two questions: a diagnosis, and confidence in this diagnosis. Image interpretation score (IIS) for these competencies was compared to number of clinical scans in that area completed at time of study, as well as post-graduate year (PGY). Analyses were performed on multiple relationships using Pearson correlations and t-tests. Results: In total, 88 residents participated, representing PGY1-4 (n = 32, 23, 29, 4, respectively) in 3-year and 4year EM residencies. On average, residents had previously completed 15.7 ± 18.9 RUQ, 20.1 ± 23.8 FAST, and 7.9 ± 12.0 AAA scans. Residents’ IIS was 73%, 93%, and 92% for RUQ, FAST, and AAA. For the RUQ and AAA competency, there was a significant association between the number of previous scans completed and IIS (q = 0.283 and p < 0.008, and q = 0.280 and p < 0.009; for FAST, q = 0.193 and p < 0.07). Those with Comparison of Resident Self, Peer, and Faculty Evaluations in a Simulation-based Curriculum Dylan Cooper, Lee Wilbur, Kevin Rodgers, Jennifer Walthall, Katie Pettit, Gretchen Huffman, Carey Chisholm Indiana University School of Medicine, Indianapolis, IN Background: The Accreditation Council of Graduate Medical Education recommends incorporation of 360degree evaluations into residency training. This includes multiple perspectives which can improve evaluation validity and reliability. Objectives: To compare self, peer, and faculty evaluations of residents in a simulation session. Methods: Emergency medicine (EM) residents at Indiana University participate in bimonthly simulation sessions throughout their training. Each session involves several simulation scenarios with one resident functioning as the team leader in each case. Resident team leaders are evaluated with a written evaluation form of competency categories: medical knowledge, clinical competence, patient advocacy/ professionalism, systems-based practice, and interpersonal/ communication skills. Each category is scored on a 1–10 scale: 1–4 (below expected level), 5–7 (expected level), 8–10 (above expected level). For this study, the team leader completed a ‘‘self’’ evaluation, fellow residents observing the case completed a ‘‘peer’’ evaluation, and faculty trainers completed a ‘‘faculty’’ evaluation. Scores were compared based upon evaluation type and year of training using various models: mixed effects regression, fixed effect for evaluation type and year of training, and random effect for the resident being evaluated. Results: There were a total of 76 residents with 146 self, 157 peer, and 174 faculty evaluations completed. All evaluation types differed significantly, with self evaluations significantly lower and peer evaluations significantly higher than faculty evaluations. When level of training was taken into account, EM 1st year residents scored significantly lower than EM 3rd year residents within each type of evaluation. Conclusion: There was no correlation between self, peer, and faculty evaluations in this study. This finding ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 was consistent across all levels of training. Residents scored themselves significantly lower than faculty, while peers rated fellow residents higher than faculty evaluations. This tendency should be taken into account when peer and self evaluations are completed in 360-degree evaluations during graduate medical education. • Do Evaluations Of Residents Change When The Evaluator Becomes Known? Jonathan S. Jones, Bentley W. Curry, Mary J. Johnson University of Mississippi Medical Center, Jackson, MS Background: Effective evaluation of residents requires honest and candid assignment of grades as well as individualized feedback. Truly candid assignment of grades may be dependent on whether the identity of the evaluator is anonymous or made known to the resident. Objectives: This study examines the grade distribution of resident evaluations when the identity of the evaluator was anonymous as compared to when the identity of the evaluator was known to the resident. We hypothesize that there will be no change in the grades assigned to residents. Methods: We retrospectively reviewed all faculty evaluations of residents and grades assigned from July 1, 2008 through November 15, 2011. Prior to July 1, 2010 the identity of the faculty evaluators was anonymous, while after this date, the identity of the faculty evaluators was made known to the residents. Throughout this time period, residents were graded on a five-point scale. Each resident evaluation included grades in the six ACGME core competencies as well as in select other abilities. Specific abilities evaluated varied over the dates analyzed. Evaluations of residents were assigned to two groups, based on whether the evaluator was anonymous or made known to the resident. Grades were compared between the two groups. Results: A total of 10,760 grades were assigned in the anonymous group, with an average grade of 3.90 (95CI 3.88, 3.91). A total of 7,122 grades were assigned in the known group with an average grade of 3.77 (95CI 3.75, 3.79). Specific attention was paid to assignment of unsatisfactory grades (1 or 2 on the five-point scale). The anonymous group assigned 355 grades in this category, comprising 3.3% of all grades assigned. The known group assigned 100 grades in this category, comprising 1.4% of all grades assigned. Unsatisfactory grades were assigned by the anonymous group 1.9% (95CI 1.5, 2.3) more often. Additionally, 5.8% (95CI 3.8, 6.8) fewer exceptional grades (4 or 5 on the five-point scale) were assigned by the anonymous group. Conclusion: The average grade assigned was closer to average (3 on a five-point scale) when the identity of the evaluator was made known to the residents. S173 Additionally, fewer unsatisfactory and exceptional grades were assigned in this group. This decrease of both unsatisfactory and exceptional grades may make it more difficult for program directors to effectively identify struggling and strong residents respectively. 319 318 www.aemj.org Testing to Improve Knowledge Retention from Traditional Didactic Presentations: A Pilot Study David Saloum, Amish Aghera, Brian Gillett Maimonides Medical Center, Brooklyn, NY Background: The ACGME requires an average of at least 5 hours of planned educational experiences each week for EM residents, which traditionally consists of formal lecture based instruction. However, retention by adult learners is limited when presented material in a lecture format. More effective methods such as small group sessions, simulation, and other active learning modalities are time- and resource-intensive and therefore not practical as a primary method of instruction. Thus, the traditional lecture format remains heavily relied upon. Efficient strategies to improve the effectiveness of lectures are needed. Testing utilized as a learning tool to force immediate recall of lecture material is an example of such a strategy. Objectives: To evaluate the effect of immediate postlecture short answer quizzes on EM residents’ retention of lecture content. Methods: In this prospective randomized controlled study, EM residents from a community based 3-year training program were randomized into two groups. Block randomization provided a similar distribution of postgraduate year training levels and performance on both USMLE and in-training examinations between the two groups. Each group received two identical 50-minute lectures on ECG interpretation and aortic disease. One group of residents completed a five-question short answer quiz immediately following each lecture (n = 13), while the other group received the lectures without subsequent quizzes (n = 16). The quizzes were not scored or reviewed with the residents. Two weeks later, retention was assessed by testing both groups with a 20-question multiple choice test (MCT) derived in equal part from each lecture. Mean and median test results were then compared between groups. Statistical significance was determined using a paired t-test of median test scores from each group. Results: Residents who received immediate post-lecture quizzes demonstrated significantly higher MCT scores (mean = 57%, median 58%, n = 10) compared to those receiving lectures alone (mean = 48%, median = 50%, n = 15); p = 0.023. Conclusion: Short answer testing immediately after a traditional didactic lecture improves knowledge retention at a 2-week interval. Limitations of the study are that it is a single center study and long term retention was not assessed. S174 320 2012 SAEM ANNUAL MEETING ABSTRACTS Pediatric Emergency Medicine Core Education Modules: Utility in Asynchronous Learning Michael Preis, Jessica Salzman, Zac Kahler, Sean Fox Carolinas Medical Center, Charlotte, NC Background: The task of educating the next generation of physicians is steadily becoming more difficult with the inherent obstacles that exist for faculty educators and the work hour restrictions that students must adhere to. The obstacles make developing curricula that not only cover important topics but also do so in a fashion that helps support and reinforce the clinical experiences very difficult. Several areas of medical education are using more asynchronous techniques and self-directed online educational modules to overcome these obstacles. Objectives: The aim of this study was to demonstrate that educational information pertaining to core pediatric emergency medicine topics could be as effectively disseminated to medical students via self-directed online educational modules as it could through traditional didactic lectures. Methods: This was a prospective study conducted from August 1, 2010 through December 31, 2010. Students participating in the emergency medicine rotation at Carolinas Medical Center were enrolled and received education in a total of eight core concepts. The students were divided into two groups which changed on a monthly basis. Group 1 was taught four concepts via self-directed online modules and four traditional didactic lectures. Group 2 was taught the same core concepts, but in opposite fashion to Group 1. Each student was given a pre-test, post-test, and survey at the conclusion of the rotation. Results: A total of 28 students participated in the study. Students, regardless of which group assigned, performed similarly on the pre-test, with no statistical difference among scores. When looking at the summative total scores between online and traditional didactic lectures, there was a trend towards significance for more improvement among those taught online. The student’s assessment of the online modules showed that the majority either felt neutral or preferred the online method. The majority thought the depth and length of the modules were perfect. Most students thought having access to the online modules was valuable and all but one stated that they would use them again. Conclusion: This study demonstrates that self-directed, online educational modules are able to convey important concepts in emergency medicine similar to traditional didactics. It is an effective learning technique that offers several advantages to both the educator and student. Table - Abstract 320: Difference in Total Scores Pre- and Post-Test and Overall Improvement in Online vs. Standard Method Pre-Test Post-Test Overall Improvement N Mean Standard Deviation P Score 28 21 21 )0.46 1.33 1.33 3.2 3.0 3.9 0.45 0.05 0.13 321 Asynchronous vs Didactic Education: It’s Too Early To Throw In The Towel On Tradition Jaime Jordan1, Azadeh Jalali2, Samuel Clarke1, Pamela Dyne3, Tahlia Spector2, Sebastian Uijtdehaage2, Wendy Coates1 1 Harbor-UCLA Medical Center, Torrance, CA; 2 David Geffen School of Medicine at UCLA, Los Angeles, CA; 3OliveView-UCLA Medical Center, Sylmar, CA Background: Computer-based asynchronous learning (ASYNCH) is cost-effective, allows self-directed pacing and review, and addresses preferences of millennial learners. Current research suggests there is no significant difference from traditional classroom instruction. Data are limited for novice learners in emergency medicine. Objectives: We compared ASYNCH modules with traditional didactics for rising fourth year medical students during a week-long intensive course in acute care, and hypothesized they would be equivalent. Methods: Prospective observational quasi-experimental study of fourth year medical students who were novice learners with minimal prior exposure to curricular elements. A pre-test assessed baseline knowledge. The acute care curriculum was delivered in either traditional classroom lecture format (shock, acute abdomen, dyspnea, field trauma) or via ASYNCH modules (chest pain, EKG interpretation, pain management, ED trauma). An interactive review of all topics was followed by a posttest. Retention at 10 weeks was also measured. Preand post-test items were written and validated by a panel of medical educators. Mean scores were analyzed using dependent t-test and attitudes were assessed by a five-point Likert scale. Results: 44 of 48 students completed the protocol. Students initially acquired more knowledge from didactic education as demonstrated by mean gain test scores (didactic: 28.39% ± 18.06; asynchronous 9.93% ± 23.22). Mean difference between didactic and ASYNCH = 18.45% with 95% CI 10.40–26.50; p = 0.0001. Retention testing demonstrated similar knowledge attrition: mean gain scores )14.94% (didactic); )17.61% (ASYNCH), which was not significantly different: 2.68% ± 20.85, 95% CI )3.66 to 9.02, p = 0.399. Students had positive attitudes towards ASYNCH; 60.4% believed asynchronous modules were educational. 95.8% enjoyed flexibility. 39.6% preferred asynchronous education for required didactics; 37.5% were neutral; 23% preferred traditional lectures. Conclusion: ASYNCH education was not equivalent to traditional lectures for novice learners of acute care topics. Interactive didactic education was valuable; however, retention rates were similar. Students had mixed attitudes towards ASYNCH. We urge caution in trading in traditional didactic lectures in favor of ASYNCH for novice learners. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 322 Access To Pediatric Equipment And Medications In Critical Access Hospitals: Is A Lack Of Resources A Valid Concern? Jessica Katznelson1, C. Scott Forsythe2, William A. Mills2 1 Johns Hopkins School of Medicine, Baltimore, MD; 2University of North Carolina School of Medicine, Chapel Hill, NC Background: Critical access hospitals (CAH) provide crucial emergency care to rural populations that would otherwise be without ready access to health care. Data show that many CAH do not meet standard adult quality metrics. Adults treated at CAH often have inferior outcomes to comparable patients cared for at other community-based emergency departments (EDs). Similar data do not exist for pediatric patients. Objectives: As part of a pilot project to improve pediatric emergency care at CAH, we sought to determine whether these institutions stock the equipment and medications necessary to treat any ill or injured child who presents to the ED. Methods: Five North Carolina CAH volunteered to participate in an intensive educational program targeting pediatric emergency care. At the initial site visit to each hospital, an investigator, in conjunction with the ED nurse manager, completed a 109-item checklist of commonly required ED equipment and medications based on the 2009 ACEP ‘‘Guidelines for Care of Children in the Emergency Department’’. The list was categorized into monitoring and respiratory equipment, vascular access supplies, fracture and trauma management devices, and specialized kits. If available, adult and pediatric sizes were listed. Only hospitals stocking appropriate pediatric sizes of an item were counted as having that item. The pharmaceutical supply list included antibiotics, antidotes, antiemetics, antiepileptics, intubation and respiratory medications, IV fluids, and miscellaneous drugs not otherwise categorized. Results: Overall, the hospitals reported having 91% of the items listed (range 87–96%). The two greatest deficiencies were fracture devices (range 33–66%), with no hospital stocking infant-sized cervical collars, and antidotes, with no hospital stocking Pralidoxime, 1/5 hospitals stocking Fomepizole, and 2/5 hospitals stocking pyridoxine and methylene blue. Only one of the five institutions had access to Prostaglandin E. The hospitals stated cost and rarity of use as the reason for not stocking these medications. Conclusion: The ability of CAH to care for pediatric patients does not appear to be hampered by a lack of equipment. Ready access to infrequently used, but potentially lifesaving, medications is a concern. Tertiary care centers preparing to accept these patients should be aware of these potential limitations as transport decisions are made. • 323 www.aemj.org S175 PEWS Program In The Pediatric Emergency Department Halves Unanticipated In-hospital Transfers To A Higher Level Of Care For Patients With Respiratory Complaints Patrick Crocker1, Eric Higginbotham1, Diane Taylor1, Ben King2, Fernanda Santos1, Manasi Nabar2, Truman J. Milling2 1 Dell Children’s Medical Center, Austin, TX; 2 University Medical Center at Brackenridge, Austin, TX Background: Pediatric Early Warning Scores (PEWS) were developed to evaluate floor patients for impending clinical deterioration necessitating transfer to a higher level of care, but it has not been integrated into the pediatric emergency department (PED) as a triage tool. Objectives: To measure the effect of a program using PEWS, as an admission level of care triage tool in the PED, on unanticipated transfers to a higher level of care in the first 12 hours after admission, and proportions of level of care admissions. Methods: This is an analysis of the year before and after (2008 and 2009) implementation of a PEWS program for patients being admitted with respiratory complaints to a dedicated children’s hospital. Chief complaints from ED triage were used to identify all patients presenting with respiratory complaints in those two years, and the year groups were divided first by discharge versus admission, then in admitted patients by whether they experienced an unanticipated transfer to a higher level of care (floor to IMC, floor to ICU, IMC to ICU) in the first 12 hours after admission. The z test for differences in proportions was used, and p values less than 0.05 were considered significant. Results: There were 8,021 patients presenting to the PED with respiratory complaints in 2008, the pre-PEWS group (PPG), with 6,823 discharges (85%) and 1,198 admissions (15%). There were 14,199 patients presenting with the same complaints in 2009, the PEWS group (PG), with 12,723 discharges (90%) and 1,476 admission (10%).There were 40 unanticipated transfers in the PPG (25 floor to IMC; 7 floor to ICU, 8 IMC to ICU), and 24 in the PG (16 floor to IMC; 5 floor to ICU; 3 IMC to ICU) (p value less than 0.05). The distribution of the admissions in the PPG (1075 or 90% to floor; 99 or 8% to IMC; 7 or 0.6% to ICU) and the PG (1344 or 91% to floor; 109 or 7% to IMC; 7 or 0.5% to ICU) were not significantly different (p value greater than 0.05), indicating patients in the PG were not routinely overtriaged. Conclusion: Implementation of a PEWS program in the PED using PEWS as a triage tool for in-patient admission nearly halved the number of unanticipated transfers to a higher level of care without significantly over-triaging. S176 324 2012 SAEM ANNUAL MEETING ABSTRACTS Children’s Emergency Department Recidivism in Infancy and Early Childhood Whitney V. Cabey1, Emily MacNeill2, Lindsey N. White3, James Norton2, Alice M. Mitchell2 1 University of Michigan, Ann Arbor, MI; 2 Carolinas Medical Center, Charlotte, NC; 3 Washington Hospital Center, Washington, DC Background: Data defining frequent children’s emergency department (CED) use, or recidivism, is lacking particularly in early childhood. Objectives: To define the threshold and risk factors for CED recidivism in the first 36 months of life in patients treated at an urban tertiary care center. Methods: We conducted a retrospective cross sectional study of children born between 2003 and 2006 who were seen at a single CED (40,000 visits per year) at any time during the first 36 months of life. Exclusion criteria included age greater than 36 months at time of index visit, or residency outside of the CED’s county during the study period. Patients were followed longitudinally for CED utilization in the first 36 months of life. The distribution of data was used to determine the threshold for recidivism, defined by the 90th percentile for frequency of CED visits. Risk factors associated with CED recidivism were identified by multivariate logistic regression analysis. Results: 16,664 patients met the inclusion/exclusion criteria. Data skewed heavily toward patients with no more than two CED visits, with 70% of patients falling within this range (range 1 to 39 visits, IQR 2 visits). The threshold for recidivism, 5 or more visits within the first 36 months of life defines the 90th percentile for frequency of CED visits within the study population and occurred with 14% (95% CI 13 to 15%) of patients. Multivariate analysis identified the following risk-factors: black race (OR 2.8, 95% CI 2.4 to 3.4), Hispanic ethnicity (OR 1.6, 95% CI 1.4 to 1.9), public insurance or lack of insurance (OR 3.5, 95% CI 2.9 to 4.2), residence in a zip code associated with greater than 70% poverty (OR 2.0, 95% CI 1.8 to 2.2), and an index visit in the first 12 months of life (OR 8.4, 95% CI 7.2 to 9.8). Conclusion: CED recidivism, defined by the 90th percentile for CED utilization within the first 36 months of life, was 5 or more visits. Black race, Hispanic ethnicity, public or absence of insurance, living in a zip code associated with greater than 70% poverty, and an index visit in the first 12 months of life were associated with CED recidivism. These findings identify a population for which frequent CED use begins early. Further study is needed to better understand the connection between socioeconomic and sociocultural risk factors for CED recidivism and patterns of health care utilization, and to identify potentially effective interventions. 325 Prevalence of Urinary Tract Infections in Febrile Infants <90 days old with RSV Antonio Muniz Dallas Regional Medical Center, Mesquite, TX Background: The prevalence of significant bacterial infections in infants <90 days old who have RSV is unknown in the era after the introduction of H. influenzae and S. pneumoniae vaccinations. Objectives: The study’s hypothesis is that infants with fever and RSV are at low risk for secondary bacterial infections and may not require extensive and invasive laboratory evaluations. Methods: Prospective evaluation of all infants <90 days old who had sepsis evaluations for fever and an RSV antigen test performed. Data were analyzed using Stata 11 with continuous variables expressed as means, while categorical variables were summarized as frequencies of occurrence and assessed for statistical significance using chi-squared test or Fisher’s exact. Results: There were 360 infants; 140 (38.8%) were RSV (+) and 220 (61.1%) were RSV (-). In the RSV (+) group, the median temperature was 38.6 ± 0.6C, while in the RSV (-) group it was 38.5 ± 0.5C. There was no significant difference in vital signs between groups. There were more coryza and coughing in the RSV (+) group (p < 0.05). There were 16 (11.4%) infants admitted to the PICU in the RSV (+) group versus 4 (1.8%) in the RSV (-). The laboratory evaluation in the RSV (+) included a CBC 130 (92.8%), urinalysis and culture 124 (88.5%), blood culture 126 (90.0%), chest radiograph 102 (72.8%), and CSF fluid analysis 62 (44.2%). In the RSV (-) group, the laboratory evaluation included a CBC 200 (90.9%), urinalysis and urine culture 200 (90.9%), blood culture 204 (92.7%), chest radiograph 100 (45.4%), and CSF fluid analysis 148 (67.2%). There were more chest radiographs performed in the RSV (+) group (p < 0.05), while all other tests were similar. There were 20 (14.2%) significant infections in the RSV (+) group versus 24 (10.9%) in the RSV (-) group. In the RSV (+) group there were 14 (10%) pneumonias, 6 (4.2%) UTI. There were no significant (+) blood cultures in the RSV (+) group, but there were 12 (8.5%) contaminants. In the RSV (-) group there were 12 (5.4%) pneumonias, 8 (3.6%) UTI, 2 (0.9%) bacteremia with group B streptococcus, and 2 (0.9%) meningitis with enterococcus. Conclusion: The risk of serious bacterial infection is low in the RSV (+) group, especially if one excludes pneumonias. There were no cases of bacteremia or meningitis, however UTIs still are significant. Therefore, it may be prudent to exclude an UTI in febrile infants with a (+) RSV antigen test. 326 Impact of an English-Based Pediatric Software on Physician Decision Making: A Multicenter Study in Vietnam Michelle Lin1, Trevor N. Brooks2, Alex C. Miller2, Jamie L. Sharp2, Le Thanh Hai3, Tu Nguyen3, Daniel R. Kievlan1, Ronald A. Dieckmann1 1 University of California, San Francisco, San Francisco, CA; 2KidsCareEverywhere, Oakland, CA; 3National Hospital of Pediatrics, Hanoi, Viet Nam Background: Global health agencies and the Vietnam Ministry of Health have identified pediatric health care and universal health information technology access as high priority goals. Medical decision support software ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 (MDSS) provides physicians with rapid access to current, evidence-based literature, but there are limited data about its effectiveness in developing countries. Objectives: We hypothesized that Vietnamese physicians, with only brief software training, can understand and effectively apply MDSS information to answer clinical questions about pediatric emergencies. Methods: This multicenter, prospective, pretest-posttest crossover study was conducted in 11 Vietnamese hospitals during November 2010-April 2011. A convenience sample of physicians volunteered to attend software training on a pediatric MDSS (PEMSoft; Brisbane, Australia) in English. Two 15-question, multiple-choice exams (Test A and B) were administered before and after a brief 80-minute software training session in Vietnamese and English. Participants who received Test A as a pretest received Test B as a posttest, and vice versa. The primary outcome measure was the difference between pretest and posttest scores, as calculated by the difference in two means using the NewcombeWilson method without continuity correction. Results: For the 208 physician participants, the mean pretest, posttest, and improvement scores were 38% (95% CI: 36–40%), 70% (95% CI: 69–72%), and 32% (95% CI: 30–35%), respectively, after the brief software training session. Subgroup analysis of physician practice setting, years of clinical experience, and comfort level with written English and computer proficiency showed no association. Conclusion: Vietnamese physicians can effectively use a pediatric MDSS (PEMSoft) written in English, after receiving minimal utilization training, for pediatric emergency care decisions on written tests. MDSS technologies may offer a highly scalable, sustainable, and potentially transformative tool in resource-poor environments. • Ability of Acutely Ill Children with Asthma to Perform Acceptable Forced Expiratory Maneuvers in the Emergency Department Christopher A. Hollweg1, Jennifer Eng-Lunt2, Maria T. Santiago2, Jahn T. Avarello1, Megan McCullough1, Alvin Lomibao1, Mae F. Ward1, Robert A. Silverman1 1 NSLIJ Emergency Department, Long Island Jewish Medical Ctr, New Hyde Park, NY; 2 Steven and Alexandra Cohen Children’s Medical Center, New Hyde Park, NY Background: Treatment guidelines recommend that objective measures of pulmonary function be used to assess pulmonary function in children with acute asthma. Although FEV1 is considered the best noninvasive measure of obstruction, in practice it is not often measured in the emergency department, in part because of the perception that good performance is not achievable in the acute care setting. Objectives: Our goal is to determine if acceptable and reproducible FEV1 measurements can be obtained in acutely ill asthmatic children. Methods: A convenience sample of asthmatic children was taken from an academic ED of a pediatric hospital S177 with an annual census of 35,000 visits. Participants, ages 6–18 years, presenting with shortness of breath had spirometry performed by a trained research associate. Those with developmental or behavioral diagnoses were excluded. The first set of efforts was obtained on ED arrival depending on research associate availability; if not, then after treatment was given. A minimum of two efforts were performed at each time point. Acceptability for a time point was defined as two or more efforts with time to peak <120ms, back extrapolated volume <5%, and expiratory time >1s. Reproducibility was defined as the two best acceptable FEV1 efforts within 10% of each other. Chi-square tests were used for comparison; for multiple visits only the first data set was included in the analysis. Results: 61 children were eligible with one child too ill to participate. The average age was 10 years (SD 3.1; range 6–18); 45% were female; and 32% required hospitalization. At the first testing time point, 53/60 (88%) children met acceptability criteria and 49/60 (82%) met reproducibility criteria. Reproducibility in younger children (£10 years) trended lower compared to older children (76% vs. 95%, p = 0.08). Children with more severe obstruction (FEV1 £ 50% predicted) less frequently met reproducibility criteria than children with higher FEV1% predicted (67% vs. 90%, p = 0.03). Performance was similar when FEV1 was assessed at subsequent time points (data not shown). Conclusion: In this pilot study the majority of children with acute asthma in an ED setting obtained acceptable and reproducible FEV1 measurements. FEV1 determination should be considered as an option to objectively assess airway obstruction in acutely ill children. 328 327 www.aemj.org Antibiotics For The Treatment Of Abscesses: A Meta-analysis Jahan Fahimi1, Amandeep Singh2, Bradley Frazee2 1 Alameda County Medical Center - Highland Hospital; University of California, San Francisco, Oakland/San Francisco, CA; 2 Alameda County Medical Center - Highland Hospital, Oakland, CA Background: While incision and drainage (I&D) alone has been the mainstay of management of uncomplicated abscesses for decades, some advocate for adjunct antibiotic use, arguing that available trials are underpowered and that antibiotics reduce treatment failures and recurrence. Objectives: To investigate the role of antibiotics in addition to I&D in reducing treatment failure as compared to management with I&D alone. Methods: We performed a search using MEDLINE, EMBASE, Web of Knowledge, and Google Scholar databases (with a medical librarian) to include trials and observational studies analyzing the effect of antibiotics in human subjects with skin and soft-tissue abscesses. Two investigators independently reviewed all the records. We performed three overlapping meta-analy- S178 2012 SAEM ANNUAL MEETING ABSTRACTS ses: 1. Only randomized trials comparing antibiotics to placebo on improvement of the abscess during standard follow-up. 2. Trials and observational studies comparing appropriate antibiotics to placebo, no antibiotics, or inappropriate antibiotics (as gauged by wound culture) on improvement during standard follow-up. 3. Only trials, but broadened outcome to include recurrence or new lesions during a longer follow-up period as treatment failure. We report pooled risk ratios (RR) using a fixed-effects model for our point estimates with Shore-adjusted 95% confidence intervals (CI). Results: We screened 1,937 records, of which 12 studies fit inclusion criteria, 9 of which were meta-analyzed (5 trials, 4 observational studies) because they reported results that could be pooled. Of the 9 studies, 5 enrolled subjects from the ED, 2 from a soft-tissue infection clinic, and 2 from a general hospital without definition of enrollment site. Five studies enrolled primarily adults, 3 pediatrics, and 1 without specification of ages. After pooling results for all randomized trials only, the RR = 1.03 (95% CI: 0.97–1.08). Exposure being ‘‘appropriate’’ antibiotics (using trials and observational studies) resulted in a pooled RR = 1.01 (95% CI: 0.98–1.03). When we broadened our treatment failure criteria to include recurrence or new lesions at longer lengths of follow-up (trials only), we noted a RR = 1.05 (95% CI: 0.97–1.15). Conclusion: Based on available literature pooled for this analysis, there is no evidence to suggest any benefit from antibiotics in addition to I&D in the treatment of skin and soft tissue abscesses. (Originally submitted as a ‘‘late-breaker.’’) 329 Primary Versus Secondary Closure of Cutaneous Abscesses in the Emergency Department: A RCT Adam J. Singer, Breena R. Taira, Stuart Chale, Anna Domingo, Gillian Schmidt Stony Brook University, Stony Brook, NY Background: Cutaneous abscesses are common and have traditionally been treated with incision and drainage (I&D) and left to heal by secondary intention. Prior studies outside the emergency department (ED) and the US have shown faster healing and comparable low recurrent rates when drained abscesses are closed primarily. Objectives: To compare wound healing and recurrence rates after primary vs. secondary closure of drained abscesses. We hypothesized the percentage of drained ED abscesses that would be completely healed at 7 days would be higher after primary closure. Methods: This randomized clinical trial was undertaken in two academic emergency departments. Immunocompetent adult patients with simple, localized cutaneous abscesses were randomly assigned to I & D followed by primary or secondary closure. Randomization was balanced by center, with an allocation sequence based on a block size of four, generated by a computer random number generator. The primary outcome was percentage of healed wounds seven days after drainage. A sample of 50 patients had 80% power to detect an absolute difference of 40% in healing rates assuming a baseline rate of 25%. All analyses were by intention to treat. Results: Twenty-seven patients were allocated to primary and 29 to secondary closure, of whom 23 and 27, respectively, were followed to study completion. Healing rates at seven days were similar between the primary and secondary closure groups (69.6% [95% CI 49.1 to 84.4] vs. 59.3% [95% CI 40.7 to 75.5]; difference 10.3% [95% CI )15.8 to 34.1]). The rate of abscess recurrence at 7 days was also similar between the primary and secondary closure groups (4.3% [95% CI 0.8 to 21.0] vs. 14.3% [95% CI 5.7 to 31.5]; difference )9.9% [95% CI )27.5 to 8.8]). Conclusion: The rates of wound healing and abscess recurrence seven days after I&D of simple abscesses are similar with either primary or secondary closure. 330 Low Incidence of MRSA Colonization Among House Officers Philip Giordano, Kurt Weber, Rachel Semmons, Josef Thundiyil, Jay Ladde, Jordan Rogers, Brian Batt Orlando Regional Medical Center, Orlando, FL Background: Due to the length of time spent in hospitals, house officers may be at high risk for becoming colonized with MRSA and could potentially act as vectors for transmission. The risk of MRSA conversion by house officers has not been widely studied. Objectives: We sought to determine the incidence of MRSA conversion among house officers after 18 months of training. Methods: We conducted a prospective cohort study at an urban, tertiary-care teaching hospital over an 18month period. The study population included all incoming residents, including emergency medicine, internal medicine, obstetrics/gynecology, general surgery, orthopedics, and pediatrics. Subjects with an active skin infection were excluded. Trained investigators sampled bilateral nares of each subject using standard Dacron swabs. Swabs were plated onto BBL CHROMagar MRSA plates (Becton Dickinson) and incubated for 48 hours. Cultures were interpreted by certified microbiologists and recorded at 24 and 48 hours. Subjects had baseline samples collected during their orientation in July 2010 and subsequent testing in December 2011. Incidence of MRSA conversion was determined and reported with 95% confidence intervals using z statisitcs. Results: In July 2010, 64 out of 70 incoming residents consented to participate. Baseline cultures were positive in 0 of 64 samples. Follow up cultures were obtained 18 months later on 45 (70%) subjects. Only one of these was positive. Of the 19 residents lost to follow-up, 15 had left the institution. The incidence of MRSA conversion was 2.2% (95% CI )2.1, 6.5). This contrasts with the 50% MRSA positivity rate for all hospital cultures during the study period. Conclusion: We had a low baseline prevalence of MRSA colonization when compared to reported rates ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 for house officers and other health care workers. Our results suggest a low risk to house officers of becoming colonized with MRSA at our institution. This is consistent with published data from a different geographic location. 331 Do Physical Exam Findings Correlate with the Presence of Fever in Patients with Skin and Soft Tissue Infections? Jillian Mongelluzo, Brian Tu, Jonathan Fortman, Robert Rodriguez University of California San Francisco, San Francisco, CA Background: Skin and soft tissue infections (SSTI) are common reasons for presentation to acute care facilities and admission to inpatient hospital facilities. Few investigations have actually been conducted to understand the relationship between clinical examination findings and the occurrence of fever. Objectives: Given the effect that the presence of fever has on treatment and admission to hospital decisions, the objective of our study was to determine whether physical exam findings and laboratory tests with regards to SSTIs are associated with the development of fever. Methods: We conducted a prospective observational study at an urban county trauma center, from June 2007 until October 2011, enrolling adults >18 years who presented to the ED for evaluation of SSTIs. Treating providers prospectively completed a data sheet (with an attached tape measure for accuracy) recording demographic information (intravenous drug use (IDU), HIV, diabetes), area of erythema (cm-sq), location, presence of adenopathy, streaking, necrosis, bullae, and joint involvement. The highest temperature within the first 6 hours of ED presentation was recorded. Fever was defined as temperature ‡ 38 C. Enrolled subjects were followed through ED and hospital stays to determine laboratory, radiologic, microbiologic, and admission data. Results: Of 734 patients enrolled, 96 (13%) were febrile. Febrile patients were found to have a significantly larger area of erythema compared to those patients who were afebrile (Kruskal-Wallis test p = 0.0001). Patients with upper extremity infections were more likely to be febrile when compared to patients with infections in other locations (chi-square = 12.8, p < 0.0001). A leukocytosis was significantly more common in patients who were febrile than in afebrile patients (chi-square = 27.2, p < 0.0001). IDU patients were significantly more likely to be febrile than non-IDU patients (chi-square = 9.2, p = 0.0025). Conclusion: Fever is uncommon in patients presenting to the ED for evaluation of suspected SSTI. Larger area of erythema, location of infection, leukocytosis, and history of IDU are all correlated with the presence of fever. Understanding of these clinical factors that are associated with higher rates of fever may aid in development of treatment and admission decision guidelines. • www.aemj.org S179 Table - Abstract 331: Characteristics of patients with and without fever All Patients Febrile Afebrile Male 45 years (IQR 35–55) 529 (72%) IDU 211 (29%) 44 years (IQR 36–53) 71/529 (13%) 36/211 (17%) 15/124 (12%) 45 (IQR 35–55) 457/529 (86%) 175/211 (83%) 109/124 (88%) 50/60 (83%) 63/68 (93%) Age (years) Diabetes 124 (17%) mellitus HIV 60 (8%) Face/neck 68 (9%) infection Area of erythema Mean 159 cm-sq Median 32 cm-sq Upper 197 (24%) extremity infection Leukocytosis 151/337* (45%) 10/60 (17%) 5/68 (7%) 266 cm-sq 80 cm-sq 40/197 (20%) 143 cm-sq 30 cm-sq 157/197 (80%) 48/151 (32%) 103/151 (68%) *Total CBCs sent 332 Preliminary Clinical Feasibility of an Improved Blood Culture Time To Detection Using a Novel Viability Protein Linked PCR Assay Enabling Universal Detection of Viable BSI Hematopathogens Three-fold Earlier Than The Gold standard Jason Rubino1, John Morrison1, Zweitzig Daniel2, Nichol Riccardello2, Bruce Sodowich2, Jennifer Axelband1, Rebecca Jeanmonod1, Mark Kopnitsky2, S. Mark O’Hara2 1 St. Luke’s Hospital and Health Network, Bethlehem, PA; 2Zeus Scientific, Branchburg, NJ Background: The diagnosis of blood stream infection (BSI) is usually made on clinical grounds due to the length of time to detection needed to obtain blood culture (BC) results, the gold standard test. Objectives: Zeus Scientific Inc. has developed a viability protein linked-PCR (VP-PCR) assay that detects hematopathogens in model systems three times faster than BC. The aim of this study was to validate this novel test with suspected BSI patients in a clinical setting. Methods: This prospective cohort study was performed at a Level I community trauma center with an annual census of 75,000. It was reviewed and approved by the IRB. After informed consent, a convenience sample of patients with suspected BSI were enrolled. For each enrolled patient, routine hospital BC were obtained (4 BC bottles: 2 aerobic and 2 anaerobic). For the Zeus (Z) VP-PCR test, one additional aerobic BC bottle was obtained. The single Z BC bottle was blind coded, refrigerated, and twice weekly transported 50 miles for independent incubation and VP-PCR time course testing. VP-PCR test was performed on 1 ml time course aliquots from the single Z BC bottle. Patients’ hospital BC laboratory results were decoded and compared to the Z BC and VP-PCR results. Sensitivity and specificity of Z VP-PCR was determined as compared to the gold S180 2012 SAEM ANNUAL MEETING ABSTRACTS standard of hospital lab BC results as well as to Z BC results. Results: Preliminary data from 223 patients were analyzed comparing hospital BC to Z VP-PCR as well as comparing Z BC to Z VP-PCR. There were a total of 23/ 223 (10.3%) gold standard positive hospital BC. Z VP-PCR performed with a sensitivity of 88% and a specificity of 99% as compared to hospital BC. Z VP-PCR performed with a sensitivity of 100% and a specificity of 100% when compared to Z BC. VP-PCR positives detected BSI three-fold earlier than any of its five corresponding BC bottle incubator flip times. Conclusion: Z VP-PCR test provides early detection of BSI with a high rate of sensitivity and specificity. Discordance in Z BC and hospital BC may be secondary to specimen handling, but requires further study. With successful validation, Z VP-PCR should be studied further to determine if its use improves patient outcomes. 333 Operating Characteristics of History, Physical, and Urinalysis for the Diagnosis of Urinary Tract Infections in Adult Women Presenting to the Emergency Department: An Evidence-Based Review Lisa Meister, Diane Scheer, Eric J. Morley, Richard Sinert SUNY Downstate, Brooklyn, NY Background: Women often present to the ED with symptoms suggestive of urinary tract infections (UTI). History, physical exam, and laboratory studies are often used to diagnosis UTI. Objectives: To perform a systematic review of the literature to determine the utility of history, physical, and urinalysis in diagnosing uncomplicated female UTI. Methods: The medical literature was searched from January 1965 to November 2011 in PUBMED and EMBASE using a strategy derived from the following PICO formulation of our clinical question: Patients: Females greater than 18 years in the ED suspected of having a UTI. Interventions: Historical, physical exam and laboratory findings commonly used to diagnose a UTI. Comparator: UTI was defined as a urine culture with greater than 100,000 colonies of bacteria per ml of urine. Outcome: The operating characteristics of the interventions in diagnosing a UTI. Studies were assessed using the Quality Assessment Tool for Diagnostic Accuracy Studies. Data analysis was performed using Meta-DiSc and a random-effects model. Results: Four studies with a total of 670 patients were included for review. The prevalence of UTI varied from 40% to 67%. History of (previous UTI, dysuria, urgency, back pain, abdominal pain, fever, and hematuria) and physical exam findings (temperature >37.2 C, and costovertebral angle tenderness) had positive likelihood ratios (LE+) from 2.2–0.8 and negative likelihood ratios (LR-) from 0.7–1.0. On rine test strip analysis (leukocyte esterase (LE), nitrite (N), blood (B), and protein (P)) only a positive nitrite reaction (LR+ = 7.5–24.5) was useful to rule in a UTI. To rule out a UTI only the absence of any leukocyte esterase or blood reaction were significantly robust (LR- 0.2–0.4). When the LE, N, and B were all non-reactive the LR- is 0.1. Combining the elements of the urine test strip analysis did not significantly improve the LR+. Microscopic urine analysis showed that the presence of any WBC had the highest LR+ (5.0) and the absence of both RBC and WBC had the best LR- (0.08–0.2). Conclusion: No single historical or physical exam finding can rule out or rule in UTI. Positive nitrite and the presence of any WBCs were good markers of a UTI. The combination of a negative LE, N, and B or the absence of any RBC or WBC can rule out UTI in patients with low pretest probability. 334 Comparison of 64 and 320 Slice Coronary Computed Tomography Angiography Adam J. Singer, Summer J. Ferraro, Alexander Abamowicz, Peter Viccellio, Michael Poon Stony Brook University, Stony Brook, NY Background: Many EDs have recently introduced coronary computed tomography angiography (CCTA) for the evaluation of low to intermediate risk chest pain. Objectives: We compared radiation doses and intravenous (IV) contrast volumes using 64 and 320 slice CCTA in ED patients with chest pain. We hypothesized that 320 slice CCTA would reduce radiation doses and IV contrast volumes. Methods: Study Design-Prospective, observational. Setting-Suburban, academic ED with annual census of 90,000 and dedicated 64 and 320 slice CTs. Subjects-ED patients with low-to-intermediate risk chest pain undergoing CCTA to exclude CAD. We compared 100 consecutive patients each scanned on the 64 or 320 slice CCTA in 2010–2011. Measures and outcomes-Data were prospectively collected using standardized data collection forms required prior to performing CCTA. The main outcomes were cumulative radiation doses and volumes of intravenous contrast. Data AnalysisGroups compared with t-, Mann Whitney U, and chi-square tests. Results: The mean age of patients imaged with the 64 and 320 scanners were 49 (SD 10) vs. 51 (13) (P = 0.27). Male:female ratios were also similar (57:43 vs. 51:49 respectively, P = 0.40). Both mean (P < 0.001) and median (P = 0.006) effective radiation dose were significantly lower with the 320 (6.8 and 6 mSv) vs. the 64-slice scanner (12.2 and 10 mSv) respectively. Prospective gating was successful in 100% of the 320 scans and only in 38% of the 64 scans (P < 0.001). Mean IV contrast volumes were also lower for the 320 vs. the 64-slice scanner (74 ± 10 vs. 96 ± 12 ml; P < 0.001). The % non-diagnostic scans was similarly low in both scanners (3% each). There were no differences in use of beta-blockers or nitrates. Conclusion: When compared with the 64-slice scanner, the 320-slice scanner reduces the effective radiation doses and IV contrast volumes in ED patients with CP undergoing CCTA. Need for beta-blockers and nitrates was similar and both scanners achieved excellent diagnostic image quality. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 335 Use of a Novel Ambulatory Cardiac Monitor to Detect Arrhythmias in Discharged ED Patients Ayesha Sattar1, Dorian Drigalla2, Steven Higgins3, Donald Schreiber1 1 Stanford University School of Medicine, Stanford, CA; 2Scott & White Memorial Hospital, Texas A&M College of Medicine, Temple, TX; 3Scripps Memorial Hospital, La Jolla, CA Background: Emergency department (ED) patients who present with symptoms that may be due to arrhythmias (ARR), such as palpitations or syncope, may be discharged after a negative evaluation and referred for ambulatory cardiac monitoring. This approach is costly and ineffective due to poor follow-up and compliance. Objectives: To assess the utility of a novel, single-use outpatient ambulatory cardiac monitoring patch applied in the ED at discharge to detect arrhythmias. Methods: Between February and October 2011, a continuous recording patch (Zio Patch - iRhythm Technologies, Inc.) was applied at discharge on a consecutive sample of ED patients with suspected ARR. Patients could wear the patch for up to 14 days and press the integrated marker button when symptomatic. Devices were mailed back for analysis for any significant ARR defined as: ventricular tachycardia (VT), SVT, paroxysmal atrial fibrillation (PAF), ‡3 sec pause, 2nd degree Mobitz II or 3rd degree AV block, or symptomatic bradycardia. Descriptive statistics and t-tests with 95% confidence intervals (95%CI) were used for analysis. Results: 135 patients - 65 males (48%), mean age 48.6 were enrolled and none were lost to follow-up. Palpitations (30%) or syncope (18%) were common indications. Average device wear time was 6.1 days (95%CI 5.8–6.4; max 14 days). 51 (38%) had ‡1 significant ARR and 7 (13.7%) were symptomatic at the time. Average time to first ARR episode was 1.8 days (95%CI 1.6–2.0; max 9.8 days) and first symptomatic ARR 2.1 days (95%CI 1.8–2.4; max 8.6 days). 44 SVT, 5 PAF, 3 VT, and 1 AV block were detected. Women • www.aemj.org S181 and patients with palpitations required significant longer monitoring time to detect ARR (Table 1). 81 symptomatic patients (60%) did not have any significant ARR. No patients who presented with syncope and discharged from the ED were found to have a significant ARR. Conclusion: The Zio Patch is a novel, single-use ambulatory cardiac monitor that successfully facilitates ARR diagnosis upon ED discharge. Its ease-of-use and excellent (100%) compliance detected symptomatic and asymptomatic ARR over 14 days. Patients with palpitations and women required longer wear times to detect ARR. The device ruled out arrhythmias in 60% of symptomatic patients. The potential for documenting clinical arrhythmias over a longer term than the standard 24–48 hour Holter monitor may have clinical benefit. 336 Inferior Vena Cava To Aorta Ratio Has Limited Value For Assessing Dehydration In Young Pediatric Emergency Department Patients Molly Theissen, Jonathan Theoret, Michael M. Liao, John Kendall Denver Health Medical Center, Denver, CO Background: A few studies have demonstrated that bedside ultrasound measurement of inferior vena cava to aorta (IVC-to-Ao) ratio is associated with the level of dehydration in pediatric patients and a proposed cutoff of 0.8 has been suggested, below which a patient is considered dehydrated. Objectives: We sought to externally validate the ability of IVC-to-Ao ratio to discriminate dehydration and the proposed cutoff of 0.8 in an urban pediatric emergency department (ED). Methods: This was a prospective observational study at an urban pediatric ED. We included patients aged 3 to 60 months with clinical suspicion of dehydration by the ED physician and an equal number of control patients with no clinical suspicion of dehydration. We excluded children who were hemodynamically unstable, had chronic malnutrition or failure to thrive, open abdominal wounds, or were unable to provide patient or parental consent. A validated clinical dehydration score (CDS) (range 0 to 8) was used to measure initial dehydration status. An experienced sonographer blinded to the CDS and not involved in the patient’s care measured the IVC-to-Ao ratio on the patient prior to any hydration. CDS was collapsed into a binary outcome of no dehydration or any level of dehydration (1 or higher). The ability of IVC-to-Ao ratio to discriminate dehydration was assessed using area under the receiver operating characteristic curve (AUC) and the sensitivity and specificity of IVC-to-Ao ratio was calculated for three cutoffs (0.6, 0.8, 1.0). Calculation of AUC was repeated after adjusting for age and sex. Results: 92 patients were enrolled, 39 (42%) of whom had a CDS of 1 or higher. Median age was 28 (interquartile range 16–39) months, and 53 (58%) were female. The IVCto-Ao ratio showed an unadjusted AUC of 0.66 (95% CI 0.54–0.77) and adjusted AUC of 0.67 (95% CI 0.56–0.79). For a cutoff of 0.6 sensitivity was 26% (95% CI 13%–42%) S182 2012 SAEM ANNUAL MEETING ABSTRACTS and specificity 92% (95% CI 82%–98%); for a cutoff of 0.8 sensitivity was 51% (95% CI 35%–68%) and specificity 74% (95% CI 60%–85%); for a cutoff of 1.0 sensitivity was 79% (95% CI 64%–91%) and specificity 40% (95% CI 26%–54%). Conclusion: The ability of the IVC-to-Ao ratio to discriminate dehydration in young pediatric ED patients was modest and the cutoff of 0.8 was neither sensitive nor specific. 337 Multicenter Randomized Comparative Effectiveness Trial of Cardiac CT vs Alternative Triage Strategies in Acute Chest Pain Patients in the Emergency Department: Results from the ROMICAT II Trial Udo Hoffmann1, Quynh A. Truong1, Hang Lee1, Eric T. Chou2, Shant Kalanjian2, Pamela Woodard3, John T. Nagurney1, James H. Pope4, Thomas Hauser5, Charles White6, Mike Mullens3, Nathan I. Shapiro5, Michael Bond6, Scott Weiner7, Pearl Zakroysky1, Douglas Hayden1, Stephen D. Wiviott8, Jerome Fleg9, David Schoenfeld1, James Udelson7 1 Massachusetts General Hospital, Boston, MA; 2 Kaiser Permanente Fontana, Fontana, CA; 3 Washington University School of Medicine, St. 4 Louis, MO; Baystate Medical Center, Springfield, MA; 5Beth Israel Deaconess Medical Center, Boston, MA; 6University of Maryland, Baltimore, Baltimore, MD; 7Tufts Medical Center, Boston, MA; 8Brigham and Women’s Hospital, Boston, MA; 9NHLBI, Bethesda, MD Background: While early cardiac computed tomographic angiography (CCTA) could be more effective to manage emergency department (ED) patients with acute chest pain and intermediate (>4%) risk of acute coronary syndrome (ACS) than current management strategies, it also could result in increased testing, cost, and radiation exposure. Objectives: The purpose of the study was to determine whether incorporation of CCTA early in the ED evaluation process leads to more efficient management and earlier discharge than usual care in patients with acute chest pain at intermediate risk for ACS. Methods: Randomized comparative effectiveness trial enrolling patients between 40–75 years of age without known CAD, presenting to the ED with chest pain but without ischemic ECG changes or elevated initial troponin and require further risk stratification for decision making, at nine US sites. Patients are being randomized to either CCTA as the first diagnostic test or to usual care, which could include no testing or functional testing such as exercise ECG, stress SPECT, and stress echo following serial biomarkers. Test results were provided to physicians but management in neither arm was driven by a study protocol. Data on time, diagnostic testing, and cost of index hospitalization, and the following 28 days are being collected. The primary endpoint is length of hospital stay (LOS). The trial is powered to allow for detection of a difference in LOS of 10.1 hours between competing strategies with 95% power assuming that 70% of projected LOS values are true. Secondary endpoints are cumulative radiation exposure, and cost of competing strategies. Tertiary endpoints are institutional, caregiver, and patient characteristics associated with primary and secondary outcomes. Rate of missed ACS within 28 days is the safety endpoint. Results: As of November 21st, 2011, 880 of 1000 patients have been enrolled (mean age: 54 ± 8, 46.5% female, ACS rate 7.55%). The anticipated completion of the last patient visit is 02/28/12 and the database will be locked in early March 2012. We will present the results of the primary, secondary, and some tertiary endpoints for the entire cohort. Conclusion: ROMICAT II will provide rigorous data on whether incorporation of CCTA early in the ED evaluation process leads to more efficient management and triage than usual care in patients with acute chest pain at intermediate risk for ACS. (Originally submitted as a ‘‘late-breaker.’’) 338 Meta-analysis Of Magnetic Resonance Imaging For The Diagnosis Of Appendicitis Michael D. Repplinger, James E. Svenson, Scott B. Reeder University of Wisconsin School of Medicine and Public Health, Madison, WI Background: The diagnosis of appendicitis is increasingly reliant on imaging confirmation. While CT is commonly used, it exposes patients to ionizing radiation, which increases lifetime risk of cancer. MRI has been evaluated in small studies as an alternative. Objectives: This study aims to perform a meta-analysis on all published studies since 2005 evaluating the use of MRI to diagnose appendicitis. Calculated measures include sensitivity, specificity, positive predictive value, and negative predictive value. Methods: This is a meta-analysis of all studies that evaluated the use of MRI for diagnosing appendicitis. All retrospective and prospective studies published in English and listed in PubMed since 2005 were included. Earlier studies were excluded in order to only report on modern imaging sequences. Studies were also excluded if a gold standard was not explicitly stated or if raw numbers were not provided to calculate the study outcomes. Data were abstracted by one investigator and confirmed by another. This included absolute number of true positives, true negatives, false positives, false negatives, number of equivocal cases, type of MRI sequence, and demographic data including study setting and sex distribution. Data were analyzed using Microsoft Excel. Results: There were 11 studies which met inclusion and exclusion criteria. A total of 626 subjects from six different countries were enrolled in these trials, 445 (70.1%) of whom were women. A majority of participants were pregnant. Nearly all studies reported routine use of unenhanced MRI with optional use of contrast-enhanced imaging or diffusion weighted imaging. Of the total cohort, 33 subjects (5%, 95% CI 4–6%) had equivocal imaging results. Sensitivity, specificity, positive predictive value, and negative predictive value were 96.9% (95%CI 95.6–98.2%), 96.7% (95%CI 95.8–97.6%), 94.8% (95%CI 93.5–96.1%), and 98% (95%CI 97.4–98.6%). ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Conclusion: The diagnosis of appendicitis appears to be readily assessed by MRI. Using this imaging modality instead of CT will decrease the overall burden of ionizing radiation from medical imaging. 339 Physician Variability In Positive Diagnostic Yield Of Advanced Radiography To Diagnose Pulmonary Embolus In Four Hospitals: 2006–2009 Dana Kindermann1, Jordan Sax2, Kevin Maloy3, Dave Milzman3, Jesse Pines4 1 Georgetown University Hospital / Washington Hospital Center, Washington, DC; 2Johns Hopkins Hospital, Baltimore, MD; 3Washington Hospital Center, Washington, DC; 4George Washington University Hospital, Washington, DC Background: Many studies have documented higher rates of advanced radiography utilization across U.S. emergency departments (EDs) in recent years, with an associated decrease in diagnostic yield (positive tests / total tests). Provider-to-provider variability in diagnostic yield has not been well studied, nor have the factors that may explain these differences in clinical practice. Objectives: We assessed the physician-level predictors of diagnostic yield using advanced radiography to diagnose pulmonary embolus (PE) in the ED, including demographics and D-dimer ordering rates. Methods: We conducted a retrospective chart review of all ED patients who had a CT chest or V/Q scan ordered to rule out PE from 1/06 to 12/09 in four hospitals in the Medstar health system. Attending physicians were included in the study if they had ordered 50 or more scans over the study period. The result of each CT and VQ scan was recorded as positive, negative, or indeterminate, and the identity of the ordering physician was also recorded. Data on provider sex, residency type (EM or other), and year of residency completion were collected. Each provider’s positive diagnostic yield was calculated, and logistic regression analysis was done to assess correlation between positive scans and provider characteristics. Results: During the study period, 15,015 scans (13,571 CTs and 1,443 V/Qs) were ordered by 93 providers. The physicians were an average of 9.7 years from residency, 36% were female, and 98% were EM-trained. Diagnostic yield varied significantly among physicians (p < 0.001), and ranged from 0% to 18%. The median diagnostic yield was 5.9% (IQR 3.8%–7.8%). The use of D-dimer by provider also varied significantly from 4% to 48% (p < 0.001). The odds of a positive test were significantly lower among providers less than 10 years out of residency graduation (OR 0.80, CI 0.68–0.95) after controlling for provider sex, type of residency training, D-dimer use, and total number of scans ordered. Conclusion: We found significant provider variability in diagnostic yield for PE and use of D-dimer in this study population, with 25% of providers having diagnostic yield less than or equal to 3.8%. Providers who were more recently graduated from residency appear to have a lower diagnostic yield, suggesting a more conservative approach in this group. • 340 www.aemj.org S183 Association Between the ‘‘Seat Belt Sign’’ and Intra-abdominal Injury in Children with Blunt Torso Trauma in Motor Vehicle Collisions Dominic Borgialli1, Angela Ellison2, Peter Ehrlich3, Bema Bonsu4, Jay Menaker5, David Wisner6, Shireen Atabaki7, Michelle Miskin8, Peter Sokolove6, Kathy Lillis9, Nathan Kuppermann6, James Holmes6, and the PECARN IAI Study Group10 1 University of Michigan School of Medicine and Hurley Medical Center, Flint, MI; 2 University of Pennsylvania School of Medicine, Philadelphia, PA; 3University of Michigan School of Medicine, Ann Arbor, MI; 4 Nationwide Children’s Hospital, Columbus, OH; 5University of Maryland Medical Center, Shock Trauma, Baltimore, MD; 6UC Davis School of Medicine, Sacramento, CA; 7The George Washington University School of Medicine, Washington, DC; 8University of Utah School of Medicine, Salt Lake City, UT; 9 University of New York at Buffalo School of Medicine, Buffalo, NY; 10EMSC, Washington, DC Background: The presence of the ‘‘Seat Belt Sign’’ (SBS) is believed to be associated with significant injury in children but a large, prospective multicenter study has not been performed. Objectives: To determine the association between the abdominal SBS and intra-abdominal injury (IAI) in children presenting to emergency departments (ED) with blunt torso trauma after a motor vehicle collision (MVC). Methods: We performed a prospective, multicenter, observational study of children with blunt torso trauma presenting after MVCs. Patient history and physical examination findings were documented before abdominal CT or laparotomy. SBS was defined as a continuous area of erythema, ecchymosis, or abrasion across the abdomen secondary to a seat belt restraint. We calculated the relative risk (RR) of IAI with 95% confidence intervals (CI) for children with and without SBS and, as a secondary analysis, the risk of IAI in those patients with SBS who were without abdominal pain or tenderness. Results: 3740 children with blunt torso trauma in MVCs with documentation about presence/absence of SBS from 20 EDs were enrolled; 585 (16%) had SBS. IAIs were more common in patients with SBS (14.4% vs 5.2%, RR 2.7; 95% CI: 2.1 to 3.5). Patients with SBS were more likely to have gastrointestinal injury (8.0% vs 0.5%, RR 15.8; 95% CI: 9.0 to 27.7), pancreatic injury (1.0% vs 0.3%, RR 3.6; 95% CI: 1.3 to 10.1), and solid organ injury (6.5% vs 4.4%, RR 1.5; 95% CI: 1.05 to 2.1) than children without SBS. Patients with SBS were more likely to undergo therapeutic laparotomy than those without SBS (6.3% vs 0.7%, RR 9.5; 95% CI: 5.6 to 16.1). IAI was diagnosed in 11/196 patients (5.6%, 95% CI: 2.8 to 9.8%) with SBS and no abdominal pain or tenderness. S184 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: Patients with SBS after MVC are at greater risk of IAI than those without SBS, predominately due to gastrointestinal and pancreatic injuries. In addition, those with SBS are more likely to undergo therapeutic laparotomy than those without. Although IAI is uncommon in SBS patients with no abdominal pain or tenderness, the risk of IAI is such that additional evaluation is generally necessary. 341 Pelvic CT Imaging In Adult Blunt Trauma: Does It Add Clinically Useful Information? Omayra L. Marrero, Alice M. Mitchell, Stacy Reynolds Carolinas Medical Center, Charlotte, NC Background: CT diagnostic imaging for blunt trauma routinely includes the abdomen (CTa) and pelvis (CTp) to evaluate suspected intra-abdominal injury (IAI). Objectives: We hypothesized that CTp contributes half the effective dose (E) of radiation to imaging protocols but adds to the diagnosis of IAI in less than 2% of adult blunt trauma patients without suspected pelvic fracture. Methods: We performed a retrospective study of patients older than 17 years evaluated with CTa and CTp at a Level I trauma center from March to June 2010 for blunt trauma. Two trained abstractors recorded data on standardized forms. Patients were excluded for known or suspected pelvic fracture defined by pelvic x-ray, CT bony pelvis, pelvic angiography, or an unstable or tender pelvis on exam. The primary outcomes were IAI identified on CTp and the mean percent reduction in E without CTp (Ep/Eap). Duplicate CT images were excluded from E calculations. We defined IAI as solid organ, hollow viscus, or vascular injury reported to the abdomen (dome of diaphragm to top of iliac crest) and pelvis (iliac crest to the greater trochanter). As a secondary outcome, we recorded occult pelvic fractures (OPF) on CTp. We estimated E based on the dose length product (DLP). We calculated mean E for combined CTa and CTp studies (Eap) and CTa (Ea) and CTp (Ep) studies alone. The mean effective percent dose reductions (Ep/Eap) were reported. A kappa statistic was calculated for the primary outcome of IAI. Results: 730 adult patients were evaluated at our center for blunt trauma during the study period. 210 patients met study criteria (84 female, 126 male). IAIs occurred in 23 patients (11%, 95% CI: 7 to 16%). CTp diagnosed IAI (traumatic ventral hernia) in 1 patient (0.5%, 95% CI: 0 to 2.6%), and detected 2 OPF and 1 non-displaced sacral fracture. Of the patients with OPF, 1 patient had a negative pelvic x-ray. The observed agreement was 99.5% (k = 0.66, 95% CI: 0.53 to 0.79). Eap was 10 mSv and Ea was 6 mSv. Elimination of CTp reduced E by a mean of 37% (95% CI: 29 to 46%). Conclusion: In this pilot study, pelvic imaging diagnosed IAI in 0.5% of patients. Elimination of CTp reduced the mean E by one third. The utility of CTp in low risk adult trauma patients without suspected pelvic fracture warrants additional investigation. 342 The Clinical Significance of Chest CT When the CXR is Normal in Blunt Trauma Patients Bory Kea1, Ruwan Gamarallage1, Hemamalini Vairamuthu1, Jonathan Fortman1, Eli Carrillo1, Kevin Lunney2, Gregory W. Hendey2, Robert M. Rodriguez1 1 University of California, San Francisco, School of Medicine; San Francisco General Hospital, San Francisco, CA; 2University of California, San Francisco, School of Medicine-Fresno, Fresno Regional Community Medical Center, Fresno, CA, Fresno, CA Background: Although computed tomography (CT) has been shown to detect more injuries than plain radiography in blunt trauma patients, it is unclear whether these injuries change patient management. Objectives: We sought to determine: 1) the proportion of patients with a normal initial CXR who are subsequently diagnosed with radiologic injuries on CT, 2) the proportion of patients with an abnormal initial CXR who are found not to have injuries on CT, and 3) the clinical significance (major, minor, or no clinical significance) of radiologic injuries seen on CT as determined by trauma expert panel. Methods: At two urban Level I trauma centers, blunt trauma victims over 14 years of age who received ED chest imaging as part of their evaluation were enrolled. An independent trauma expert panel consisting of six EM physicians and four trauma surgeons classified pairs of chest injuries and interventions/management (e.g. pneumothorax and chest tube placement) as major, minor, or no clinical significance. Charts of enrolled subjects were reviewed to determine their appropriate designation according to the trauma panel’s classification. Results: Of the 3639 subjects enrolled, 2808 (77.2%) had a CXR alone and 831 (22.8%) had both a CXR and chest CT. 7.0% (256/3639) had CXRs suggesting blunt trauma, of which 177 (69.1%, 95% CI: 63.2–74.5) had chest CTs confirming injury and 25 (9.8%, 95% CI: 6.7– 14.0%) revealed no injury on chest CT. Of 590 patients who had a chest CT after a normal CXR, 483 (81.6%, 95% CI: 78.6–84.8%) had CTs that were also read as normal and 18.1% (107/590) had CTs that diagnosed injuries, primarily rib fractures, pneumothorax, and hemothorax. Of the 590 patients with a normal CXR, chest CT led to a diagnosis of major injury in 13 (2.2%, 95% CI: 1.3–3.7%) and minor injury in 62 (10.5%, 95% CI: 8.3–13.2%) cases. Conclusion: Although chest CT frequently detects injuries missed on CXR in blunt trauma patients, it rarely changes patient management. Given the relatively low added value with the high costs and radiation risks, development of a guideline for selective chest CT utilization in blunt trauma is warranted. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 343 Occult Pneumothoraces Visualized in Children with Blunt Torso Trauma Lois Lee1, Alexander Rogers2, Peter Ehrlich2, Maria Kwok3, Peter Sokolove4, Stephen Blumberg5, Josh Kooistra6, Michelle Miskin7, Sandra Wootton-Gorges4, Art Cooper8, Nathan Kuppermann4, James Holmes4, PECARN IAI Study Group4 1 Children’s Hospital Boston, Boston, MA; 2Mott Children’s Hospital, Ann Arbor, MI; 3Columbia University Medical Center, New York City, NY; 4 UC Davis Children’s Hospital, Sacramento, CA; 5Jacobi Medical Center, Bronx, NY; 6 Spectrum Health Helen Devos Children’s Hospital, Grand Rapids, MI; 7University of Utah, Salt Lake City, UT; 8Columbia University Medical Center at Harlem Hospital, New York City, NY Background: Chest radiography is frequently the initial screening test to identify pneumothoraces (PTX) in trauma patients, but it has limitations. Computed tomography scans may identify PTX not seen on CXR (‘‘occult PTX’’). Objectives: The objectives of this study were to determine the prevalence of occult PTX in injured children and the rate of treatment with tube thoracostomy among those with occult PTX. Methods: We conducted a planned sub-study of children with CXRs performed from a large prospective multicenter observational cohort study of children <18 years old evaluated in emergency departments for blunt torso trauma from 5/07 to 1/10. The faculty radiologist interpretations of the initial CXRs and any subsequent imaging studies, including CT scans, were reviewed for the presence of a PTX. An ‘‘occult PTX’’ was defined as a PTX not identified on initial CXR, but visualized on CT scan. Prevalence rates and rate differences with 95% confidence intervals were calculated. Results: Of 12,044 enrolled in the parent study, 8,030 (67%) children (median age 11.3 years) underwent CXR in the ED. PTX on any imaging modality was identified in 383 (4.8%, 95% CI 4.3%, 5.3%) patients. The initial CXR demonstrated a PTX in 148 patients (1.8%, 95% CI 1.6%, 2.2%) including one false positive PTX. Occult PTX was diagnosed in 235 (2.9%, 95% CI 2.6%, 3.3%) patients. A tube thoracostomy was placed in 85 (57.8%, 95% CI 49.4%, 65.9%) patients with PTX on initial CXR and in 42 (17.9%, 95% CI 13.2%, 23.4%) patients with occult PTX (rate difference 39.9%, 95% CI 30.6%, 49.3%). Conclusion: In pediatric patients with blunt torso trauma, PTX are uncommon, but only one-third were identified on initial CXR. CT will frequently identify occult PTX, and although some undergo tube thoracostomy, most do not. • 344 www.aemj.org S185 Use and Impact of the FAST Exam in Children with Blunt Abdominal Trauma Jay Menaker1, Stephen Blumberg2, David Wisner3, Peter Dayan4, Michael Tunik5, Madelyn Garcia6, Prashant Mahajan7, Michelle Miskin8, David Monroe9, Dominic Borgialli10, Nathan Kuppermann3, James Holmes3, PECARN IAI Study Group11 1 University of Maryland, Baltimore, MD; 2Jacobi Medical Center, Bronx, NY; 3UC Davis School of Medicine, Sacramento, CA; 4Columbia Medical School, New York, NY; 5New York University, New York, NY; 6Rochester University, Rochester, NY; 7 Wayne State University, Detroit, MI; 8University of Utah, Salt Lake City, UT; 9Howard County Hospital, Columbia, MD; 10Hurley Medical Center, Flinth, MI; 11EMSC, Washington, DC Background: Use of the focused assessment sonography for trauma (FAST) is controversial in children with blunt abdominal trauma. Objectives: To evaluate the effect of clinician-performed FAST in children with abdominal trauma. We hypothesized that those undergoing a FAST would be less likely to undergo abdominal computed tomography (CT). Methods: We performed a planned analysis of a 20-center prospective study of children (<18 years) with blunt abdominal trauma. Patients with GCS scores >8 were eligible but those with hypotension and taken directly to the operating suite prior to CT were excluded. Patients from eight centers which did FAST on <5% of patients were excluded. Patients were risk stratified by clinician suspicion for intra-abdominal injury (IAI) as very low <1%, low 1–5%, moderate 5–10%, high 11–50%, and very high >50%. The relative risk (RR) for CT use based on undergoing a FAST was determined in each of these strata. Results: Of 11,025 eligible, 6,558 (median age = 10.7, IQR = 6.3, 15.5) met eligibility. 3,076 (46.9%) underwent abdominal CT and 381 (5.8%) were diagnosed with an IAI. 887 (13.7%) underwent clinician-performed FAST exam during emergency department evaluation. Use of the FAST exam among the 12 participating sites ranged from 5.5% to 58% and increased as clinician suspicion for IAI increased: 11.0% with <1% risk, 13.5% with 1–5% risk, 20.5% with 6–10% risk, 23.2% with 11–50% risk, and 30.7% with >50% risk. RRs for CT use stratified by clinician suspicion of IAI showed that patients with a low to moderate risk of IAI were less likely to receive a CT following a FAST exam compared to those not undergoing FAST: RR = 0.83 (0.67, 1.03), RR = 0.81 (0.72, 0.91), RR = 0.85 (0.78, 0.94), RR = 0.99 (0.94, 1.05), and RR = 0.97 (0.91, 1.05) for patients with suspicion for IAI of <1%, 1–5%, 5–10%, 11–50%, and >50%. Conclusion: The FAST exam is used in a small percentage of children with blunt abdominal trauma. While its use is highly variable amongst centers, use increases as clinician suspicion for IAI increases. Patients with a low to moderate clinician suspicion of IAI are less likely to undergo abdominal CT if they received a FAST exam. A randomized controlled trial is required to determine the benefits and drawbacks of the FAST exam in the evaluation of children with blunt abdominal trauma. S186 345 2012 SAEM ANNUAL MEETING ABSTRACTS History Of Anticoagulation And Head Trauma In Elderly Patients: Is There Really An Increased Risk Of Bleeding? Laura D. Melville, Ronald Tanaka New York Methodist Hospital, Brooklyn, NY Background: The literature reports that anticoagulation increases the risk of mortality in patients presenting to emergency departments (ED) with head trauma (HT). It has been suggested that such patients should be treated in a protocolized fashion, including CT within 15 minutes, and anticipatory preparation of FFP before CT results are available. There are significant logistical and financial implications associated with implementation of such a protocol. Objectives: Our primary objective was to determine the effect of anticoagulant therapy on the risk of intracranial hemorrhage (ICH) in elderly patients presenting to our urban community hospital following bunt head injury. Methods: This was a retrospective chart review study of HT patients >60 years of age presenting to our ED over a 6-month period. Charts reviewed were identified using our electronic medical record via chief complaints and ICD-9 codes and cross referencing with written CT logs. Research assistants underwent review of at least 25% of their contributing data to validate reliability. We collected information regarding use of warfarin, clopidogrel, and aspirin and CT findings of ICH. Using univariate logistic regression, we calculated odds ratios (OR) for ICH with 95% CI. Results: We identified 363 elderly HT patients. The mean age of our population was 72, 34 (8.3%) admitted to using anticoagulant therapy, and 23% were on antiplatelet drugs. 14 (3.8%) of the cohort had ICB, 3 patients required neurosurgical intervention, and 1 had transfusion of blood products. Of the non-anticoagulated patients, 12 (3.6%) were found to have ICH, half of those (6) were on an anti-platelet medication. Thirtyfour patients on warfarin, 2 (5.9%) had ICH. Six patients (7.1%) taking anti-platelet medication had ICH. OR for warfarin = 1.33 (95% CI, 0.20, 5.17) p = 0.71 OR for clopidogrel = 2.31 (0.34, 9.25) p = 0.30. OR for ASA = 2.260 (0.72, 6.71) p = 0.14 and for any combination of these treatments was OR = 2.33 (0.79, 7.26) p = 0.13. Conclusion: In our patient population, the risk of ICH in our anticoagulated patients was not significantly higher than those not taking warfarin. Anticipatory preparation of FFP may not be ideal resource utilization in our setting. Although not statistically significant, the trend toward higher risk in the patients on antiplatelet therapy warrants continues data collection and further investigation. (Originally submitted as a ‘‘latebreaker.’’) 346 The Value of microRNA for the Diagnosis and Prognosis of Emergency Department Patients with Sepsis Mike Puskarich1, Nathan Shapiro2, Stephen Trzeciak3, Jeffrey Kline4, Alan Jones1 1 University of Mississippi Medical Center, Jackson, MS; 2BIDMC, Boston, MA; 3Cooper University Hospital, Camden, NJ; 4Carolinas Medical Centera, Charlotte, NC Background: MircoRNAs are short, non-coding RNA sequences that regulate gene expression and protein synthesis, and have been proposed as biomarkers for a number of disease processes, including sepsis. Objectives: Determine diagnostic and prognostic ability of three previously identified microRNAs in ED patients with sepsis. Methods: Prospective observational study of a convenience sample of patients presenting to one of three large, urban, tertiary care EDs. Inclusion criteria: 1) Septic shock: suspected infection, two or more systemic inflammatory response (SIRS) criteria, and systolic blood pressure (SBP) <90 mmHg despite a 20 mL/kg fluid bolus; 2) Sepsis: suspected infection, two or more SIRS criteria, and SBP >90 mmHg; and 3) Control: ED patients without suspected infection, no SIRS criteria, and SBP >90 mmHg. Three microRNAs (miR-150, miR146a, and miR-223) were measured using real-time quantitative PCR from serum drawn at enrollment. IL-6, IL-10, and TNF-a were measured using a Bio-Plex suspension system. Baseline characteristics, IL-6, IL-10, TNF-a and microRNAs were compared using one way ANOVA or Fisher exact test, as appropriate. Correlations between miRNAs and SOFA scores, IL-6, IL-10, and TNF-a were determined using Spearman’s rank. A logistic regression model was constructed using in-hospital mortality as the dependent variable and miRNAs as the independent variables of interest. Bonferroni adjustments were made for multiple comparisons. Results: Of 93 patients, 24 were controls, 29 had sepsis, and 40 had septic shock. We found no difference in serum miR-146a or miR-223 between cohorts, and found no association between these microRNAs and either inflammatory markers or SOFA score. miR-150 demonstrated a significant correlation with SOFA score (q = 0.31, p = 0.01), IL-10 (q = 0.37, p = 0.001), but not IL-6 or TNF-a (p = 0.046, p = 0.59). Logistic regression demonstrated miR-150 to be associated with mortality, even after adjusting for SOFA score (p = 0.003). Conclusion: miR-146a or miR-223 failed to demonstrate any diagnostic or prognostic ability in this cohort. miR150 was associated with inflammation, increasing severity of illness, and mortality, and may represent a novel prognostic marker for diagnosis and prognosis of sepsis. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 347 Prognostic Value of Significantly Elevated Serum Lactate Measurements in Emergency Department Patients with Suspected Infection Michael Puskarich1, Jeffrey Kline2, Richard Summers1, Alan Jones1 1 University of Mississippi Medical Center, Jackson, MS; 2Carolinas Medical Center, Charlotte, NC Background: Previous studies have confirmed the prognostic significance of lactate concentrations categorized into groups (low, intermediate, high) among emergency department (ED) patients with suspected infection. Although the relationship between lactate concentrations categorized into groups and mortality appears to be linear, the relationship between lactate as a continuous measurement and mortality is uncertain. Objectives: We sought to evaluate the association between blood lactate concentrations along an incremental continuum, up to a maximum value of 20 mmol/ L, and mortality in ED patients with suspected infection. Methods: Retrospective cohort analysis of adult ED patients with suspected infection from a large urban emergency department during 2007–2010. Inclusion criteria: suspected infection evidenced by administration of antibiotics in the ED and measurement of whole blood lactate in the ED. Data were identified using a query of the electronic medical record via a previously compiled quality assurance database. All antibiotics available on formulary over the study period were included. If multiple lactate values were available for a patient, only the first value was included. The primary outcome was in-hospital mortality. Fractional polynomial regression was used to model the relationship between lactate concentration and in-hospital mortality. Results: 2,596 patients met inclusion criteria and were analyzed. The initial median lactate concentration was 2.1 mmol/L (IQR 1.3, 3.3) and the overall mortality rate was 14.4%. Of the entire cohort, 459 (17.6%) had an initial lactate >4 mmol/L. Mortality continued to rise across the continuum of incremental elevations, 6% for lactate <1.0 mmol/L up to 39% for a lactate of 18– 20 mmol/L. Polynomial regression analysis showed a strong curvilinear correlation between lactate and mortality (R = 0.92, p < 0.0001). Conclusion: In ED patients with suspected infection, we found a strong curvilinear relationship between incremental elevations in lactate concentration and mortality up to 20 mmol/L. These data support the use of lactate as a continuous variable rather than categorical variable when used for prognostic purposes. 348 Does Documented Physician Consideration Of Sepsis Lead To More Aggressive Treatment? Aaron M. Stutz, Jenifer Luman, John Winkler, Uwe Stolz, Lisa Stoneking, Kurt Denninghoff University of Arizona, Tucson, AZ Background: Timely treatment of sepsis is critical for patient outcomes. • www.aemj.org S187 Objectives: To examine the association between emergency physician recognition of SIRS and sepsis and subsequent treatment of septic patients. Methods: A retrospective cohort study of all-age patient medical records with positive blood cultures drawn in the emergency department from 11/2008–1/ 2009 at a Level I trauma center. Patient parameters were reviewed including vital signs, mental status, imaging, and laboratory data. Criteria for SIRS, sepsis, severe sepsis, and septic shock were applied according to established guidelines for pediatrics and adults. These data were compared to physician differential diagnosis documentation. The Mann-Whitney test was used to compare time to antibiotic administration and total volume of fluid resuscitation between two groups of patients: those with recognized sepsis and those with unrecognized sepsis. Results: SIRS criteria were present in 233/338 reviewed cases. Sepsis criteria were identified in 215/338 cases and considered in the differential diagnosis in 121/215 septic patients. Severe sepsis was present in 89/338 cases and septic shock was present in 42/338 cases. The sepsis 6-hour resuscitation bundle was completed in the emergency department in 16 cases of severe sepsis or septic shock. 121 patients who met sepsis criteria and were recognized by the ED physician had a median time to antibiotics of 150 minutes (IQR: 89–282) and a median IVF of 1500 ml (IQR: 500–3000). The 94 patients who met sepsis criteria but went unrecognized in the documentation had a median time to antibiotics of 225 minutes (IQR: 135–355) and median volume of fluid resuscitation of 1000 ml (IQR: 300–2000). Median time to antibiotics and median volume of fluid resuscitation differed significantly between recognized and unrecognized septic patients (p = 0.003 and p = 0.002, respectively). Conclusion: Emergency physicians correctly identify and treat infection in most cases, but frequently do not document SIRS and sepsis. Lack of documentation of sepsis in the differential diagnosis is associated with increased time to antibiotic delivery and a smaller total volume of fluid administration, which may explain poor sepsis bundle compliance in the emergency department. 349 Elevated Inter-alpha Trypsin Inhibitor (ITI) levels in Emergency Department Patients with Severe Sepsis and Septic Shock Anthony M. Napoli, Ling Zhang, Fenwick Gardiner, Patrick Salome Warren Alpert Medical School of Brown University, Providence, RI Background: Serine protease inhibitors, specficially Inter-alpha trypsin inhibitor (ITI), are thought to be protective in sepsis by mitigating the neutrophil-derived proteinases that are upregulated during states of severe inflammation. Prolonged negative feedback of the inflammatory cascade is thought to exhaust the body’s reserves leading to dysregulation of cellular immunity seen in sepsis. Levels of ITI have been associated with mortality in ICU sepsis patients while small studies have suggested improved mortality in animals and hemodynamic stability in animals when repleted. S188 2012 SAEM ANNUAL MEETING ABSTRACTS Objectives: Characterize ITI levels in ED patients and determine if these levels are associated with severity of sepsis. Methods: Prospective cohort of controls(C), acutely ill inflammatory non-septic illnesses (AINS) with SIRS, and patients with severe sepsis or septic shock (SS) (lactate >4, SBP<90 after 2L normal saline). A competitive ELISA assay using murine 69.26 antibody was used to measure ITI levels. Based on prior results from inpatients with sepsis, a sample size of 100 patients would be necessary (two-tail, p < 0.05,b = 0.8), assuming a standard error of difference of 140. Results: 96 patients are enrolled to date with completed assay analysis on 66. ITI levels are lower in patients with SS vs. C (262 ± 73 vs. 318 ± 89, p = 0.03) but not vs. AINS (264 ± 93, p = 0.94). There is a trend toward an association of ITI levels with APACHE II score (r = 0.33, p = 0.11). Conclusion: ITI levels in ED patients presenting with sepsis are higher than control patients but not AINS. The degree of overlap with control patients and the lack of distinction of ITI levels in SS vs. AINS may limit its utility as a diagnostic and prognostic marker. Larger studies are needed to confirm these results. (Originally submitted as a ‘‘late-breaker.’’) 350 The Microcirculation is preserved in Sepsis Patients without Organ Dysfunction or Shock Michael R. Filbin1, Peter Hou2, Apurv Barche1, Katherine Wai2, Allen Gold2, Siddharth Parma2, Michael Massey3, Alex Bracey3, Nicolaj Duujas3, Hendrick Zijlstra3, Nathan I. Shapiro3 1 Massachusetts General Hospital, Boston, MA; 2 Brigham & Women’s Hospital, Boston, MA; 3 Beth Israel Deaconess Hospital, Boston, MA Background: The advent of sidestream darkfield imaging (SDF) has allowed the direct observation of the microcirculation at the bedside in a manner that was previously not possible. Recent research with SDF has shown that the microcirculation is dysfunctional in critically ill patients with septic shock; however, it is unstudied in patients of lower acuity presenting to the ED with uncomplicated sepsis. Objectives: We hypothesize that patients with early sepsis (without organ dysfunction or shock) have signs of microcirculatory dysfunction compared to noninfected control patients. Methods: Prospective, observational study of a convenience sample of sepsis patients meeting the following criteria: clinical suspicion of infection, age 18 or older, two or more SIRS criteria, no organ dysfunction or shock; control patients were ED patients without infection or SIRS criteria. The study was conducted across three urban tertiary care EDs. The images were obtained using a videomicroscope and sent to a coordinating center for offline analysis. A trained, blinded scorer analyzed tapes of microcirculatory flow using a previously validated semi-quantitative scale (0–3) for small vessel blood flow velocity. The flow velocity was compared using an unpaired t-test alpha set at 0.05. Results: There were 98 patients enrolled: 59 sepsis patients and 39 controls. The mean age of the sepsis (58.0, 95% CI 53.6–62.4) and control patients (58.3, 52.3– 64.3) was similar between groups, p = 0.94. There was no difference in average flow velocity for the infected patients 2.94 (2.90–2.97) versus non-infected controls 2.92 (2.88–2.96), with a mean difference of 0.017 ()0.067 to 0.033), p = 0.51. Conclusion: We did not observe microcirculatory flow disturbances in patients with sepsis by SIRS criteria without organ dysfunction or shock. One may postulate that infection with SIRS criteria alone is not indicative of microcirculatory dysfunction. Further study is warranted as to the role of microcirculatory dysfunction in progression of disease in sepsis. 351 Benchmarking The Incidence And Mortality Of Severe Sepsis In The United States David F. Gaieski, J. Matthew Edwards, Michael J. Kallan, Brendan G. Carr University of Pennsylvania School of Medicine, Philadelphia, PA Background: Severe sepsis is a common clinical syndrome with substantial human and financial impact. In 1992 the first consensus definition of sepsis was published. Subsequent epidemiologic estimates were collected using administrative data, but ongoing discrepancies in the definition of severe sepsis led to large differences in estimates. Objectives: We seek to describe the variations in incidence and mortality of severe sepsis in the US using four methods of database abstraction. Methods: Using a nationally representative sample, four previously published methods (Angus, Martin, Dombrovskiy, Wang) were used to gather cases of severe sepsis over a 6-year period (2004–2009). In addition, the use of new ICD-9 sepsis codes was compared to previous methods. Our main outcome measure was annual national incidence and in-hospital mortality of severe sepsis. Results: The average annual incidence varied by as much as 3.5 fold depending on method used and ranged from 894,013 (300 / 100,000 population) to 3,110,630 (1,031 / 100,000) using the methods of Dombrovskiy and Wang, respectively. Average annual increase in the incidence of severe sepsis was similar (13.0–13.3%) across all methods. Total mortality mirrored the increase in incidence over the 6-year period (168,403– 229,044 [Dombrovskiy] and 299,939–375,608 [Wang]). In-hospital mortality ranged from 14.7% to 29.9% using abstraction methods of Wang and Dombrovskiy, respectively. Using all methods, there was a decrease in in-hospital mortality across the 6-year period (35.2– 25.6% [Dombrovskiy] and 17.8–12.1% [Wang]). Use of ICD-9 sepsis codes more than doubled over the 6-year period [158,722–489,632 (995.92 severe sepsis), 131,719– 303,615 (785.52 septic shock)]. Conclusion: There is substantial variability in incidence and mortality of severe sepsis depending on method of database abstraction used. A uniform, consistent ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 definition of severe sepsis is needed for use in national registries to facilitate accurate assessment of clinical interventions and outcomes comparisons between hospitals and regions. 352 Emergency Medical Services Focused Assessment with Sonography in Trauma and Cardiac Ultrasound in Cardiac Arrest: The Training Phase Barry Knapp, Donald Byars, Virginia Stewart, Rebecca Ryszkiewicz, David Evans Eastern Virginia Medical School, Suffolk, VA Background: With improvement in ultrasound (US) technology, portable units are now able to endure the rigors of the prehospital setting. The ability of prehospital providers using US to make an accurate assessment of free abdominal and pericardial fluid in trauma victims or to determine cardiac activity in arrest patients using US may improve both resuscitative efforts and resource utilization. No studies address how best to train emergency medical services (EMS) providers. Objectives: The purpose of this study is to determine whether EMS providers at the EMT-Intermediate and and EMT-Paramedic levels can, in a four-hour training session, acquire the knowledge and skill to operate a portable US machine and achieve a high level of accuracy performing cardiac and FAST exams. Methods: This was a prospective cohort educational study conducted on Paramedic and Intermediate EMS providers in an urban city. Participants used a Sonosite M-Turbo US unit with a phased-array probe. We recruited 90 EMS providers, 70 EMT-Ps and 20 EMT-Is. All enrolled participants were required to complete a one-hour online web-based home study program. The four hour training program consisted of: 1) didactic lecture, 2) practice scanning, and 3) testing scenarios. All subjects also underwent pre- and post-training written tests. The testing scenarios included one normal and abnormal cardiac, and one normal and abnormal FAST. All scenarios were performed on live standardized patients (SP) and graded in an Observed Standardized Clinical Examination (OSCE) format. In abnormal scenarios, clips were displayed once proper ultrasound views were obtained on the live SP. Results: The average score on the pre-test was 73% (14.62 ± 2.34), compared with the post-test score of 95% (19.09 ± 1.29) (p-value <0.0001). Each station was graded with an 18-point tool for the cardiac exam and a 32-point tool for the FAST exam. In total, EMS providers (n = 90) scored on average 98.9 points out of 100 on the OSCE testing stations. EMT-Ps (n = 70) scored, on average 98.9 points out of 100 on the OSCE stations. Average score for EMT-Is (n = 20) was 99.1 points out of 100 on the OSCE stations. There was no statistical difference between the performance of EMT-Is and Ps (p-value of 0.86). Conclusion: EMS providers were able to successfully demonstrate proper image acquisition and interpretation of both basic cardiac and FAST US exams after a four-hour training module. • 353 www.aemj.org S189 Videolaryngoscopy as an Educational Tool for the Novice Pediatric Intubator: A Comparative Study Jendi L. Haug, Geetanjali Srivastava University of Texas Southwestern, Dallas, TX Background: Prior studies show that novice residents have low intubation success rates. Videolaryngoscopy is a technology that allows for enhanced airway visualization. Objectives: We hypothesized that residents who receive videolaryngoscope training (video group) would have a higher success rate in intubating high-fidelity simulation mannequins compared to residents trained with traditional methods. We also predicted that the video group would have a decrease in the mean time to intubation, number of intubation attempts, and right main stem (RMS) intubation rate, as well as a higher rate of intubation before the onset of oxygen desaturation and an improved view of the glottis. Methods: This was a randomized controlled study comparing a traditional method of teaching intubation to instruction with the adjunct of a videolaryngoscope. Participants received a didactic presentation reviewing intubation basics. They were then randomized into two groups: traditional and video. All were taught and practiced direct laryngoscopy on a low-fidelity mannequin. The video group received additional training with a videolaryngoscope. Afterwards, each participant intubated an infant and adult high-fidelity mannequin. Data were collected on length of time to intubation, number of attempts, desaturation before intubation, Cormack Lehane (CL) grade view, and RMS intubations. We defined successful intubation as tracheal intubation in £ 3 attempts. Results: 32 residents (17 pediatric, 15 emergency medicine) presented for participation. There was no difference in intubation success rate between the groups for either the adult (71% traditional vs. 77% video, p = 1) or infant (100% vs. 100%, p = 1) mannequin. There were differences in the following secondary outcomes for the infant mannequin: mean intubation time (traditional vs. video: 35 vs. 19 seconds, 15.8 second difference, 95% CI 7.9–23.7) and the rate of intubation before the onset of desaturations (53% traditional vs. 100% video, p = 0.003). There was no difference for the following: number of attempts, CL grade, and RMS intubation. Conclusion: There was no difference in intubation success rate between the group trained with videolaryngoscopy and those trained with traditional method. The video group had both a lower mean time to intubation and higher rate of intubation before the onset of oxygen desaturation with the infant mannequin. 354 CT Imaging In The ED: Perception Of Radiation Risks By Provider Richard Bounds1, Michael Trotter2, Wendy Nichols1, James F. Reed III1 1 Christiana Care Health System, Newark, DE; 2 Taylor Hospital, Ridley Park, NJ Background: Radiation exposure from medical imaging has been the subject of many major journal articles, as well as the topic of mainstream media. Some S190 2012 SAEM ANNUAL MEETING ABSTRACTS estimate that one-third of all CT scans are not medically justified. It is important for practitioners ordering these scans to be knowledgeable of currently discussed risks. Objectives: To compare the knowledge, opinions, and practice patterns of three groups of providers in regards to CTs in the ED. Methods: An anonymous electronic survey was sent to all residents, physician assistants, and attending physicians in emergency medicine (EM), surgery, and internal medicine (IM) at a single academic tertiary care referral Level I trauma center with an annual ED volume of over 160,000 visits. The survey was pilot tested and validated. All data were analyzed using the Pearson’s chi-square test. Results: There was a response rate of 32% (220/668). Data from surgery respondents were excluded due to a low response rate. In comparison to IM, EM respondents correctly equated one abdominal CT to between 100 and 500 chest x-rays, reported receiving formal training regarding the risks of radiation from CTs, believe that excessive medical imaging is associated with an increased lifetime risk of cancer, and routinely discuss the risks of CT imaging with stable patients more often (see Table 1). Particular patient factors influence whether radiation risks are discussed with patients by 60% in each specialty (see Table 2). Before ordering an abdominal CT in a stable patient, IM providers routinely review the patient’s medical imaging history less often than EM providers surveyed. Overall, 67% of respondents felt that ordering an abdominal CT in a stable ED patient is a clinical decision that should be discussed with the patient, but should not require consent. Conclusion: Compared with IM, EM practitioners report greater awareness of the risks of radiation from CTs and discuss risks with patients more often. They also review patients’ imaging history more often and take this, as well as patients’ age, into account when ordering CTs. These results indicate a need for improved education for both EM and IM providers in regards to the risks of radiation from CT imaging. Table 1 - Abstract 354: Question EM n = 89 IM n = 115 p-value Formal training? Inc life risk of cancer belief? Correctly identified CT comparison with chest x-rays Discussion of risks with particular patients Discussion of risks with all stable patients History review prior to order 63 (70.8%) 81 (91.0%) 45 (39.1%) 83 (72.2%) 0.001 0.002 30 (34%) 21 (18%) 0.004 71 (80%) 55 (49%) 0.001 14 (16%) 10 (9%) 0.001 70 (79%) 67 (62%) 0.014 Table 2 - Abstract 354: Decision Factors EM n = 57 IM n = 45 p-value Female sex Previous # CTs 53 (93%) 57 (100%) 33 (73%) 38 (84%) 0.001 0.001 355 A Mobile Lightly-embalmed Cadaver Lab As A Possible Model For Training Rural Providers Wesley Zeger1, Paul Travis2, Michael Wadman1, Carol Lomneth1, Sara Keim1, Stephanie Vandermuelen1 1 UNMC, Omaha, NE; 2Creighton Medical School, Omaha, NE Background: In Nebraska, 80% of emergency departments have annual visits less than 10,000, and the predominance are in rural settings. General practitioners working in rural emergency departments have reported low confidence in several emergency medicine skills. Current staffing patterns include using midlevels as the primary provider with non-emergency medicine trained physicians as back-up. Lightly-embalmed cadaver labs are used for resident’s procedural training. Objectives: To describe the effect of a lightlyembalmed cadaver workshop on physician assistants’ (PA) reported level of confidence in selected emergency medicine procedures. Methods: An emergency medicine procedure lab was offered at the Nebraska Association of Physician Assistants annual conference. Each lab consisted of a 2-hour hands-on session teaching endotracheal intubation techniques, tube thoracostomy, intraosseous access, and arthrocentesis of the knee, shoulder, ankle, and wrist to PAs. IRB-approved surveys were distributed pre-lab and a post-lab survey was distributed after lab completion. Baseline demographic experience was collected. Pre- and post-lab procedural confidence was rated on a six-point Likert scale (1–6) with 1 representing no confidence. The Wilcoxon Signed-Rank Test was use to calculate p values. Results: 26 PAs participated in the course. All completed a pre- and post-lab assessment. No PA had done any one procedure more than 5 times in their career. Pre-lab modes of confidence level were £3 for each procedure. Post-lab modes were >4 for each procedure except arthrocentesis of the ankle and wrist. However, post lab assessments of procedural confidence significantly improved for all procedures with p values <0.05. Conclusion: Midlevel providers’ level of confidence improved for emergent procedures after completion of a procedure lab using lightly-embalmed cadavers. A mobile cadaver lab would be beneficial to train rural providers with minimal experience. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 356 Creation Of A Valid And Reliable Competency Assessment For Advanced Airway Management James K. Takayesu1, Demian Szyld2, Calvin Brown III3, Benjamin Sandefur3, Eric Nadel3, Ron M. Walls3 1 Massachusetts General Hospital, Boston, MA; 2 NYU, New York City, NY; 3Brigham and Women’s Hospital, Boston, MA Background: Difficult airway management is a core competency in EM. Current assessment of individual resident performance in managing the difficult airway typically is by non-systematic global assessments by varying faculty members in the ED, rather than by controlled, reliable assessment methods. Objectives: To design an assessment method for difficult airway management that is reliable and valid to determine competence during residency training. Methods: Expert consensus was used to derive three difficult airway simulated scenarios: unanticipated difficult airway (UDA) requiring alternative method, anticipated difficult airway (ADA) using flexible fiberoptics, and failed airway requiring cricothyrotomy (FAC). PGY3 resident performance on each of the three scenarios was assessed using a standardized performance checklist, including cognitive and psychomotor items, by each of two faculty assessors. Interobserver agreement was calculated for each checklist item to determine reliability of assessment. Validity was evaluated by global assessment of performance. Results: The checklist instruments had 24 items for UAD, 31 items for ADA, 14 items for FAC. Using a cutoff of either kappa agreement greater than 0.6 or interobserver agreement greater than 90%, the checklists were refined to include 19 items for UAD, 18 items for ADA, 8 items for FAC. Conclusion: A simple, standardized checklist with high inter-rater reliability can be used to determine competency in difficult airway management. • 357 www.aemj.org S191 Implementing and Measuring Efficacy of a Capnography Simulation Training Program for Prehospital Health Care Providers Jared T. Roeckner, Ivette Motola, Angel A. Brotons, Hector F. Rivera, Robert E. Soto, S. Barry Issenberg University of Miami Miller School of Medicine, Miami, FL Background: Capnography is increasingly used in prehospital and emergency department settings to aid diagnosis and management of patients with respiratory emergencies. End-tidal CO2 (ETCO2) monitoring can verify correct endotracheal tube placement, determine hypo/hyperventilation, and monitor effective CPR. While there is evidence for the effectiveness of capnography in the prehospital setting, little has been published on educational methodologies for training EMS personnel. To address this void, the University of Miami Gordon Center for Research in Medical education incorporated capnography training as part of its 8-hour simulation-based airway management course. Objectives: Determine the effect of capnography skills training on knowledge acquisition in practicing EMS personnel. We hypothesized that there would be no difference between the results of the precourse and postcourse assessments. Methods: From August 2008 to September 2011, 424 paramedics throughout Florida participated in the airway course. The educational components included a didactic session of capnography principles utilizing tracings, demonstration on a live capnography monitor, case-based scenarios using an audience-response system, and participation in six simulation scenarios. We measured cognitive outcomes using precourse and postcourse assessments. Eleven of the 30 questions incorporated capnogram interpretation or decisionmaking that relied on participants’ ability to interpret ETCO2. Statistical analysis was performed using a paired, two-sample t-test. S192 2012 SAEM ANNUAL MEETING ABSTRACTS Results: Eighteen of the 424 learners were excluded from analysis because of incomplete assessments. The precourse mean score for 406 participants was 68% (SD 0.150) for all of the capnography questions and the postcourse mean was 91% (SD 0.104). The mean improvement was 22.8% (SD 0.181, 99% CI 20.4–25.2, p < 0.0001). Questions depicting capnograms illustrating esophageal intubation (Pre 29%, Post 82%) and rebreathing (Pre 29%, Post 92%) showed the greatest score improvement. Conclusion: Training that incorporated didactic sessions and hands-on simulation scenarios significantly increased knowledge of capnography. Marked improvement was seen in the interpretation of capnograms of esophageal intubation, which has direct application in verifying proper intubation. 358 Current Knowledge of and Willingness to Perform Hands-Only CPR in Laypersons Jennifer Urban, Adam J. Singer, Henry C. Thode Jr Stony Brook University, Stony Brook, NY Background: Sudden cardiac arrest is the leading cause of death. Early and effective cardiopulmonary resuscitation (CPR) improves mortality, although participation in CPR remains low. Recent simplified guidelines recommend hands-only CPR (HCPR) for laypersons. Objectives: We determined current knowledge of and willingness to perform hands-only CPR. Methods: Design-Prospective anonymous survey. Setting-Academic suburban emergency department. Subjects-Adult patients and visitors in suburban ED. Survey instrument-33 item closed question format based on prior studies that included baseline demographics and knowledge and experience of CPR. Main outcomeKnowledge of and willingness to perform hands-only CPR. Data Analysis-Descriptive statistics. Univariate and multivariate analyses were performed to determine the association between predictor variables and knowledge of and willingness to perform HCPR. Results: We surveyed 532 subjects; mean age was 44 ± 16; 53.2% were female, 75.6% were white. 45.5% were college graduates, and 44.4% had an annual income of greater than $50,000. 41.9% had received prior CPR training; only 10.3% had performed CPR. Of all subjects, 124 (23.3%) had knowledge of HCPR, yet 414 (77.8%) would be willing to perform HCPR on a stranger. Age (p = 0.003) and income (p = 0.014) predicted knowledge of HCPR. A history of a cardiacrelated event in the family (p = 0.003) and previous CPR training (p = 0.01) were associated with likelihood to perform HCPR . Conclusion: Less than one fifth of surveyed laypersons knew of hands-only CPR, yet three quarters would be willing to perform hands only CPR even on a stranger. Efforts to increase layperson education are required to enhance CPR performance. 359 Use of Automated External Defibrillators for Pediatric Out-of-Hospital Cardiac Arrests: A Comparison to Adult Patients Austin Johnson1, Brian Harahan2, Jason S. Haukoos1, David Slattery3, Bryan McNally4, Comilla Sasson5 1 Denver Health Medical Center, Denver, CO; 2 University of Minnesota, Minneapolis, MN; 3 University of Nevada, Las Vegas, NV; 4Emory 5 University, Atlanta, GA; University of Colorado, Aurora, CO Background: Use of automated external defibrillators (AED) improves survival in out-of-hospital cardiopulmonary arrest (OHCA). Since 2005, the American Heart Association has recommended that individuals one year of age or older who sustain OHCA have an AED applied. Little is known about how often this occurs and what factors are associated with AED use in the pediatric population. Objectives: Our objective was to describe AED use in the pediatric population and to assess predictors of AED use when compared to adult patients. Methods: We conducted a secondary analysis of prospectively collected data from 29 U.S. cities that participate in the Cardiac Arrest Registry to Enhance Survival (CARES). Patients were included if they had a documented resuscitation attempt from October 1, 2005 through December 31, 2009 and were ‡1 year old. Patients were considered pediatric if they were less than 19 years old. AED use included application by laypersons and first responders. Hierarchical multivariable logistic regression analysis was used to estimate the associations between age and AED use. Results: There were 19,559 OHCAs included in this analysis, of which 239 (1.2%) occurred in pediatric patients. Overall AED use in the final sample was 5,517, with 1,751 (8.9%) total survivors. AEDs were applied less often in pediatric patients (19.7%, 95% CI: 14.6%–24.7% vs 28.3%, 95% CI: 27.7%–29.0%). Within the pediatric population, only 35.4% of patients with a shockable rhythm had an AED used. In all pediatric patients, regardless of presenting rhythm, AED use demonstrated a statistically significant increase in return of spontaneous circulation (AED used 29.8%, 95% CI: 16.2–43.4 vs AED not used 16.8%, 95% CI: 11.4–22.1, p < 0.05), although there was no significant increase in survival to hospital discharge (AED used 12.8%; AED not used 5.2%; p = 0.057). In the adjusted model, pediatric age was independently associated with failure to use an AED (OR 0.61, 95% CI: 0.42– 0.87) as was female sex (OR 0.88, 95% CI: 0.81–0.95). Patients who had a public arrest (OR 1.35, 95% CI: 1.24–1.46) or one that was witnessed by a bystander (OR 1.20. 95%: CI 1.11–1.29) were also predictive of AED use. Conclusion: Pediatric patients who experience OHCA are less likely to have an AED used. Continued education of first responders and the lay public to increase AED use in this population is necessary. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 360 Does Implementation of a Therapeutic Hypothermia Protocol Improve Survival and Neurologic Outcomes in all Comatose Survivors of Sudden Cardiac Arrest? Ken Will, Michael Nelson, Abishek Vedavalli, Renaud Gueret, John Bailitz Cook County (Stroger), Chicago, IL Background: The American Heart Association (AHA) currently recommends therapeutic hypothermia (TH) for out of hospital comatose survivors of sudden cardiac arrest (CSSCA) with an initial rhythm of ventricular fibrillation (VF). Based on currently limited data, the AHA further recommends that physicians consider TH for CSSCA, from both the out and inpatient settings, with an initial non-VF rhythm. Objectives: Investigate whether a TH protocol improves both survival and neurologic outcomes for CSSCA, for out and inpatients, with any initial rhythm, in comparison to outcomes previously reported in literature prior to TH. Methods: We conducted a prospective observational study of CSSCA between August 2009 and May 2011 whose care included TH. The study enrolled eligible consecutive CSSCA survivors, from both out and inpatient settings with any initial arrest rhythm. Primary endpoints included survival to hospital discharge and neurologic outcomes, stratified by SCA location, and by initial arrest rhythm. Results: Overall, of 27 eligible patients, 11 (41%, 95% CI 22–66%) survived to discharge, 7 (26%, 95% CI 9– 43%) with at least a good neurologic outcome. Twelve were out and 15 were inpatients. Among the 12 outpatients, 6 (50%, 95% CI 22–78%) survived to discharge, 5 (41%, 95% CI 13–69%) with at least a good neurologic outcome. Among the 15 inpatients, 5 (33%, 95% CI 9– 57) survived to discharge, 2 (13%, 95% CI 0–30%) with at least a good neurologic outcome. By initial rhythm, 6 patients had an initial rhythm of VF/T and 21 non-VF/T. Among the 6 patients with an initial rhythm of VF/T, 4 (67%, CI 39–100%) survived to discharge, all 4 with at least a good outcome, including 3 out and 1 inpatients. Among the 21 patients with an initial rhythm of nonVF/T, 7 (33%, CI 22–53%) survived to discharge, 3 (14%, CI 0–28%) with at least a good neurologic outcome, including 2 out and 1 inpatients. Conclusion: Our preliminary data initially suggest that local implementation of a TH protocol improves survival and neurologic outcomes for CSSCA, for out and inpatients, with any initial rhythm, in comparison to outcomes previously reported in literature prior to TH. Subsequent research will include comparison to local historical controls, additional data from other regional TH centers, as well as comparison of different cooling methods. • 361 www.aemj.org S193 Protocolized Use of Sedation and Paralysis with Therapeutic Hypothermia Following Cardiac Arrest Timothy J. Ellender1, Dustin Spencer2, Judith Jacobi3, Michelle Deckard2, Elizabeth Taber2 1 Indiana University Department of Emergency 2 Medicine, Indianapolis, IN; IU Health Methodist Hospital, Indianapolis, IN; 3Indiana University, Indianapolis, IN Background: Therapeutic hypothermia (TH) has been shown to improve the neurologic recovery of cardiac arrest patients who experience return of spontaneous circulation (ROSC). It remains unclear as to how earlier cooling and treatment optimization influence outcomes. Objectives: To evaluate the effects of a protocolized use of early sedation and paralysis on cooling optimization and clinical outcomes in survivors of cardiac arrest. Methods: A 3-year (2008–2010), pre-post intervention study of patients with ROSC after cardiac arrest treated with TH was performed. Those patients treated with a standardized order set which lacked a uniform sedation and paralytic order were included in the pre-intervention group, and those with a standardized order set which included a uniform sedation and paralytic order were included in the post-intervention group. Patient demographics, initial and discharge Glasgow Coma Scale (GCS) scores, resuscitation details, cooling time variables, severity of illness as measured by the APACHE II score, discharge disposition, functional status, and days to death were collected and analyzed using Student’s t-tests, Man-Whitney U tests, and the Log-Rank test. Results: 232 patients treated with TH after ROSC were included, with 107 patients in the pre-intervention group and 125 in the post-intervention group. The average time to goal temperature (33C) was 227 minutes (pre-intervention) and 168 minutes (post-intervention) (p = 0.001). A 2-hour time target was achieved in 38.6% of the patients (post-intervention) compared to 24.5% in the pre-group (p = 0.029). Twenty-eight day mortality was similar between groups (65.4% and 65.3%) though hospital length of stay (10 days pre- and 8 days post-intervention) and discharge GCS (13 preand 14-post-intervention) differed between cohorts. More post-intervention patients were discharged to home (55.8%) compared to 43.2% in the pre-intervention group. Conclusion: Protocolized use of sedation and paralysis improved time to goal temperature achievement. These improved TH time targets were associated with improved neuroprotection, GCS recovery, and disposition outcome. Standardized sedation and paralysis appears to be a useful adjunct in induced TH. S194 362 2012 SAEM ANNUAL MEETING ABSTRACTS Physician Clinical Impression Compared to a Novel Clinical Decision Rule for Use in Sparing Pediatric Patients with Signs and Symptoms of Acute Appendicitis Exposure to Computed Tomography Michael Wandell1, Michael Brown2, Harold Simon3, Karen Copeland4, David Huckins5, Brent Blumenstein6 1 AspenBio Pharma, Castle Rock, CO; 2Michigan State University, Grand Rapids, MI; 3 Emergency Medicine Children’s Healthcare of Atlanta, Atlanta, GA; 4AspenBio Pharma, Castle Rock, CO; 5Newton Wellesley, Newton, MA; 6 TriArch Consulting, Washington, DC Background: CT is increasingly used to assess children with signs and symptoms of acute appendicitis (AA) though concerns regarding long-term risk of exposure to ionizing radiation have generated interest in methods to identify children at low risk. Objectives: We sought to derive a clinical decision rule (CDR) of a minimum set of commonly used signs and symptoms from prior studies to predict which children with acute abdominal pain have a low likelihood of AA and compared it to physician clinical impression (PCI). Methods: We prospectively analyzed 420 subjects aged 2 to 20 years in 11 U.S. emergency departments with abdominal pain plus signs and symptoms suspicious for AA within the prior 72 hours. Subjects were assessed by study staff unaware of their diagnosis for 17 clinical attributes drawn from published appendicitis scoring systems and physicians responsible for physical examination estimated the probability of AA based on PCI prior to their medical disposition. Based on medical record entry rate, frequently used CDR attributes were evaluated using recursive partitioning and logistic regression to select the best minimum set capable of discriminating subjects with and without AA. Subjects were followed to determine whether imaging was used and use was tabulated by both PCI and the CDR to assess their ability to identify patients who did or did not benefit based on diagnosis. Results: This cohort had a 27.3% prevalence (118/431 subjects) of AA. We derived a CDR based on the absence of two out of three of the following attributes: abdominal tenderness, pain migration, and rigidity/ guarding had a sensitivity of 89.8% (95% CI: 83.1–94.1), specificity of 47.6% (95% CI: 42.1–53.1), NPV of 92.5% (95% CI: 87.4–95.7), and negative likelihood ratio of 0.21 (95% CI: 0.12–0.37). The PCI set at AA <30% pre-test probability had a sensitivity of 94.1% (95% CI: 88.3– 97.1), specificity of 49.4% (95% CI: 43.9–54.9), NPV of 95.7% (95% CI: 91.3–97.9), and negative likelihood ratio of 0.12 (95% CI: 0.06–0.25). The methods each classified 37% of the patients as low risk for AA. Our CDR identified 29.1% (43/148) of low risk subjects who received CT but being AA (-), could have been spared CT, while the PCI identified 20.1% (30/149). Conclusion: Compared to physician clinical impression, our clinical decision rule can identify more children at low risk for appendicitis who could be managed more conservatively with careful observation and avoidance of CT. 363 Negative Predictive Value of a Low Modified Alvarado Score For Adult ED Patients with Suspected Appendicitis Andrew C. Meltzer1, Brigitte M. Baumann2, Esther H. Chen3, Frances S. Shofer4, Angela M. Mills4 1 George Washington University, Washington, DC; 2Cooper University Hospital, Camden, NJ; 3 UCSF, San Francisco, CA; 4University of Pennsylvania, Philadelphia, PA Background: Abdominal pain is the most common complaint in the ED and appendicitis is the most common indication for emergency surgery. A clinical decision rule (CDR) identifying abdominal pain patients at a low risk for appendicitis could lead to a significant reduction in CT scans and could have a significant public health impact. The Alvarado score is one of the most widely applied CDRs for suspected appendicitis, and a low modified Alvarado score (less than 4) is sometimes used to rule out acute appendicitis. The modified Alvarado score has not been prospectively validated in ED patients with suspected appendicitis. Objectives: We sought to prospectively evaluate the negative predictive value of a low modified Alvarado score (MAS) in ED patients with suspected appendicitis. We hypothesized that a low MAS (less than 4) would have a sufficiently high NPV (>95%) to rule out acute appendicitis. Methods: We enrolled patients greater than or equal to 18 years old who were suspected of having appendicitis (listed as one of the top three diagnosis by the treating physician before ancillary testing) as part of a prospective cohort study in two urban academic EDs from August 2009 to April 2010. Elements of the MAS and the final diagnosis were recorded on a standard data form for each subject. The sensitivity, specificity, negative predictive value (NPV), and positive predictive value (PPV) were calculated with 95% CI for a low MAS and final diagnosis of appendicitis. Results: Of 290 enrolled patients, 28 were excluded for missing a MAS variable. The remaining 262 patients were included for analysis (mean age 35 years [range 18–89], 68% female, 52% white), of whom 54 (21%) had acute appendicitis. The test characteristics were as follows: sensitivity 72.2% (95% CI 58–84%); specificity 54.3% (95% CI 47–61%); PPV 29.1% (95% CI 22–38%); and NPV 88.3% (95% CI 81–93%). Conclusion: The negative predictive value of a MAS less than 4 was 88.3%. Given the serious complications of a missed appendicitis, we believe that the NPV is too low to recommend adoption of the low MAS to clinically rule out appendicitis. These observations lay the groundwork for future studies to improve clinical diagnosis of appendicitis and safely reduce the amount of CT scans for patients at low risk for appendicitis. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 364 A Novel BioMarker Panel to Rule Out Acute Appendicitis in Pediatric Patients with Abdominal Pain Michael Wandell1, Michael Brown2, Roger Lewis3, Harold Simon4, David Huckins5, Karen Copeland1, Brent Blumenstein5, David Spiro6 1 AspenBio Pharma, Castle Rock, CO; 2Michigan State University, Grand Rapids, MI; 3UCLA Harbor, Los Angeles, CA; 4Emergency Medicine Children’s Healthcare of Atlanta, Atlanta, GA; 5Newton Wellesley, Newton, MA; 6 Pediatric Emergency Services, Oregon Health Sciences University, Portland, OR Background: Evaluating children for appendicitis is difficult and strategies have been sought to improve the precision of the diagnosis. Computed tomography is now widely used but remains controversial due to the large dose of ionizing radiation and risk of subsequent radiation-induced malignancy. Objectives: We sought to identify a biomarker panel for use in ruling out pediatric acute appendicitis as a means of reducing exposure to ionizing radiation. Methods: We prospectively enrolled 431 subjects aged 2 to 20 years presenting in 11 U.S. emergency departments with abdominal pain and other signs and symptoms suspicious for acute appendicitis within the prior 72 hours. Subjects were assessed by study staff unaware of their diagnosis for 17 clinical attributes drawn from appendicitis scoring systems and blood samples were analyzed for CBC differential and 5 candidate proteins. Based on discharge diagnosis or post-surgical pathology, the cohort exhibited a 27.3% prevalence (118/431 subjects) of appendicitis. Clinical attributes and biomarker values were evaluated using principal component, recursive partitioning, and logistic regression to select the combination that best discriminated between those subjects with and without disease. Mathematical combination of three inflammation-related markers in a panel comprised of myeloid-related protein 8/14 complex (MRP), C-reactive protein (CRP), and white blood cell count (WBC) provided optimal discrimination. Results: This panel exhibited a sensitivity of 98% (95% CI, 94–100%), a specificity of 48% (95% CI, 42–53%), and a negative predictive value of 99% (95% CI, 95– 100%) in this cohort. The observed performance was then verified by testing the panel against a pediatric subset drawn from an independent cohort of all ages enrolled in an earlier study. In this cohort, the panel exhibited a sensitivity of 95% (95% CI, 87–98%), a specificity of 41% (95% CI, 34–50%), and a negative predictive value of 95% (95% CI, 87–98%). Conclusion: AppyScore is highly predictive of the absence of acute appendicitis in these two cohorts. If these results are confirmed by a prospective evaluation currently underway, the AppyScore panel may be useful to classify pediatric patients presenting to the emergency department with signs and symptoms suggestive of, or consistent with, acute appendicitis and thereby sparing many patients ionizing radiation. • 365 www.aemj.org S195 Video Capsule Endoscopy in the Emergency Department: A Novel Approach to Diagnosing Acute Upper Gastrointestinal Hemorrhage Andrew C. Meltzer1, Gayatri Patel1, Jeff Smith1, Payal Shah1, Amir Ali1, Roderick Kresiberg1, Meaghan Smith1, David Fleischer2 1 George Washington University, Washington, DC; 2Mayo Clinic, Scottsdale, AZ Background: Video capsule endoscopy (VCE) has an established role in diagnosing occult gastrointestinal hemorrhage and other small bowel disease. It is a novel and potentially useful method to diagnose an acute upper GI hemorrhage. Potential advantages include the ability to be performed 24 hours a day without sedation and to be interpreted at the bedside by the ED physician. Objectives: Our objectives were to demonstrate (1) ED patient tolerance for VCE, (2) the agreement of VCE interpretation between ED and GI physicians, and (3) the ability of VCE to detect active bleeding compared to subsequent upper endoscopy (EGD) or patient follow-up. Methods: This study was conducted over a 6-month period at an urban academic ED. Investigators performed VCE (Pillcam Eso2, Given Imaging) on subjects identified to have suspected acute upper GI hemorrhage (melena, hematemesis, or coffee-ground emesis within past 24 hours). Following the VCE, subjects completed a short survey regarding procedure tolerance. Approximately 30 minutes of video were recorded per subject and reviewed by four blinded physicians (two ED physicians with training in reading VCE studies but no formal endoscopic training, and two GI physicians with VCE experience.) Subjects were followed for clinical outcomes and EGD results. Results: Twenty-five subjects (mean age 52 years old, 13 female) were enrolled. No eligible subjects declined and 96% stated the VCE was well tolerated. No subjects suffered any complications related to the VCE. Between the two GI physicians, there was good agreement on the presence of fresh blood (j = 0.84). Compared to the GI physicians’ interpretation, each of the two ED physicians demonstrated good agreement regarding the presence of fresh blood (j = 0.83 and j = 0.90). The presence or absence of fresh blood on VCE showed a sensitivity of 83%, specificity of 84%, PPV of 0.63, and NPV of 0.94 compared to the gold standard of active bleeding on EGD within 24 hours (20 subjects) or patient follow-up (5 subjects). Conclusion: VCE was well tolerated in ED patients with suspected acute upper GI hemorrhage. ED physicians with VCE training were able to interpret the presence of fresh blood with good agreement with experienced GI physicians. Finally, VCE was accurate compared to the gold standard of EGD or patient follow-up for detection of active bleeding. These observations lay the groundwork for prospective multi-center studies which could allow ED physicians and GI physicians to further collaborate in the care of acute upper GI hemorrhage. (Originally submitted as a ‘‘late-breaker.’’) S196 366 2012 SAEM ANNUAL MEETING ABSTRACTS A Novel Way to Track Patterns of ED Patient Dispersal to Nearby Hospitals When a Major ED Closes Thomas Nguyen1, Okechukwu Echezona1, Arit Onyile2, Gilad Kuperman3, Jason Shapiro2 1 Beth Israel Medical Center, New York, NY; 2 Mount Sinai Medical Center, NYCLIX Inc., 3 New York, NY; Columbia-Presbyterian Hospital, New York, NY Background: There are no current studies on the tracking of emergency department (ED) patient dispersal when a major ED closes. This study demonstrates a novel way to track where patients sought emergency care following the closure of Saint Vincent’s Catholic Medical Center (SVCMC) in Manhattan by using de-identified data from a health information exchange, the New York Clinical Information Exchange (NYCLIX). NYCLIX matches patients who have visited multiple sites using their demographic information. On April 30, 2010, SVCMC officially stopped providing emergency and outpatient services. We report the patterns in which patients from SVCMC visited other sites within NYCLIX. Objectives: We hypothesize that patients often seek emergency care based on geography when a hospital closes. Methods: A retrospective pre- and post-closure analysis was performed of SVCMC patients visiting other hospital sites. The pre-closure study dates were January 1, 2010–March 31, 2010. The post closure study dates were May 1, 2010–July 31, 2010. A SVCMC patient was defined as a patient with any SVCMC encounter prior to its closure. Using de-identified aggregate count data, we calculated the average number of visits per week by SVCMC patients at each site (Hospital A-H). We ran a paired t-test to compare the pre- and post-closure averages by site. The following specifications were used to write the database queries: Of patients who had one or more prior visits to SVCMC for each day within the study return the following: a. EID: a unique and meaningless proprietary ID generated within the NYCLIX Master Patient Index (MPI). b. Age: Thru the age of 89. Persons over 90 were listed as ‘‘90 + ’’ c. Ethnicity/Race d. Type of visit: emergency e. Location of visit: specific NYCLIX site. Results: Nearby hospitals within 2 miles saw the highest number of increased ED visits after SVCMC closed. This increase was seen until about 5 miles. Hospitals >5 miles away did not see any significant changes in ED visits. See table. Conclusion: When a hospital and its ED close down, patients seem to seek emergency care at the nearest hospital based on geography. Other factors may include the patient’s primary doctor, availabilities of outpatient specialty clinics, insurance contracts, or preference of ambulance transports. This study is limited by the inclusion of data from only the eight hospitals participating in NYCLIX at the time of the SVCMC closure. Upstream Relief: Benefits On EMS Offload Delay Of A Provincial ED Overcapacity Protocol Aimed At Reducing ED Boarding Andrew D. McRae1, Dongmei Wang2, Ian E. Blanchard2, Wadhah Almansoori2, Andrew Anton1, Eddy Lang1, Grant Innes1 1 University of Calgary, Calgary, AB, Canada; 2 Alberta Health Services, Calgary, AB, Canada 367 Background: EMS offload delays resulting from ED overcrowding contribute to EMS system costs, operational inefficiencies, and compromised patient safety. Overcapacity protocols (OCP) that enhance ED outflow to inpatient units may improve EMS offload delays for arriving patients. Objectives: To examine the effect of a provincial, system-wide OCP policy implemented in December 2010 on ambulance offload delays at three urban EDs. Methods: Data were collected on all ED EMS arrivals from the metro Calgary (population 1.1 million) area to its three urban adult hospitals. The study phases consisted of the 7 months from February to October 2010 (pre-OCP) compared against the same months in 2011 (post-OCP). Data from the EMS operational database and the Regional Emergency Department Information System (REDIS) database were linked. The primary analysis examined the change in EMS offload delay defined as the time from EMS triage arrival until patient transfer to an ED bed. A secondary analysis evaluated variability in EMS offload delay between receiving EDs. Table - Abstract 366: SVCMC Patients Visiting Other Hospitals Pre and Post Closure Site SVCMC Hospital Hospital Hospital Hospital Hospital Hospital Hospital Hospital A B C D E F G H Distance from SVCMC (miles) Pre-Closure Patients/week Post-Closure Patients/week Changes in the number of visits/week from SVCMC NA 1.3 2.0 2.8 5.0 5.7 7.4 8.8 11.2 354.8 101.3 39.5 60.8 25.2 38.8 14.2 69.5 5.5 460 177.5 53.5 68.8 30.8 38.1 14.5 70.8 6.0 NA 76.2 14 8.0 5.6 )0.7 0.3 1.3 0.4 Percentage change (p-value) NA 75.2% (<0.0001) 35.4% (<0.0001) 13.1% (0.0179) 22.2% (0.0068) )1.80% (0.7849) )2.10% (0.8521) 1.86% (0.4679) 9.09% (0.4669) ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 Results: 11431 patients had linked data in both the EMS and REDIS databases. The mean EMS offload delay time was reduced following OCP implementation. Mean EMS offload delay decreased by 18.2 minutes (95%CI 16.4–19.9) from 33.5 minutes to 15.8 minutes. The decrease in EMS offload delay following OCP implementation was observed at all three receiving EDs to varying degrees. At site one, which has the highest acuity, offload delay improved by 7.5 minutes, from 14.4 minutes (95%CI 13.1–15.7) to 6.9 minutes (95%CI 6.1–7.7). At site two, which has the next-highest acuity, offload delay improved by 22.0 minutes from 36.8 minutes (95%CI 33.7–39.9) to 14.8 minutes (95%CI 13.8–16.0). At site three, which has the lowest acuity, offload delay improved by 29.6 minutes from 59.2 (95%CI 54.9–63.7) minutes to 29.6 minutes (95%CI 27.4– 31.8). Conclusion: Implementation of a regional overcapacity protocol to reduce ED crowding was associated with an important reduction in EMS offload delay, suggesting that policies that target hospital processes have bearing on EMS operations. Variability in offload delay improvements is likely due to site-specific issues, and the gains in efficiency correlate inversely with acuity. 368 What is the Impact of a Rapid Assessment Zone on Wait Times to Care for the Acute Care Unit of the Emergency Department? Alex Guttman, Marc Afilalo, Antoinette Colacone, Xiaoqing Xue, Nathalie Soucy, Eli Segal, Bernard Unger Jewish General Hospital, McGill University, Montreal, QC, Canada Background: Timely access to ED care is a serious and persistent problem that continues to challenge health care providers to identify new management strategies that optimize patient flow. Objectives: Evaluate the effect of a Rapid Assessment Zone (RAZ) on wait times to cubicle access and nurse • www.aemj.org S197 and physician assessment for patients directed to the Acute Care Unit (ACU) of the ED. Methods: A pre-post intervention study was conducted in the ED of an adult university teaching hospital in Montreal (annual visits = 69 000). The RAZ unit (intervention), created to offload the ACU of the main ED, started operating in January, 2011. Using a split flow management strategy, patients were directed to the RAZ unit based on patient acuity level (CTAS code 3 and certain code 2), likelihood to be discharged within 12 hours, and not requiring an ED bed for continued care. Data were collected weekdays from 9:00 to 21:00 for 4 months (September December 2008) (pre-RAZ) and for 1.5 months (February - March 2011) (post-RAZ). In the ACU of the main ED, research assistants observed and recorded cubicle access time, and nurse and physician assessment times. Databases were used to extract socio-demographics, ambulance arrival, triage code, chief complaint, triage and registration time, length of stay, and ED occupancy. Multiple linear regression analysis was used to compare the wait times (calculated from Triage-End Time) between pre-RAZ and post-RAZ periods with adjustment of potential confounding factors: age, triage code and ED occupancy (at Triage-End Time of a new patient). Results: During the pre-RAZ and post-RAZ periods, the ACU received 1692 and 876 visits respectively, with mean age (±SD) 68 (±18) vs. 70 (±17); Triage code 1–2: 30% vs. 35%; ambulance arrival 36% vs. 46%; and % ED occupancy 115% vs. 159%. ED staffing was re-distributed but not increased during the post-RAZ period and hospital admission policy remained unchanged. Compared to pre-RAZ, the post-RAZ period wait times (in minutes) to cubicle access, nurse assessment, and physician assessment had decreased on average by 50 (95% CI: 41–60), 46 (95% CI: 38–55), and 22 (95% CI: 13–31) respectively. Other factors associated with these wait times are ED occupancy, age, and triage code. Conclusion: Implementation of the RAZ unit resulted in a significant reduction in wait time to cubicle access, nurse and physician assessment for patients directed to the ACU of the main ED. S198 369 2012 SAEM ANNUAL MEETING ABSTRACTS Factors Influencing Completion of a Follow-Up Telephone Interview of Emergency Department Patients One Week after ED Visit Sara C. Bessman1, Julius C. Pham1, Ru Ding2, Melissa L. McCarthy2, EDIP Study Group1 1 Johns Hopkins University School of Medicine, Baltimore, MD; 2George Washington University Medical Center, Washington, DC Background: Telephone follow-up after discharge from the ED is useful for treatment and quality assurance purposes. ED follow-up studies frequently do not achieve high (i.e. ‡ 80%) completion rates. Objectives: To determine the influence of different factors on the telephone follow-up rate of ED patients. We hypothesized that with a rigorous follow-up system we could achieve a high follow-up rate in a socioeconomically diverse study population. Methods: Research assistants (RAs) prospectively enrolled adult ED patients discharged with a medication prescription between November 15, 2010 and September 9, 2011 from one of three EDs affiliated with one health care system: (A) academic Level I trauma center, (B) community teaching affiliate, and (C) community hospital. Patients unable to provide informed consent, non-English speaking, or previously enrolled were excluded. RAs interviewed subjects prior to ED discharge and conducted a telephone follow-up interview 1 week later. Follow-up procedures were standardized (e.g. number of calls per day, times to place calls, obtaining alternative numbers) and each subject’s follow-up status was monitored and updated daily through a shared, web-based data system. Subjects who completed follow-up were mailed a $10 gift card. We examined the influence of patient (age, sex, race, insurance, income, marital status, usual major activity, education, literacy level, health status), clinical (acuity, discharge diagnosis, ED length of stay, site), and procedural factors (number and type of phone numbers received from subjects, offering two gift cards for difficult to reach subjects) on the odds of successful followup using multivariate logistic regression. Results: Of the 3,940 enrolled, 45% were white, 59% were covered by Medicaid or uninsured, and 44% reported an annual household income of <$26,000. 86% completed telephone follow-up with 41% completing on the first attempt. The table displays the factors associated with successful follow-up. In addition to patient demographics and lower acuity, obtaining a cell phone or multiple phone numbers as well as offering two gift cards to a small number of subjects increased the odds of successful follow-up. Conclusion: With a rigorous follow-up system and a small monetary incentive, a high telephone follow-up rate is achievable one week after an ED visit. Table - Abstract 369: Adjusted OR and 95% CI for Significant Predictors of Completing Follow-Up (N = 3386) Characteristic Age (10 year increase) Female vs. Male Private vs. Medicaid, Medicare, Self-Pay Completed some college/ beyond vs. 12th grade/below Triage acuity 3–5 vs. acuity 1–2 Multiple phones vs. home/ work/other only Cell phone only vs. home/work/ other only Policy change (2 gift cards for difficult to reach subjects) 370 Follow-Up Completion OR 95% CI 1.1 1.4 1.6 1.1, 1.2 1.2, 1.7 1.3, 2.0 1.5 1.2, 1.9 1.9 1.8 1.2, 2.8 1.3, 2.5 1.5 1.2, 1.9 1.6 1.3, 2.0 Effect of the Implementation of an Electronic Clinical Decision Support Tool on Adherence to Joint Commission Pneumonia Core Measures in an Academic Emergency Department Michael A. Gibbs1, Michael R. Baumann2, James Lyden2, Tania D. Strout2, Daniel Knowles2 1 Carolinas Medical Center, Charlotte, NC; 2 Maine Medical Center, Portland, ME Background: In 2005, the Centers for Medicare and Medicaid Services (CMS) introduced a series of ‘‘core measures’’ designed to standardize the care of hospitalized patients with pneumonia (PNA). Several core measures are related to care provided in the emergency department (ED), where work flow and complex patient presentations may make adherence difficult. Objectives: To evaluate the effect of the implementation of an enhanced electronic clinical decision-support tool on adherence to CMS pneumonia core measures. Methods: An interrupted time-series design was used to evaluate the study question. Data regarding adherence with the following pneumonia core measures were collected pre-and post-implementation of the enhanced decision-support tool: blood cultures prior to antibiotic, antibiotic within 6 hours of arrival, appropriate antibiotic selection, and mean time to antibiotic administration. Prescribing clinicians were educated on the use of the decision-support tool at departmental meetings and via direct feedback on their cases. Results: During the 33-month study period, complete data were collected for 1185 patients diagnosed with CAP: 613 in the pre-implementation phase and 572 post-implementation. The mean time to antibiotic administration decreased by approximately one minute from the pre- to post-implementation phase, a change that was not statistically significant (p = 0.824). The proportion of patients receiving blood cultures prior to antibiotics improved significantly (p < 0.001) as did the proportion of patients receiving antibiotics within 6 hours of ED arrival (p = 0.004). A significant improvement in appropriate antibiotic selection was noted with 100% of patients experiencing appropriate selection in ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 the post-phase, p = 0.0112. Use of the available support tool increased throughout the study period, v2 = 78.13, df = 1, p < 0.0001. All improvements were maintained 15 months following the study intervention. Conclusion: In this academic ED, introduction of an enhanced electronic clinical decision support tool significantly improved adherence to CMS pneumonia core measures. The proportion of patients receiving blood cultures prior to antibiotics, antibiotics within 6 hours, and appropriate antibiotics all improved significantly after the introduction of an enhanced electronic clinical decision support tool. 371 Continued Rise In The Use Of Midlevel Providers In US Emergency Departments, 1993 To 2009 David F.M Brown, Ashley F. Sullivan, Janice A. Espinola, Carlos A. Camargo Jr. Massachusetts General Hospital, Boston, MA Background: ED visits in the US have risen dramatically over the past two decades. In order the meet the growing demand for emergency care, mid-level providers (MLPs) - both physician assistants (PAs) and nurse practitioners (NPs) - have been utilized in EDs in various ways. We previously demonstrated a striking increase in MLP utilization between 1993 and 2005. The extent to which MLPs currently are used in US EDs and the degree to which this use has changed since 2005 are not known. Objectives: To test the hypothesis that MLP usage in US EDs continues to rise. Methods: We analyzed ED visits from the National Hospital Ambulatory Medical Care Survey (NHAMCS) to identify those seen by mid-level providers (MLP). MLP visits were defined as those seen by PAs and/or NPs, with or without the involvement of physicians. Trends in all MLP visits were examined over the 17-year study period. Also, MLP-only visits, defined as visits where the patient was seen by a MLP without being seen by a physician, were compared with those seen by physicians only. We compared weighted proportions with 95%CI and analyzed trends using weighted logistic regression. Results: During 1993 to 2009, 8.4% (95%CI, 7.6–9.2%) of all US ED visits were seen by MLPs; 6.3% (95%CI, 5.5–7.0%) were seen by PAs and 2.5% (95%CI, 2.1– 2.8%) by NPs. These summary data include marked changes in MLP utilization. During the 17-year study period, PA visits rose more than 3-fold, from 2.9% to 9.9%, while NP visits rose more than 4-fold, from 1.1% to 4.7% (both Ptrend<0.001). Together, MLP visits accounted for approximately 15% of all US ED visits. Among all ED visits involving MLPs during the study period, most (58%) were seen by MLPs with physicians, while 42% were seen only by MLPs only. Compared to physician-only visits, those seen by MLPs only arrived by ambulance less frequently (6% vs. 16%), had lower urgent acuity (34% vs. 60%), and were admitted less often (3% vs. 14%). Conclusion: MLP use has increased in US EDs, with increases seen for both PAs and NPs. By 2009, approximately 1 in 7 ED visits involved a MLP, with a • www.aemj.org S199 substantial number seen by MLPs only. Although ED visits seen by MLPs only are generally of lower acuity, their involvement in some ED urgent visits and those requiring admission confirms that the role of MLPs extends beyond minor presentations. 372 Should Osteopathic Students Applying to Emergency Medicine Take the USMLE Exam? Moshe Weizberg, Dara Kass, Abbas Husain, Jennifer Cohen, Barry Hahn Staten Island University Hospital, Staten Island, NY Background: Board scores represent an important aspect of an emergency medicine (EM) residency application. Whereas allopathic (MD) students take the United States Medical Licensing Examination (USMLE), osteopathic (DO) students take the Comprehensive Osteopathic Medical Licensing Examination (COMLEX). It is difficult to compare these board scores. Previous literature proposed an equation to predict USMLE scores based on COMLEX. Recent analyses suggested that this may no longer be accurate. DO students applying to allopathic programs frequently ask whether they should take USMLE in addition to COMLEX. Objectives: #1: Compare the likelihood of MD and DO students to match in allopathic EM residencies. #2: Compare the likelihood to match of DO applicants who took USMLE to those who did not. Methods: IRB-approved review of ERAS and NRMP data for application season 2010–2011 in conjunction with a survey of all EM residency program leadership. ERAS provided the number of DO applicants and how many reported USMLE. NRMP supplied how many DO students ranked and matched in allopathic EM programs and whether they reported USMLE. A questionnaire was sent to all allopathic EM residency programs asking about the importance of DO students taking USMLE. Results: 1,482 MD students ranked EM programs; 1,277 (86%, 95% CI 84.3–87.9) matched. 350 DO students ranked EM programs; 181 (52%, 95% CI 46.4– 57.0) matched (95% CI 29.8–39.0, p < 0.0001). 208 DO students reported a USMLE score; 126 (61%, 95% CI 53.6–67.2) matched. 142 did not report a USMLE score; 55 (39%, 95% CI 30.7–47.3) matched (95% CI 11.2–32.5, p < 0.0001). Operating characteristics of USMLE to prevent not matching: absolute risk reduction = 21.9%, relative risk reduction = 55.6%, number needed to ‘‘take’’ = 4.6. Programs were surveyed about the importance of DO students taking USMLE: extremely important 40%, somewhat important 38%, not at all important 22%. Conclusion: In the 2010–2011 application season, MD students were more likely than DO students to match in an allopathic EM residency. DO students who took USMLE were more likely to match than those who did not. DO students applying to allopathic EM programs should consider taking USMLE to improve their chances of a successful match. Limitations: Single application season, students may not have reported USMLE, S200 2012 SAEM ANNUAL MEETING ABSTRACTS positions obtained outside the match, factors other than boards may have contributed. 373 Emergency Medicine Residents’ Association (EMRA) Emergency Medicine Qualifying and Certification Exam Preparation Survey Todd Guth University of Colorado, Aurora, CO Background: Emergency medicine (EM) residency graduates need to pass both the written qualifying exam and oral certification exam as the final benchmark to achieve board certification. The purpose of this project is to obtain information about the exam preparation habits of recent EM graduates to allow current residents to make informed decisions about their individual preparation for the ABEM written qualifying and oral certification exams. Objectives: The study sought to determine the amount of residency and individual preparation, to determine the extent of the use of various board review products, and to elicit evaluations of the various board review products used for the ABEM qualifying and certification exams. Methods: Design: An online survey instrument was used to ask respondents questions about residency preparation and individual preparation habits, as well as the types of board review products used in preparing for the EM boards. Participants: As greater than 95% of all EM graduates are EMRA members, an online survey was sent to all EMRA members who have graduated for the past three years. Observations: Descriptive statistics of types of preparation, types of resources, time, and quantitative and qualitative ratings for the various board preparation products were obtained from respondents. Results: A total of 520 respondents spent an average of 9.1 weeks and 15 hours per week preparing for the written qualifying exam and spent an average of 5 weeks and 7.8 hours per week preparing for the oral certification exam. In preparing for the written qualification exam, 90% used a preparation textbook with 16% using more than one textbook and 47% using a board preparation course. In preparing for the oral qualifying exam, 56% used a preparation textbook while 34% used a preparation course. Sixty-seven percent of respondents reported that their residency programs had a formalized written qualifying exam preparation curriculum of which 48% was centered on the annual in-training exam. Eight-five percent of residency programs had a formalized oral certification exam preparation. Respondents reported spending on average $715 preparing for the qualifying exam and $509 for the certification exam. Conclusion: EM residents spend significant amounts of time and money and make use of a wide range of residency and commercially available resources in preparing for the ABEM qualifying and certification exams. 374 Use Of The Multiple Mini Interview (MMI) For Emergency Medicine Resident Selection: Acceptability To Participants And Comparison With Application Data Laura R. Hopson1, Eve D. Losman1, R. Brent Stansfield1, Taher Vohra2, Danielle TurnerLawrence3, John C. Burkhardt1 1 University of Michigan, Ann Arbor, MI; 2 Henry Ford Hospital, Detroit, MI; 3William Beaumont Hospital, Royal Oak, MI Background: Communication and professionalism skills are essential for EM residents but are not wellmeasured by selection processes. The Multiple MiniInterview (MMI) uses multiple, short structured contacts to measure these skills. It predicts medical school success better than the interview and application. Its acceptability and utility in EM residency selection is unknown. Objectives: We theorized that the MMI would provide novel information and be acceptable to participants. Methods: 71 interns from three programs in the first month of training completed an eight-station MMI developed to focus on EM topics. Pre- and post-surveys assessed reactions using five-point scales. MMI scores were compared to application data. Results: EM grades correlated with MMI performance (F(1.66) = 4:18, p < 0.05) with honors students having higher MMI summary scores. Higher third year clerkship grades trended to higher MMI performance means, although not significantly. MMI performance did not correlate with a match desirability rating and did not predict other individual components of the application including USMLE Step 1 or USMLE Step 2. Participants preferred a traditional interview (mean difference = 1.36, p < 0.0001). A mixed format was preferred over a pure MMI (mean difference = 1.1, p < 0.0001). Preference for a mixed format was similar to a traditional interview. MMI performance did not significantly correlate with preference for the MMI; however, there was a trend for higher performance to associate with higher preference (r = 0.15, t(65) = 1.19, n.s.) Performance was not associated with preference for a mix of interview methods (r = 0.08, t(65) = 0.63, n.s.). Conclusion: While the MMI alone was viewed less favorably than a traditional interview, participants were receptive to a mixed methods interview. The MMI appears to measure skills important in successful completion of an EM clerkship and thus likely EM residency. Future work will determine whether MMI performance correlates with clinical performance during residency. ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 375 Novel Comprehensive Emergency Medicine In-Training Exam Course Can Improve Residency-Wide Scores Rahul Sharma, Jeremy D. Sperling, Peter W. Greenwald, Wallace A. Carter Weill Cornell Medical College / NewYorkPresbyterian Hospital, New York, NY Background: The annual American Board of Emergency Medicine (ABEM) in-training exam is a tool to assess resident progress and knowledge. When the New York-Presbyterian (NYP) EM Residency Program started in 2003, the exam was not emphasized and resident performance was lower than expected. A course was implemented to improve residency-wide scores despite previous EM literature failing to exhibit improvements with residency-sponsored in-training exam interventions. Objectives: To evaluate the effect of a comprehensive, multi-faceted course on residency-wide in-training exam performance. Methods: The NYP EM Residency Program, associated with Cornell and Columbia medical schools, has a 4year format with 10–12 residents per year. An intensive 14-week in-training exam preparation program was instituted outside of the required weekly residency conferences. The program included lectures, pre-tests, high-yield study sheets, and remediation programs. Lectures were interactive, utilizing an audience response system, and consisted of 13 core lectures (2– 2.5 hours) and three review sessions. Residents with previous in-training exam difficulty were counseled on designing their own study programs. The effect on intraining exam scores was measured by comparing each resident’s score to the national mean for their • www.aemj.org S201 postgraduate year (PGY). Scores before and after course implementation were evaluated by repeat measures regression modeling. Overall residency performance was evaluated by comparing residency average to the national average each year and by tracking ABEM national written examination pass rates. Results: Resident performance improved following course implementation. Following the course’s introduction, the odds of a resident beating the national mean increased by 3.9 (95% CI 1.9–7.3) and the percentage of residents exceeding the national mean for their PGY year increased by 37% (95% CI 23%–52%). Following course introduction, the overall residency mean score has outperformed the national exam mean annually and the first-time ABEM written exam board pass rate has been 100%. Conclusion: A multi-faceted in-training exam program centered around a 14-week course markedly improved overall residency performance on the in-training exam. Limitations: This was a before and after evaluation as randomizing residents to receive the course was not logistically or ethically feasible. 376 Difference in Rates of Medical Board Disciplinary Action between Emergency Medicine Trained and Non-Emergency Medicine Trained Physicians in the Emergency Department David J. Kammer, Jeffery A. Kline Carolinas Medical Center, Charlotte, NC Background: Quantifying the effect of residency training and board certification in emergency medicine on the quality of practice of medicine in the emergency S202 2012 SAEM ANNUAL MEETING ABSTRACTS department remains an unmet goal. The public North Carolina Medical Board (NCMB) database archives information about training status, type of practice, and disciplinary actions from any state. Objectives: Measure the frequency and incidence of state board disciplinary actions for residency-trained, board-certified or board-eligible physicians practicing emergency medicine compared with non-residency trained practitioners of emergency medicine in the state of North Carolina. Methods: We downloaded and analyzed the North Carolina Medical Board licensing database for state medical board disciplinary actions against all registered physicians who self-reported the ED as one of their areas of medical practice. A reviewer blinded to training status counted all unique disciplinary actions from all states, and years of medical practice from the reported date of medical school graduation. Groups were compared using 95% CI for difference in independent proportions. Results: The database contained 2,195 physicians who reported EM as their area of practice. Of these, 1,626 (74%) reported ABEM or AOA board certification or reported training at an accredited EM residency, with a mean of 18.0 ± 10.5 years of practice whereas 569 (26%) reported no EM training or board certification, with a 22.3 ± 11.0 years of practice. Among the nonresidency trained, non-boarded EM physicians, the percentage of individuals with board actions against them was significantly higher (6.9% vs. 1.9%, 95% CI for difference of 5.0% = 3.1 to 7.5%), but the incidence of actions was not significant (1.3 vs. 3.4 events/ 1000 years of practice, 95% CI for difference of 2.1/ 1000 = )3/1000 to +8/1000), but the power to detect a difference was 30%. Conclusion: In this study population, EM-trained physicians had significantly fewer total state medical board disciplinary actions against them than non-EM trained physicians, but when adjusted for years of practice (incidence), the difference was not significantly different at the 95% confidence level. The study was limited by low power to detect a difference in incidence. 377 Does the Residency Selection Cycle Impact What Information Is Accessed on the Web? Jim Killeen, Gary Vilke, Theodore Chan, Leslie Oyama, Maegan Carey, Edward Castillo UCSD Medical Center, San Diego, CA Background: Information on EM residency programs is widely available and most commonly accessed through the internet. Less is known on what types of information are accessed and when such access occurs given the yearly residency selection cycle. Objectives: To determine what information is most commonly accessed as measured by internet web page views, and when that access occurs for EM residency programs. Methods: A retrospective study of all internet access to an EM department and residency website for a 1 year period (7/1/10-6/30/11). Data collected included frequency of website ‘‘hits’’, number of unique visitors, duration of view, traffic source, and specific webpages accessed over the study period. Data were stratified by quarter to determine patterns of temporal variations. Statistical analysis used chi-square tests to assess differences over quarters. Results: During the study period, there were 28,827 website visits or ‘‘hits’’, from 16,581 unique visitors. The majority of traffic source came from search engines (52.7%, primarily Google), with the remaining coming from direct traffic (32.3%) or a referring site (15.0%). Visitors spent an average of 1:50 minutes on the website and visited 2.94 separate pageviews for a total of 84,610 pageviews cumulatively for 1 year. There were 820 unique visitors to the EM residency application page. 46.8% of these occurred in the quarter prior to applications being due, followed by 21.5%, 15.1%, and 16.6% in the following quarters. The percentages of visitors for the same quarter viewing specific residency info webpages were: conferences (26.4%), current residents (28.7%), curriculum (28.2%), FAQs (29.1%), and goals (25.1%). Comparison of all other pages with the application page for numbers of visitors was significantly lower for all comparisons (p < 0.001). Conclusion: In the quarter prior to residency applications being due, almost half of the unique annual website hits to the residency application page occur, which is significantly higher than access of the specific informational pages about the residency. 378 A Focused Educational Intervention Increases Paramedic Documentation of Patient Pain Complaints Herbert G. Hern1, Harrison Alter1, Joseph Barger2, Monica Teves3, Karen Hamilton3, Leslie Mueller3 1 Alameda County - Highland, Oakland, CA; 2 Contra Costa County EMS Agency, Martinez, CA; 3American Medical Response, Concord, CA Background: Patients in the EMS setting are often underassessed and undertreated for acute painful conditions. Multiple studies document the failure to assess, quantify, or treat pain by EMS providers. Objectives: We chose pain documentation as a long term project for quality improvement in our EMS system. Our objectives were to enhance the quality of pain assessment, to reduce patient suffering and pain through improved pain management, to improve pain assessment documentation, to improve capture of initial and repeat pain scales, and to improve the rate of pain medication. This study addressed the aim of improving pain assessment documentation. Methods: This was a quasi-experiment looking at paramedic documentation of the PQRST mnemonic and pain scales. Our intervention consisted of mandatory training on the importance and necessity of pain assessment and treatment. In addition to classroom training, we used rapid cycle individual feedback and public posting of pain documentation rates (with unique IDs) for individual feedback. The categories of chief ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 complaint studied were abdominal pain, blunt injury, burn, chest pain, headache, non-traumatic body pain, and penetrating injury. We compared the pain documentation rates in the 3 months prior to intervention, the 3 months of intervention, and 3 months post intervention. Using repeated-measures ANOVA, we compared rates of paramedic documentation over time. Results: Our EMS system transported 42166 patients during the study period, of whom 15490 were for painful conditions in the defined chief complaint categories. There were 168 paramedics studied, of whom 149 had complete data. Documentation increased from 1819 of 5122 painful cases (35.5%) in Qtr 1 to 4625 of 5180 painful cases (89.3%) in Qtr 3. The trend toward increased rates of pain documentation over the three quarters was strongly significant (p < 0.001). Paramedics were significantly more likely to document pain scales and PQRST assessments over the course of the study with the highest rates of documentation compliance in the final 3-month period. Conclusion: A focused intervention of education and individual feedback through classroom training, one on one training, and public posting improves paramedic documentation rates of perceived patient pain. 379 Emergency Medical Service Providers Perspectives on Management Of The Morbidly Obese Graham Ingalsbe1, John Cienki2, Kathleen Schrank1 1 University of Miami, Miami, FL; 2Jackson Memorial Hospital, Miami, FL Background: Obesity is epidemic in the United States, with an increasing burden to the health care system. Management and transport of the morbidly obese (MO) poses unique challenges for EMS providers. Though resources are being directed to improving transport of the obese, little research exists to guide these efforts. Objectives: We sought to identify EMS providers’ perspectives on the challenges of caring for MO patients in the field and areas for improvement. Methods: We administered an anonymous, web-based survey to all active providers of prehospital transport from a large, urban, fire-based EMS system to determine the specific challenges of managing MO patients. This survey looked at various components of the transport: lifting, transport time, airway management, establishing IV access, drug administration, as well as demographics, equipment, and education needs. The survey contained yes/no, rank-order, and Likert scale questions with free response available. Results: Of approximately 550 EMS providers surveyed 203 completed responses. All had transported an MO patient. Providers felt MO patients frequently call EMS for assistance around the home (29%) and non-emergent transport to a health care facility (21%), with pain (29%) and shortness of breath (23%) as the most common emergent complaints. Of specific challenges to properly care for MO patients, 91% thought lifting and/or moving the patient was most difficult, followed by IV access, airway management, transport time, and measuring accurate vital signs. Respondents felt that transporting MO • www.aemj.org S203 patients requires at least six EMS personnel (87%) . Of EMS providers who responded, 73% had attempted IV access on a MO patient, but only 36% had ever intubated, and 25% had ever calculated or adjusted medication dosages specifically for MO patients. More than 87% felt it would be beneficial to receive more training on transport, intubation, and medication dosing for MO and 96% felt they needed more equipment. Of respondents, 89% felt that MO patients were not able to receive the same standard of care as other patients. Conclusion: Surveyed EMS providers felt that difficulties in lifting, as well as lack of experience, hinders care for MO patients in the prehospital setting, and most felt that more training would be beneficial. Surveyed EMS providers reported that MO patients are not able to receive the same standard of care as non-MO patients. (Originally submitted as a ‘‘late-breaker.’’) 380 Anaphylaxis Knowledge Among Paramedics: Results of a National Survey Ryan C. Jacobsen1, Serkan Toy2, Joseph Salomone1, Aaron Bonham3, Jacob Ruthstrom1, Matthew C. Gratton1 1 Truman Medical Center, Kansas City, MO; 2 Children’s Mercy Hospitals and Clinics, Kansas City, MO; 3University of MissouriKansas City, Kansas City, MO Background: Anaphylaxis is a potentially life-threatening emergency and prompt recognition and treatment by EMS providers can be critically important for early initiation of appropriate life-saving interventions. Anaphylaxis has been shown to be both under-recognized and under-treated by physicians with limited data on the prehospital care of anaphylaxis. Objectives: The purpose of this study was to determine how well paramedics recognize and treat anaphylaxis in an adult population. Methods: Blinded survey of a random sample of paramedics registered by the National Registry of Emergency Medical Technicians (NREMT). A SurveyMonkeyTM survey was designed and validated with local paramedics and then distributed via e-mail to a random sample of NREMT-certified paramedics. The survey contained three sections: demographic, self-assessment of confidence regarding anaphylaxis care, and cognitive assessment. Paramedics were contacted a total of four times to ensure compliance with the survey. Results: 3538 of 9655 responded (36.6% response rate). 79% were male; mean age 36; mean EMS experience 12.1 years; 78.5% full time EMS; and 30.9% fire-based, 27.9% private, 17.2% county, and 12.9% hospital-based. 98% were confident that they could recognize anaphylaxis; 97% were confident that they could treat anaphylaxis; 95.4% were confident they could use an epinephrine auto-injector (EAI). 98.9% correctly recognized a case of classic anaphylaxis while only 2.5% correctly identified the atypical case. Only 46.7% gave epinephrine as the first treatment for anaphylactic shock (airway already secured) and only 38.9% gave it IM (58.4% SQ); and 60.5% gave it in the deltoid (11.6% thigh). S204 2012 SAEM ANNUAL MEETING ABSTRACTS Conclusion: While a large percentage of paramedics recognized classic anaphylaxis, a very small percentage recognized atypical anaphylaxis; less than half chose epinephrine as the initial drug of choice and most respondents were unable to identify the correct route of administration. This survey points out a number of areas for improved education and training. 381 Thermal Medication Stress in Air Ambulances: The Mercy Saint Vincent Life Flight Experience Richard Tavernetti1, Edward Tavernetti2, David J. Ledrick1 1 MSVMC, Toledo, OH; 2University of California at Davis, Davis, CA Background: Temperature variations can degrade the potency of advanced life support medications. While mitigating the effects of temperature on medications carried by ground ambulances has been well studied, there is little information to guide helicopter emergency medical services (HEMS). To date a comprehensive evaluation of thermal medication stress in the HEMS environment has not been completed. Previous studies have been limited by their brevity, attention only to maximum and minimum temperatures, lack of replication, and variability in storage practices. Objectives: The objective of this study was to provide a high-resolution picture of medication temperatures across a large HEMS system with sufficient replication to allow for correlations to be made between thermal medication stress and specific operational practices. Methods: Our program is located in Northwest Ohio and conducts approximately 2,500 annual patient transports across five aircraft. Bags containing medications are located in the cabin and tail compartments of all aircraft. Temperature loggers were secured inside all the medication bags and temperatures were recorded every 30 minutes for one year. Results: 139,523 data points were included in the analysis. The temperature range was between 46.5C and )2.5C. The mean kinetic temperature (MKT) goal of 25C was exceeded in the May-September period with cabin MKT being significantly cooler (p < 0.003). Cabin storage was associated with lower thermal stress throughout the year. A larger cabin size appeared to be associated with lower MKT. Maximum temperatures during winter and summer were similar due to the wintertime use of heaters. Freezing temperatures (<0C) were only encountered in the tail compartments. Conclusion: Current HEMS medication storage practices are not within U.S. Pharmacopeia standards, although they appear attainable with minimal changes in operational practices. Storing medications in the aircraft’s cabin can reduce thermal medication stress. Operational procedures aimed at reducing thermal medication stress appear to be most important MaySeptember. Heaters can subject medications to extremes of temperature and require greater care in their implementation. Stock rotation protocols based on previous data from the ground ambulance setting may be inadequate for HEMS. 382 Opportunities for Emergency Medical Services Care of Syncope Brit Long, Luis Serrano, Fernanda Bellolio, Erik Hess Mayo Clinic College of Medicine, Rochester, MN Background: Emergency medical services (EMS) systems are vital in the identification, assessment, and treatment of trauma, stroke, myocardial infarction, and sepsis and improving early recognition, resuscitation, and transport to adequate medical facilities. EMS personnel provide similar first-line care for patients with syncope, performing critical actions such as initial assessment and treatment as well as gathering key details of the event. Objectives: To characterize emergency department patients with syncope receiving initial care by EMS and their role as initial providers. Methods: We prospectively enrolled patients over 18 years of age who presented with syncope or near syncope to a tertiary care ED with 72,000 annual patient visits from June 2009 to June 2011. We compared patient age, sex, comorbidities, and 30-day cardiopulmonary adverse outcomes (defined as myocardial infarction, pulmonary embolism, significant cardiac arrhythmia, and major cardiovascular procedure) between EMS and non-EMS patients. Descriptive statistics, two-sided ttests, and chi-square testing were used as appropriate. Results: Of the 669 patients enrolled, 254 (38.0%) arrived by ambulance. The most common complaint in patients transported by EMS was fainting (50.4%) or dizziness (45.7%); syncope was reported in 28 (11.0%). Compared to non-EMS patients, those who arrived by ambulance were older (mean age (SD) 64.5 (18.7), vs. 60.6 (19.5) years, p = 0.012). There were no differences in the proportion of patients with hypertension (20.0% vs 32.0%, p = 0.75), coronary artery disease (8.85% vs 15.3%, p = 0.67), diabetes mellitus (6.5% vs 9.5%, p = 0.57), or congestive heart failure (3.8% vs 6.6%, p = 0.74). Sixtynine (10.8%) patients experienced a cardiopulmonary event within 30 days. Twenty-eight (4.4%) patients who arrived by ambulance and 41 (6.4%) non-EMS patients had a subsequent cardiopulmonary adverse event (RR 1.08, 95%CI 0.68–1.69) within 30 days. The table tabulates interventions provided by EMS prior to ED arrival. Conclusion: EMS providers care for more than one third of ED syncope patients and often perform key interventions. EMS systems offer opportunities for advancing diagnosis, treatment, and risk stratification in syncope patients. Table - Abstract 382: ACADEMIC EMERGENCY MEDICINE • April 2012, Vol. 19, No. 4, Suppl. 1 383 Electronic Accountability Tools Reduce CT Overutilization in ED Patients with Abdominal Pain Angela M. Mills1, Daniel N. Holena1, Caroline Kerner1, Hanna M. Zafar1, Frances S. Shofer1, Brigitte M. Baumann2 1 University of Pennsylvania, Philadelphia, PA; 2 Cooper University Hospital, Camden, NJ Background: Abdominal pain is the most common reason for visiting an emergency department (ED), and abdominopelvic computed tomography (APCT) use has increased dramatically over the past decade. Despite this, there has been no significant change in rates of admission or diagnosis of surgical conditions. Objectives: To assess whether an electronic accountability tool affects APCT ordering in ED patients with abdominal or flank pain. We hypothesized that implementation of an accountability tool would decrease APCT ordering in these patients. Methods: Before and after study design using an electronic medical record at an urban academic ED from Jul-Nov 2011, with the electronic accountability tool implemented in Oct 2011 for any APCT order. Inclusion criteria: age >= 18 years, non-pregnant, and chief complaint or triage pain location of abdominal or flank pain. Starting Oct 17th, 2011, resident attempts to order APCT triggered an electronic accountability tool which only allowed the order to proceed if approved by the ED attending physician. The attending was prompted to enter the primary and secondary diagnoses indicating APCT, agreement with need for CT and, if no agreement, who was requesting this CT (admitting or consulting physician), and their pretest probability (0–100) of the primary diagnosis. Patients were placed into two groups: those who presented prior to (PRE) and after (POST) the deployment of the accountability tool. Primary outcome was percentage of APCT performed by group. Continuous data were compared using Mann Whitney test while categorical data were compared using Fisher’s exact test. Results: Of 1419 patients enrolled (mean age 46.7 ± 17 years, 54% male, 58% black, 33% admitted), the majority had a chief complaint of abdominal pain (81%). There were no statistically significant differences in age, sex, or race between the PRE (N = 1014; 71%) and POST (N = 405; 29%) groups. There was a significant reduction in APCT use after the tool’s implementation (37.8% PRE vs. 31.9% POST, difference 6%, 95% CI 0.4–11%) with no change in the rate of admission (33.5% PRE vs. 30.9% POST, p = 0.35). Conclusion: An electronic accountability tool with attending forced completion reduced unnecessary APCT in ED patients with abdominal pain and did not lead to increased rates of admission. Extension of these tools into other areas could lead to further reductions in unnecessary radiologic testing. • 384 www.aemj.org S205 Identification Of Patients Less Likely To Have Significant Alternate Diagnoses On CT For Renal Colic Brock Daniels, Cary Gross, Dinesh Singh, Chris Moore Yale University School of Medicine, New Haven, CT Background: Previous studies on unselected populations have reported that more than 10% of CT scans for suspected kidney stone may reveal significant alternate pathology. Objectives: To determine the relative effect on prevalence of siginificant alternate diagnoses in CT for suspected renal colic after excluding patients without flank/back pain; or with evidence of infection (fever and/or leukocytes on urine dip), active malignancy, known renal disease, or prior urologic intervention. Methods: Retrospective record review. All CT ‘‘flank pain protocol’’ (FPP) done in patients >18y.o. at an academic ED between 4/05–11/10 were electronically retrieved (n = 5379). 4585 records (85%) were randomly selected for review by trained research assistants who categorized dictated CT reports as ‘‘normal’’, ‘‘symptomatic stone’’, or ‘‘other’’. All ‘‘other’’ CT findings were reviewed along with medical records by one of three physician authors to determine if CT was diagnostic, and if so, if there was an intervention or followup (f/u) ‘‘required’’, ‘‘recommended’’, or ‘‘not needed’’. A random subset (1868/4585) of records underwent full review for exclusion criteria, as well as all records with f/u ‘‘required’’ or ‘‘recommended’’. Any categorization or exclusion that was not clear was adjudicated by the three physicians. A subset of records randomly selected from each category was blindly double reviewed by two physicians to ensure agreement. Results: Results are shown in the table. Overall, 58.2% (95% CI 56.8–59.7) of the CTs reviewed were diagnostic, with 51.8% (50.4–53.6) of all CTs revealing symptomatic kidney stones, 4.1% (3.6–4.7) intervention or f/u required, 1.4% (1.1–1.7) f/u recommended, and 1.6% (0.7–1.3) not needed. Of the 188 intervention or f/u required, 139 or 73.9% (67.2–79.7) would be screened by the exclusion rule, while on full record review 42.4% (39.2–43.6) would be excluded. The data suggest that if applied prospectively, the prevalence of alternate diagnoses requiring intervention or follow-up in CTs for renal colic could be reduced from 4.1% to under 1.6% (50 significant diagnoses out of 3101 not excluded). Conclusion: On thorough record review, less than five percent of CTs for renal colic reveal alternate diagnoses requiring intervention or follow-up. Applying a rule with basic historical data and point-of-care urine dip could identify a population where this prevalence is under 2%. Prospective study is needed. S206 2012 SAEM ANNUAL MEETING ABSTRACTS Table- Abstract 384: Review of CT FPPs (n = 4585) n= Age ± SD Female CTs diagnostic Symptomatic stone Intervention or f/u required f/u recommended f/u not needed Expected* not excluded (derived from sample) Not excluded intervention or f/u required (actual) 385 44.9 ± 15.8 2363 2670 2376 188 62 44 3101 50 % 51.5% 85.2% 51.8% 4.1% 1.4% 1.0% 57.4% 1.6% Delayed Outcomes For Patients With Suspected Renal Colic After Discharge From The Emergency Department Justin Yan, Marcia Edmonds, Shelley McLeod, Rob Sedran, Karl Theakston The University of Western Ontario, London, ON, Canada Background: Although renal colic is a relatively benign disease, it can have a significant effect on patientrelated outcomes that may not be immediately apparent during an isolated emergency department (ED) visit. Objectives: The objective of this study was to determine the burden of disease for patients with suspected renal colic after discharge from their initial ED visits. Methods: This was a prospective observational cohort study involving all adult (‡ 18 years) patients who presented to the EDs of a tertiary care centre (combined census 150,000) with suspected renal colic over a oneyear period (Oct 2010–Oct 2011). Patients were contacted by telephone by trained research personnel at 48–72 hours and 10–14 days after their ED visit to determine use of analgesics, missed days of school or work, and repeat visits to a health care provider. Electronic patient charts were reviewed at 90 days after the initial visit to determine if urologic intervention (lithotripsy, stents, etc.) was eventually required. Results: Of 397 patients enrolled, 38 (9.6%) were excluded for definite alternate diagnoses. Of 359 remaining patients, 12 (3.3%) were lost to follow-up, leaving 347 (96.7%) providing post-discharge outcomes. Ten patients (2.9%) were admitted from the ED. Of those patients contacted, 95 (27.4%) reported passing a stone within 72 hours and 31 additional patients (36.3%) reported passing a stone within 14 days. 270 (77.8%) patients required analgesia for a median (IQR) duration of 4 (2, 7) days. There were 88 (25.4%) patients who still