Praslova Assessment of Long-Term Outcomes Poster Handout
Transcription
Praslova Assessment of Long-Term Outcomes Poster Handout
Assessment of long-term educational outcomes: Insights from the Kirkpatrick’s four level model of training criteria. Ludmila N. Praslova Vanguard University of Southern California Accrediting organizations, governments, and employers increasingly emphasize the long-term outcomes of education and expect institutions of Higher Education to prepare students for the labor force through development of relevant competencies, while also preparing them for civil participation (Geary Schneider, 2012; Toutkoushian, 2005; Voorhees and Harvey 2005). Documenting student achievement at the individual and societal levels is a complex task. Reactions are easily measured with indirect methods of assessment, but are considered insufficient for evaluation of outcomes (Kirkpatrick, 1959, 1976, 1996). Institutions are making progress in assessing learning with direct methods. Behavior and especially results criteria, while extremely important, are very difficult to evaluate and conceptualize. While organizations focus on organizational outcomes, current evaluation on Higher Education effectiveness is often focused on both societal and The classic model used for evaluation of training in individual outcomes (e.g., individual Return on business organizations, the Kirkpatrick’s four level Investment (ROI) (Gallup/Purdue, 2014; Chronicle of model of training criteria (Kirkpatrick, 1959, 1976, Higher Education, 2015). 1996), might provide a versatile framework for assessment in Higher Education (Praslova, 2010). Comparison between evaluation of training in organizations Higher Education suggests several The four levels of evaluation in Kirkpatrick’s model important questions to consider in evaluation of are reaction criteria, learning criteria, behavior results-level criteria: criteria, and results criteria. 1) Criteria In organizations In Higher Education Reaction (internal to program) trainee affective reactions & utility judgments student affective reactions & utility judgments Indirect assessment/student opinion surveys Learning (internal to program) direct measures of learning outcomes, typically knowledge tests or performance tasks direct assessment of learning outcomes (tests, papers, performance tasks, other graded work, national tests) Behavior (external to program) measures of actual on-the-job performance: supervisor ratings or objective indicators of performance/job outputs. Beyond course or program level. Evidence of student use of knowledge and skills learned early in the program in subsequent work, e.g., research projects or creative productions, application of learning during internship, through community involvement, and in the workplace. Other behaviors outside the context in which the initial learning occurred. Related documentation. Results (external to program) productivity gains, increased customer satisfaction, employee morale for management training, profit value gained by organization Table 1; in part based on Praslova, 2010. Individual level/Societal level: alumni career success, graduate school admission, service to society, personal stability, fulfillment etc. Alumni surveys, employer feedback, samples of scholarly or artistic accomplishments, notices of awards, recognition of service, etc. [email protected] 2) 3) 4) 5) 6) Are individual level vs. group-level results the appropriate level of analysis? If both, then what is the appropriate model to account for interactions between these? How does behavior translate into results, and how learning process can best influence behavior? Do measures such as jobs, ROI, and earnings tap into results level as intended by the proponents? Is national level productivity an appropriate outcome to assess? How much variance in individual outcomes is explained by education, individual characteristics, individual behavior, and contextual influences? How much variance in societal outcomes is explained by education vs. contextual influences of the national political and economic model and the global trends? Careful consideration of these questions should benefit attempts to evaluate the effectiveness of Higher Education. References: Gallup/Purdue (2014). Great Jobs, Great Lives. The 2014 Gallup-Purdue Index Report. Geary Schneider, C. (2012). Moving Democratic Learning from the Margins to the Core. Remarks Delivered at the White House. http://www.aacu.org/press_room/cgs_perspectives/documents/MovingDemocraticLearningfromtheMarginstotheCore.pdf Kirkpatrick, D. L. (1959). Techniques for evaluating training programs. Journal of the American Society of Training Directors, 13, 3– 9. Kirkpatrick, D. L. (1976). Evaluation of training. In R. L. Craig (Ed.), Training and development handbook: A guide to human resource development (2nd ed., pp. 301–319). New York: McGraw-Hill. Kirkpatrick, D. L. (1996). Invited reaction: reaction to Holton article. Human Resource Development Quarterly, 7, 23–25. Praslova, L. (2010). Adaptation of Kirkpatrick’s four level model of training criteria to assessment of learning outcomes and program evaluation in Higher Education. Educational Assessment, Evaluation and Accountability, 22, 215-125, Springer Netherlands. DOI: 10.1007/s11092-010-9098-7 Toutkoushian, R. K. (2005). What can institutional research do to help colleges meet the workforce needs of states and nations? Research in Higher Education, 46, 955–984. doi:10.1007/s11162-005-6935-5. Voorhees, R. A., & Harvey, L. (2005). Higher education and workforce development: a strategic role for institutional research. New Directions for Institutional Research, 128, 5–12.