Abstract
A rubric was developed to assess student posters as a mechanism to evaluate learning outcomes for a senior capstone course. The analytic rubric allows for the efficient and systematic collection of data from posters by students who worked across a variety of disciplines including the physical, biological, Earth sciences, social science, and the humanities. The rubric effectively addressed a fundamental assumption and requirement put forth during rubric development, that is, it needs to be relatively easy to use without training while at the same time producing consistent results across evaluators. The overall Chronbach’s alpha of 0.80 across semesters indicates acceptable inter-rater reliability. Data generated by assessment of 106 interdisciplinary posters indicates a general, yet not statistically significant, improvement, in total scores from the spring 2014 to spring 2018, documents student proficiency, and captures the variability in the quality of the various projects. The rubric was primarily developed as a tool to inform formative assessment, but it is also a teaching tool. Its use in providing feedback and as reflection tool enhances the learning experience for students and increases the impact of the senior thesis process on their professional development. The use of a student feedback questionnaire has informed reflective instructional practice. This resulted in an increased emphasis within the capstone course on the inclusion of reference citations, use of informal writing activities, and frequency of meetings with faculty mentors. The results from our approach should be encouraging to other interdisciplinary environmental studies and science programs that seek to efficiently and effectively impact student learning outcomes and evaluate the impact of course changes over several semesters.
References
Akister J, Bannon A, Mullender-Lock H (2000) Poster presentations in social work education assessment: a case study. Innov Educ Train Int 37(3):229–233
Behar-Horenstein LS, Roberts KW, Dix AC (2010) Mentoring undergraduate researchers: an exploratory study of students’ and professors’ perceptions. Mentor Tutor Partnership Learn 18(3):269–291
Brown S Knight P (1994) Assessing learners in higher education. Kogan Page, London, pp 161
Connelly LM (2018) Designing effective conference posters. Medsurg Nurs 27(1):64–65
Diller KR, Phelps SF (2008) Learning outcomes, portfolios, and rubrics, oh my! Authentic assessment of an information literacy program. Portal Libr Acad 8(1):75–89. https://doi.org/10.1353/pla.2008.0000
Gosselin D, Cooper S, Bonnstetter RJ, Bonnstetter B (2013) Exploring the assessment of 21st century professional competencies of undergraduate students in environmental studies through a business – academic partnership. J Environ Stud Sci 3:359–368. https://doi.org/10.1007/s13412-013-0140-1.PlusErratum
Harvey M, Coulson D, McMaugh A (2016) Towards a theory of the ecology of reflection: reflective practice for experiential learning in higher education. J Univ Teach Learn Pract 13(2):2
Hawthorne J, Zerr R, Kelsch AV (2018) Assessment of general education: an unexpected (but effective) faculty development opportunity. Chapter 7. New Directions for Teaching and Learning: 155. https://doi.org/10.1002/tl.20304
Howard C (2015) The role of posters as a means of summative assessment. Worcester Journal of Teaching and Learning:10
Kinikin JN, Hench K (2012) Poster presentations as an assessment tool in a third/college level information literacy course: an effective method of measuring student understanding of library research skills. J Inform Lit 6(2)
Kuh GD, Ikenberry SO, Jankowski N, Cain TR, Ewell PT, Hutchings P, Kinzie J (2015) Using evidence of student learning to improve higher education. Jossey-Bass, San Francisco
MacIntosh-Murray A (2007) Poster presentations as a genre in knowledge communication: a case study of forms, norms, and values. Sci Commun 28(3):347–376. https://doi.org/10.1177/1075547006298251
Menke JL (2014) Implementation of online poster sessions in online and face-to-face classrooms as a unique assessment tool. J Chem Educ 91(3):414–416
Moskal BM (2003) Recommendations for developing classroom performance assessments and scoring Rubrics. Pract Assess Res Evaluation 8(14). http://PAREonlinenet/getvnasp?v=8&n=14. Accessed 20 Mar 2019
Palomba C, Banta TW (1999) Assessment essentials: planning, implementing, and improving assessment in higher education. Jossey-Bass, Inc, San Francisco
Skocpol T (2009) Foreword. In Assessment in Political Science. ed. Deardorff MD, Hamann K., and Ishiyama J., xi–xiii. Washington. American Political Science Association
Taber KS (2018) The use of Cronbach’s alpha when developing and reporting research instruments in science education. Res Sci Educ 48(6):1273–1296
Thiry H, Laursen SL (2011) The role of student-advisor interactions in apprenticing undergraduate researchers into a scientific community of practice. J Sci Educ Technol 20:771–784. https://doi.org/10.1007/s10956-010-9271-2
Todd M, Bannister P, Clegg Sc (2004) Independent inquiry and the undergraduate dissertation: perceptions and experiences of final-year social science students. Assessment & Evaluation in Higher Education 29(3):335–355
Tziner A, Murphy KR, Cleveland JN (2005) Contextual and rater factors affecting rating behavior. Group Organ Manag 30:89–98
Volz T, Saterbak A (2009) Students’ strengths and weaknesses in evaluating technical arguments as revealed through implementing Calibrated Peer Review™ in a bioengineering laboratory. Across Discip 6(1)
Weaver KF, Morales V, Nelson M, Weaver PF, Toledo A, Godde K (2016) The benefits of peer review and a multisemester capstone writing series on inquiry and analysis skills in an undergraduate thesis. CBE Life Sci Educ 15(4):ar51
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Gosselin, D.C., Golick, D. Posters as an effective assessment tool for a capstone course. J Environ Stud Sci 10, 426–437 (2020). https://doi.org/10.1007/s13412-020-00612-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13412-020-00612-x