=Paper=
{{Paper
|id=Vol-1633/ws1-paper4
|storemode=property
|title=Teaching Students to Give and To Receive: Improving Disciplinary Writing Through Peer Review
|pdfUrl=https://ceur-ws.org/Vol-1633/ws1-paper4.pdf
|volume=Vol-1633
|authors=Julia Morris,Jennifer Kidd
|dblpUrl=https://dblp.org/rec/conf/edm/MorrisK16
}}
==Teaching Students to Give and To Receive: Improving Disciplinary Writing Through Peer Review==
Teaching Students To Give And To Receive: Improving Interdisciplinary Writing Through Peer Review Julia Morris Jennifer Kidd Old Dominion University Old Dominion University 5115 Hampton Boulevard 5115 Hampton Boulevard Norfolk, Virginia 23529 Norfolk, Virginia 23529 1 (704) 490 2068 1 (757) 683 3248 Jmorr005@odu.edu Jkidd@odu.edu ABSTRACT collaboration of six faculty members. These faculty, representing The context for this study is a multi-disciplinary collaboration of four disciplines, implemented peer review into their classrooms in six faculty members using peer review in their respective order to improve undergraduate student writing. Students disciplines with the goal of improved student writing. Faculty engaged in formative peer review and reflected on their members developed their own assignments and methods for experience in an online survey. The results of the initial fall 2015 implementing peer review, but each followed the same guidelines. study showed variation between classes, indicating the effect of Students submitted drafts to peers who made comments and used the instructor and assignment for the peer review process. a rubric to provide formative feedback. The instructors used a The results from the spring 2016 student data show a more variety of tools to support peer review, including Google Drive, complete picture of how peer review can be implemented to Blackboard, and Expertiza, a dedicated peer-review system. promote student investment in this autonomous learning process. Students reflected on the peer review process in an online survey Further, results show that students’ can, and do, recognize that after each round of peer review. The survey results varied peer review should promote substantive changes in their writing. considerably between the classes, suggesting the importance of This study’s hypothesis stated that feedback from the fall 2015 the instructor, assignment, and peer review process. There were students would facilitate increased satisfaction with and utility of also common themes that emerged across courses, such as the the peer review process in spring 2016 students. This paper common value of giving reviews. This paper examines one describes the progression of responses from education students in participating faculty member’s fall 2015 and spring 2016 one class between the fall and spring semesters. education course and how students’ perceptions of peer review evolved positively across the two semesters. 2. REVIEW OF LITERATURE Keywords Once considered the “neglected variable in education,” peer review and student interaction have the potential to transform Peer review, feedback, student writing, multi-disciplinary writing practices in higher education [1]. Comer et al. discuss how peer- to-peer interactions in two MOOCs enhance learners’ 1. INTRODUCTION understanding of course learning objectives and highlight the Technology enhanced peer review enables students the value of online learning environments. This is due to the fact that opportunity to work collaboratively with each other, more than interactions occur almost entirely in written form. Comer et al.’s ever before. Online peer review systems have increased students’ findings indicate that peer review fosters a networked learning opportunities to provide and be given expeditious feedback. experience as online interactions require the primary form of Writers are able to benefit from multiple perspectives more communication to occur through writing, thus improving both immediately through the various modalities of technology, further course specific and composition skills. expanding upon present classroom peer review processes. Peer Lui and Sadler [2] also demonstrate the success of technology- review promotes learning autonomy for students during the infused peer review for higher education students by comparing writing process. The positive effects of learning for both face-to-face peer review with online peer review. The authors find reviewers and reviewees, facilitated by advances in classroom that technology greatly enhanced the number of comments, the technology, have allowed students to improve their writing number of revision-oriented comments, and the number of through peer review. revisions made by students after engaging the technology- This study describes the data from one faculty member who enhanced peer review process. However, the incorporation of participated in this yearlong exploration into student perceptions technology builds upon but does not eradicate pre-existing issues of peer review. This study describes the second phase of an prevailing in face-to-face peer review. One of the most prevalent exploration into these perceptions through the multi-disciplinary points discussed is the concept of anonymity. The literature is varied in its stance on anonymity. Lee [3] discusses how anonymity is peer review’s most significant inhibitor as it does not encourage self-regulated learning. Similarly, Huahui et al. [4] found that non-anonymous peer review partners encourage a social presence, born of an optimal level of participation and interaction, which promoted a “more supportive learning environment” [4, p.812]. However, in contrast, Raes et al. [5] concludes that increased anonymity can decrease peer pressure and increase comfortability with peer review. Similarly, Lu and (pre-writing) assignments or they could be complete papers in an Bol [6] found students participating in anonymous peer review initial state. Reviewers provided formative feedback via outperformed students in identifiable peer review, and provided comments and a rubric. Authors made changes as desired and more critical feedback to their peers. With these antithetical submitted a final product. This was graded by either the conclusions, the best approach to anonymity in peer review is still instructor, or by students via a summative round of peer undetermined. evaluations. The instructors used a variety of tools to support peer The skills required and refined by peer review prove advantageous review including Google Docs, Blackboard, and Expertiza, a across classrooms. It is understood that peer review has ventured peer-review system developed at North Carolina State University. beyond English and education courses; it is now a regularly Students reflected on their peer review experience in an utilized tool in computer programming, engineering, anonymous online survey with scaled and open-ended items after environmental science, and business and entrepreneurship courses each round of peer review. To assess the impact of peer review on [7,8]. However, as peer review transcends disciplines, it remains student writing, instructors graded a selection of student a cooperative effort between students and faculty. Peer review is submissions before and after the peer reviews using a common no longer an educational arrangement devised to benefit students interdisciplinary writing rubric. This enabled them to judge the and alleviate staff workload, but instead, a “rather complex quality of student writing and to assess the extent to which undertaking,” that merges students’ ability to assess and students’ students revised (and hopefully, improved) their writing after knowledge of course content [7, p. 181]. With the unavoidability participating in peer reviews. Faculty also completed mid and end of variability amongst students in any given classroom, faculty of semester surveys to gauge their perceptions of the costs and must mediate reviews in order to assure students of reliability. benefits of peer review. As the data collection from faculty is still The nature of student perceptions of peer review is limitedly underway, this paper presents the results of the student reflection studied. Studies at an Australian university found that 90% of survey, focusing on the data from one instructor. surveyed students expect peer review to be helpful [7]. Half of The instructor discussed in this study has taught these students expressed anxieties about peer review due to an foundational educational assessment to pre-service teachers for uncertainty of how to be constructive without seeming too harsh. 10(+) semesters. After implementing peer review during the first Conversely, some students were worried that reviews they semester of this study and reflecting on the corresponding fall received would be too nice and thus provide little substance. Post data, changes were made in order to respond to students’ concerns peer review, the number of students who considered peer review and suggestions for the spring semester. Several changes were to be helpful dropped to 70%, indicating what Mulder et al. refer implemented to bring about change. Specifically, instead of a to as the “modest downward shift” in positive perceptions of peer variety of tools being used (i.e. Google Docs, Expertiza, face-to- review. No study has yet simultaneously considered student face), students conducted all peer review in one consistent tool. perceptions in conjunction with what feedback is incorporated The assignments were all submitted, peer reviewed, received, and affects learning outcomes [8]. In a recent study, students’ most meta-reviewed through Expertiza. The instructor also added a critical impressions found formative peer review to be plagued calibration assignment where students were given the opportunity with inconsistencies in quality and reliability [6]. As a result, less to simulate peer review, comparing their attempts at review with than one-third of surveyed first-year students felt they had helpful the instructor’s expert review. Students were also guided through feedback via peer review, complaining of “unreliable” and the peer review process gradually. Peer review stages were time “inconsistent credibility.” More than half of students reported restricted meaning students could not work ahead of the current that they were disillusioned with partners who lacked expertise. state of review (i.e. submission, peer review, and/or meta-review). These students instead preferred an “expert review” from faculty This is a semester-long course and therefore methods described or staff. These impressions, according to the authors, suggest that from the fall 2015 semester were amended for the new students peer review can be effective, as students are looking for ways to enrolled in the spring 2016 semester. The results discussed reflect improve their artifacts, but further research must be done that the trend in data from fall to the spring semester. explores more effective implementation. When this study is complete it will provide new insight into students’ attitudes and behaviors, helping to delineating what type of feedback students 4. RESULTS find most helpful and to uncover the processes and procedures that prompt students to revise their work. This preliminary report 4.1 Peer Review Design of student attitudes begins this work. This excerpt of the study into student perceptions of peer reviews discusses results from an education course geared towards pre- service teachers. Students were asked to participate in multiple 3. METHODOLOGY rounds of formative peer review, culminating in a final round of Six faculty members agreed to incorporate peer review into their summative assessment where peers graded each other’s developed undergraduate courses with the intention of improving student lesson. Students were asked to create a digital lesson to teach writing. Although two of the participants had significant peers about a topic in education. Students completed multiple experience peer review, the intention was to create a learning stages for this scaffolded assignment, each followed by a round of community with the faculty learning from and with one another as peer reviews: a 3-paragraph research essay on their topic, they designed and later revised writing prompts, rubrics, and peer followed by formative peer reviews; an initial draft of their lesson, review protocols. Each faculty member developed their own followed by formative peer reviews; and, their final lesson, assignments and peer review process, but followed the same submitted for summative peer review during which students general guidelines. Students were required to submit a draft(s) on assigned grades to one another. All peer review was conducted a major writing assignment (instructors could institute a single or through Expertiza and was designed to be anonymous. multiple rounds of peer review). The drafts could be preliminary Figure 1: Peer Review Procedures by Course o Round 1 52% (N=91) Subject/Course Number Technology Purpose & (N=86) o Round 2 44% of Process of o Round 2 43% (N=74) Responses Peer Review (N=71) o Round 3 o Round 3 0% (N= 84) (N/A) Anonymity Education 331 Expertiza Three Anonymity o I knew who Rounds: o I knew who reviewed my work 1 & 2 were reviewed my work Yes 2% formative; 3 Yes 37% (N=4) was (N=58) No 98% summative No 63% (N=163) (N=102) o I knew whose -Round 1 = 3- o I knew whose work I reviewed paragraph work I reviewed Yes 14% essay Yes 55% (N=24) (N=88) No 86% -Round 2 = No 45% (N=143) complete draft (N=72) -Round 3 = final draft 4.3 Quantitative Survey Results All reviews The quantitative survey items revealed that education students’ were impressions varied between semesters; however, the overall anonymous impressions were positive (see table below). Across both semesters, students found being reviewed and reviewing to be 4.2 Participants beneficial and reported comfortability in both roles. During the fall semester, the lowest scores were reported when students were In the fall 2015 semester and spring 2016 semester, students in an asked whether they would like to see a similar peer review process education course agreed to participate in the research (see table implemented by more instructors. In variation, the lowest scores below). The students were asked to complete an anonymous in the spring semester were reported when students were asked online survey after engaging in each round of peer review. whether they received new insight into their work. Highest scores Students completed multiple rounds of peer review, and thus were reported in the fall semester when students were asked if completed the survey multiple times during the semester. they intended to change, or had already changed, their work based Accordingly, the numbers reported below indicate survey on the peer review process. The spring semester’s students responses rather than numbers of students. reported the highest scores when asked about the peer review Figure 2: Demographics/Logistics of Participants system’s ease of use. The overall rating of the peer review experience improved between the two semesters. Fall 2015 Spring 2016 Gender Gender Figure 3: Mean Scores from the Quantitative Survey Items o Female 88% o Female 82% Fall Sprin (N=142) (N=137) Formative Assessment Survey Questions 2015 g o Male 11% o Male 18% (N=17) (N=30) 2016 o Prefer not to answer <1% (N=1) Student Status N=160 N=16 Student Status o Full-time 92% 7 o Full-time 88% (N=153) (N=141) o Part-time 8% o Part-time 12% (N=14) 1. The reviews I received addressed the 3.90 3.77 (N=19) Age questions/ concerns I had about my work. Age o 18-22 56% o 18-22 66% (N=93) 2. The reviews I received gave me new 3.86 3.76 (N=104) o 23 or older 44% insight into my work. o 23 or older 34% (N=74) (N=56) Delivery Modality Delivery Modality o Face to Face 3. The reviews I received helped me 3.97 3.81 o Face to Face 73% (N=122) understand what I needed to change about 70% (N=110) o Online 27% my work. o Online 30% (N=45) (N=50) Round of Peer Review 4. I trust the feedback I received. 3.87 3.86 Round of Peer Review o Round 1 54% between the semesters support current literature that students * 5. I plan to change (or already changed) 4.11 4.20 benefit more meaningfully while acting as the reviewer. The my work based on the review process. increase in Question 11 indicates that spring students were more satisfied with the peer review experience than their fall * 6. I felt comfortable giving feedback to 3.86 4.11 counterparts. The decrease in Questions 1, 2, 3, 4, and 9 indicate my peers. that students did not receive the type of helpful feedback they were anticipating. These results are further enforced by the open- * 7. I felt comfortable receiving feedback 3.97 4.10 ended responses. Similarly, the increase in Questions 5, 6, 7, 8, from my peers. 10, and 11 indicate that spring students were more comfortable with the peer review process than fall students. * 8. The peer review system was easy to 3.82 4.29 use. 4.4 Summary of Qualitative Data The survey included five open-ended questions. Students were asked to consider both what they liked and what they found 9. The reviews I received were beneficial 3.90 3.80 challenging about the peer review process, what kind of feedback to me. they valued most, and what suggestions they had for peer review. They also had an opportunity to add additional information not * 10. The process of reviewing other 3.89 4.17 specifically requested. In both the fall and spring semesters, students’ work was beneficial to me. education students saw peer review in a positive light. One fall student observed, “When I am reviewing the work of someone else, I find my self [sic] noticing things that I need to work on in * 11. I wish more of my instructors would 3.34 3.81 my own work; I end up with a whole sheet of paper of revisions I use this type of peer review in their classes. need to make on my own work.” Similarly, a spring semester student stated, “I was able to understand how to review as well as when I reviewed what I should improve.” Students valued the peer review feedback as a form of copy Overall Evaluation of Peer Review Process editing much more in the fall semester than those students in the (Composite of 1-11) 42.5 43.7 spring. While students from both groups mentioned the benefits of input on formatting, grammar, and sentence structures, spring students appreciated, and were looking for, more substantive changes. One spring student stated in response to the item that asked what kind of feedback was most beneficial: “The feedback * = Increase occurred about what was good, the feedback about what I could change, > All means are from a 1-5 scale AND [sic] advice on how to do so.” Another stated, “One reader stated what she learned from my paper and I think having someone reflect what they see in your lesson is about is helpful, The greatest difference between the semesters occurred when making sure what your readers are getting and what you wanted to students were asked if they would like to see peer review of this communicate are lining up.” When giving feedback, however, type implemented in other courses (3.34 vs. 3.81). In contrast, the students were uncomfortable giving critical feedback and largest decrease occurred when students were asked if the reviews questioned their ability to give good feedback because they felt they received helped them identify necessary changes (3.97 vs. they lacked expertise. Assigning grades during the summative 3.81). Overall, the increased positivity between results in the fall round of review (round 3) was felt to be especially difficult for and spring semester was greater than the rate of decrease, and the this reason. Students felt it was hard to think of what to say when comprehensive results increased from the fall to the spring (42.5 they saw the work as being quality work or to not repeat what vs. 43.7). others had said when such feedback was visible to them. The consensus impression was positive, with every question Technology concerns were less prevalent in the spring responses. indicating more strongly agree/agree responses across both Where fall students struggled while experimenting with multiple semesters. Fall students responded most positively to the items technology platforms (i.e. Google Docs, Expertiza), spring related to recognizing and making changes in their artifacts (i.e. students found the technology to be an asset. The instructor “The reviews I received helped me understand what I needed to streamlined the spring peer review assignment to exist entirely in change…” and “I plan to change my work based on the review Expertiza. Students responded very positively to Expertiza, with process”). In contrast, spring students rated the utility and one student reporting: “I really enjoy using Expertiza for this type logistics of the peer review process most positively (i.e. “The of assignment.” Where technology issues were mentioned in process of reviewing other students’ work was beneficial to me,” more than 60% of open-ended responses from the fall responses, and “The peer review system was easy to use”). While the technology was only mentioned 6 times in the (4%) 167 responses responses were predominately positive, a small number of made in the spring data. Instead, the predominant theme of the students responded very negatively to certain items. Fall students spring data found students to be critical of their peers’ level of responded most negatively to the idea of using peer review in investment in the peer review process. Students were other classes. In the spring, students responded most negatively disillusioned about mismatched feedback, where “the chosen to the helpfulness of the reviews they received (i.e. “The reviews I [ratings] did not always match up with the comments.” Valid, received gave me new insight into my work”). The differences reliable, constructive, and thorough feedback emerged as students’ greatest desires, and subsequent disappointments. an average of 4.11 to 4.2 (see Figure 3). Students across the study Students were neutral/positive about receiving summative scores questioned peer review’s credibility and reliability, which reflects from one another but were dissatisfied/negative about reviewing previous scholarship [4]. While in the survey information does unexplained or mismatched feedback that accompanied that grade. present trends towards overall improvement in students’ One comment read, “I didn’t have any problem with being graded perceptions of peer review, it should be acknowledged that not by other students. I just didn’t like how I never received feedback every question showed improvement. For instance, questions 5 on why they gave me the grade [they] did.” Another student and 9 show contradictory changes. While question 5 showed that recounted, “I received a good grade but it was not a perfect score more students in the spring 2016 semester made changes to their and I wish I knew what was lacking…I heard a couple of my work based on the peer review process, question 9 showed that peers stating the same desire.” fewer students in the spring 2016 found the reviews they received during peer review to be beneficial. The instructor attributed this Positive responses from students discussed how students enjoyed difference to students’ participation in the calibration training seeing others’ work as this helped “clarify” the assignment. prior to the first round of peer review. Having practiced giving Students from both semesters valued constructive criticism more critical feedback and having seen expert feedback from the than complimentary “vague” commentary, as well as differing instructor, spring 2016 students were more critical of their peers’ perspectives on their work. Students from the spring semester felt feedback and therefore found fewer reviews to be beneficial when a deeper practical connection to the peer review process after it compared with fall 2015 students. However, due to this training, was compared to grading. A practice, or training exercise, was students found the peer review process to be more helpful in the added to the spring semester. These students completed a spring 2016 semester as they were more engaged in the peer calibration assignment in Expertiza that allowed them to assess review process. As supported by current literature, students learn two artifacts against an expert assessment. Students were asked to more from reviewing their peers than being reviewed. Further consider what was effective and ineffective in two example research and analysis should investigate how to foster reliability lessons plans. These lesson plans were created by students from a in reviews, or more precisely, how to help students trust the past semester, and each were representative of noteworthy feedback they receive from their reviewers. More significant positive and negative aspects. Students compared their rankings stakes should be placed on open-ended responses as students so to an “expert” review completed by the instructor. By evaluating highly value commentary from their reviewers. these two lessons, spring students had an advantage over fall As this snapshot is part of a larger study, the data from students; they were provided with a model to guide their own this course will be combined with the arcs from other instructors’ submissions and peer review responses. As pre-service teachers, courses in order to provide a more thorough understanding of how spring students were instructed during the assignment the structure of peer review can effectively promote student introduction that peer review is a dry run for future students. investment and learning. This research is ongoing and part of a While this point was mentioned in the fall semester, peer review larger investigation into student perceptions of peer review. The was presented as a more practical skill for these teachers during effects of these peer review processes on course instructors will the peer review training process. also be discussed. Student perceptions of peer review have shown While the data responses from both semesters were mostly a positive trend as the data from this study is further probed; positive, the negative commentary evolved from the fall to the students are feeling more and more positively about using peer spring to show a progressive direction for peer review in this review in their classes. More importantly, students are becoming instructor’s classroom. Far fewer students were disillusioned with more critical of using the peer review process in order to peer review during the spring semester. Negative-toned maximize their outcomes of making changes to their writing. commentary was centered almost entirely on students’ 6. ACKNOLWEDGEMENTS dissatisfaction with feedback (or, a lack thereof). Students wanted The Peerlogic project is funded by the National Science “slacker” peers to be held accountable for their failure to provide Foundation under grants 1432347, 1431856, 1432580, 1432690, substantive and “constructive” feedback. Of the twenty open- and 1431975. ended responses coded as “negative” (indicating unhappiness, dissatisfaction, et cetera), 16 mentioned a dissatisfaction with 7. REFERENCES incomplete, mismatched, or unreliable feedback. A student [1] Comer, D. K., Clark, C. R., & Canelas, D. A. (2014). Writing responded by stating, “I think that for the reviews, students should to Learn and Learning to Write across the Disciplines: Peer-to- be graded on completing the comment section. It was really Peer Writing in Introductory-Level MOOCs. International Review frustrating receiving grades below a [perfect] score and not have Of Research In Open And Distance Learning, 15(5), 26-82. [sic] an explanation as to why. It made me feel as though my peers were not actually taking time to assess my work.” [2] Liu, J., & Sadler, R. W. (2003). The effect and affect of peer review in electronic versus traditional modes on L2 writing. 5. CONCLUSIONS Journal Of English For Academic Purposes, 2193-227. While there are changes in students’ impressions of peer review [3] Lee, Chun-Yi. (2015). The Effects of Online Peer between the fall and spring semesters, this study’s hope to find Assessment and Family Entrepreneurship Experience on Students’ students making changes to their writing after engaging in peer Business Planning Performance. Turkish Online Journal Of review consistently occurs between both semester. Students Educational Technology, 14(1), 123-132. consistently rated that they planned to make changes to their work [4] Huahui Zhao1, Z., Sullivan, K. H., & Mellenius, I. (2014). after engaging in the peer review process. The improvements Participation, interaction and social presence: An exploratory implemented by the course instructor, including training and study of collaboration in online peer review groups. British streamlined technology, positively affected the students in the Journal Of Educational Technology, 45(5), 807-819. spring semester, increasing their agreeableness with this goal from [5] Raes, A. T. (2015). Increasing anonymity in peer assessment [8] Mulder, R. A., Pearce, J. M., & Baik, C. (2014). Peer Review by using classroom response technology within face-to-face in Higher Education: Student Perceptions before and after higher education. Studies In Higher Education, 40(1), 178-193. Participation. Active Learning In Higher Education, 15(2), 157- [6] Lu, R., & Bol, L. (2007). A Comparison of Anonymous 171. Versus Identifiable e-Peer Review on College Student Writing [9] Lundstrom, K., & Baker, W. (2009). To give is better than to Performance and the Extent of Critical Feedback. Journal Of receive: The benefits of peer review to the reviewer's own writing. Interactive Online Learning, 6(2), 100-115. Journal of Second Language Writing, 18(1), 30-43. [7] Wang, Y., Ai, W., Liang, Y., & Liu, Y. (2015). Toward Motivating Participants to Assess Peers’ Work More Fairly: Taking Programing Language Learning as an Example. Journal Of Educational Computing Research, 52(2), 180-198.