Yet more tips on moving teaching online

Chad and Amanda have already given their ideas on the crash move to online, so here’s my version too, in graphic form.

You can download a PDF version here.

As the others have said, you need to triage your teaching: what absolutely must happen and what’s just nice to have?

If you’re struggling, then ask colleagues, both within your institution and beyond: there are lots of great people on Twitter (start with the @ALPSblog follow list).

When Students Ignore Feedback

While I don’t comment on student writing nearly as much as some professors do, I expect students to at least read what I do write. A colleague recently pointed out that our Canvas LMS displays a date stamp at the top right of the Speedgrader web page when a student has viewed a previously-submitted assignment after an instructor has commented on it. I had never noticed this before, I guess because the date stamp’s font is rather small. Here it is, indicated by the red arrow:

This feature became very useful in a course in which students are required to write a series of memos that all have the same format. Last week, a student taking the course sent me this email:

I’m not sure what is expected from the assignments, my memo 3 was completely different from 2 yet your comment says see comments about memo 2. I am a second semester senior doing grad classes that had a 3.6 gpa last semester. Somehow I’m failing every single assignment in a freshman level class, while still attending every single class except one and participating in class basically every session. 

I looked at the student’s submissions for memos 1, 2, and 3 — no date stamp. My comments had been ignored. My reply to the student’s email:

The memo is a standard method of efficiently communicating information that is used in a wide variety of professional environments. I’m surprised you haven’t yet had much of an opportunity to practice this form of writing, so here is what I am willing to do: you can earn up to 10 points by emailing me by noon on Friday a memo that discusses how well you incorporated my feedback on your Memo 1, provided by my comments on your work on Canvas, into your Memo 2, and the same for Memo 3 in respect to my comments on your Memo 2.

Completion of my “extra credit opportunity” would have required the student to admit that he had not read my comments and thus ignored the feedback I had provided.

The student did not respond.

Design Fail or Attention Fail?

I recently graded a writing assignment for one of my courses and I’m wondering if it’s an example of “You can lead students to the education but you can’t make them learn.”

The instructions for the assignment:

You have been given the task of submitting a memo to the National Security Council  that answers the following question:

  • Will Nigeria remain a single state or divide into separate states?

Download the memo template; use it to format your work. Turn the question above into a declarative sentence and use it as the memo’s executive summary.  Write two subsections that support your argument using evidence taken from course readings. Each subsection should be a single paragraph. Reference sources with in-text parenthetical citations. 

The information above was repeated by the memo template itself and by the rubric attached to the assignment. From my perspective, the assignment is totally straightforward and the assessment criteria are completely transparent. Yet . . .

Several students wrote memos on the question of “Should?” rather than the actual question of “Will?

Many students also failed the “Turn the question above into a declarative sentence . . .” part. A few representative examples of what they submitted as an executive summary:

  • “In the current course, Nigeria will see a deterioration in multiple states because of the lack of agreement over these issues and the failure of the government to uphold a true democracy. Nigeria is a fledgling state on the verge of an inner collapse due to current trends.”
  • “The United States should help Nigeria dissolve into multiple sovereign states, by mediating the separation process. Nigeria is currently facing a slew of ethnic and social conflicts the country is made up of 36 states which legally sew divisions between regional outsiders and regional natives, this has sparked ethnic and religious conflicts.”
  • “The best path forward for Nigeria is to remain a single state. Splitting the nation up now would only be detrimental to Nigeria’s ability to control its sphere of influence across the African continent. Splitting Nigeria into multiple states would challenge the work that has gone on for years to make it politically equitable and would not account for the vast cultural differences of the nation.”

And so on.

I’m wondering how I should interpret what happened. Is this simply a case of failing to follow directions? If not, I don’t know how I can make things more obvious.

Modeling “Good” Writing

Today we have a guest post from Colin Brown, assistant teaching professor in the Department of Political Science at Northeastern University. He can be reached at colin [dot] brown (at) northeastern [dot] edu.

A few weeks ago, Shana Gadarian made a point on Twitter about explicitly teaching writing that I strongly endorsed. Getting our students to write better will, at minimum, make our job easier. If we’re lucky, it might actually promote better thinking.

For upper-level students, very open-ended prompts sometimes lead to really creative and insightful thinking—and a dissertation is of course the vaguest prompt of all. But our expectations often rest on implicit assumptions about what we think of as “good” political science. Amanda Rosen has written about this in the context of transparency. As she points out, telling students “write a five-page essay” doesn’t acknowledge that essay means different things in different fields, and many of our students will not yet know what this even means in political science.

Clarity is critical for essay prompts, especially for introductory students. While long, detailed instructions might help point students toward what they should think about, students new to a field often don’t have the context to know what is most important in a long prompt. To them, any sentence with a question mark might appear to be equally important—causing them to focus on what we thought was a minor point and producing a disconnect between what we want to assess and what we actually assess when our implicit expectations aren’t met.

Here are what I think were a failed and a relatively successful attempt to do this in my past semester’s intro comparative politics class. Students told me that while the instructions were explicit in the first, it was hard to know where to start and which parts to emphasize. With the latter prompt, they said it was not only clear what to do but why they were doing it.

One question I’ve found to be a bit polarizing in talking with colleagues is whether to provide model papers or templates. Is it a better way to make our aims clear, or does it cause students to just parrot back the template? I’ve always found myself on the side of providing models. Graff and Birkenstein’s They Say, I Say has solidified my thinking for why. They identify common rhetorical devices that mirror the most common academic ways of thinking, which they into templates that, for example, force students to write down the most obvious counterarguments. Experienced writers have read enough in the field to be able to pick up on these techniques implicitly, but beginners by definition have not. Graff and Birkenstein argue, and I think rightly, that this goes beyond rhetoric to actually learning the ways of thinking. Some students may not learn how to think about the flaws in their own argument, or even that they need to, until they are forced to write them with this kind of template.

In my own teaching, I’ve found it hard to explain in an abstract sense the need for clear writing over “beautiful” writing—and what students think is “beautiful” writing often feels cluttered and verbose to us. But when students see each other’s models and observe how much they actually understand from clear writing versus what they think is “good” writing, they start to self-diagnose their own excessive prose.

One challenge is that writing for politics requires some discipline-specific skills that might be more or less amenable to templates. Sarah James, George Soroka and I have a forthcoming JPSE piece on adapting tools from K-12 and composition studies for political science writing. But defining what we actually want from “good” political science writing seems often more folk wisdom than clearly defined—Ian Anson’s 2017 article on meaning-making is a notable and valuable exception—so as part of showing students what we want, there’s room for us to make this clearer to ourselves.

 Sarah, George and I will be leading a workshop at APSA TLC 2020 on implementing rubrics to improve student writing—Friday Feb. 7 @ 4:15 pm.

Writing as Learning

My last post discussed writing as a professional endeavor. Today: writing as a device for learning; i.e., why and how we as teachers assign writing to students.

Generally we present our students with some form of expository writing task. Perhaps we call it thesis-driven, discipline-oriented, argumentative, or research-based. Regardless of the label, there is an assumption of students locating relevant primary data by means of different methods that they understand how to use, evaluating the data in an appropriate manner while being aware of their own assumptions, reaching some conclusion, and effectively communicating all of this to an audience.

That’s the ideal. The reality? Students often don’t know how to find primary data, or which methods are best suited for analyzing it. They may not even know what methods are. They assume there is either one right answer, or that all possible answers are equal, because they don’t understand that some answers can be more strongly supported by data than others while even better answers await discovery in the future.

And so we default to assignments that direct students to preferred secondary or tertiary sources (a “text”), tell them to organize their explanations as competitions between two artificial, diametrically-opposed positions, or, sometimes, encourage them to dredge up arguments that arrive at positions they already favor. Students learn to hang evidence on a predetermined conclusion rather than derive a conclusion from the evidence.

This type of deductive exercise has been used by teachers since the age of the agora to build students’ rhetorical skills. Today, unfortunately, it can produce people with a facile ability to argue any position at any time without veering from a worldview that they hold to be sacrosanct.

So what’s the solution? I don’t really have one. Too few of the students I encounter are willing or able to draw reasonable conclusions from evidence they have independently located, so writing exercises that involve inductive reasoning get chucked out the window. It’s frustrating.

Why does no one come and talk to me anymore?

This guest post comes from Patrick Bijsmans of Maastricht University.

I’ve always considered myself an approachable teacher; someone students can come to with questions or worries or just for a talk. And from what I hear, I am considered to be approachable.

Still, I am noticing something that worries me. I have been having open office for about 9 years now, but fewer students have been showing up. Weeks go by when no one comes, even in periods when I am teaching and coordinating courses.

I know that I am not the first one raising this issue. It is even the topic of students’ research! But I still believe that students can learn from meeting with us for input and feedback, whether this concerns a relatively simple question or my assessment of their paper.

So, why does no one come and talk to me anymore?

Turnout during open office hours again was low during the first weeks of this year, when I coordinated and taught a first-year course on academic research and writing. At the end, students write a short paper. These are randomly distributed among teaching staff, myself plus 10 other colleagues – together we teach 25 problem-based learning groups of about 12 students. As soon as results are out, all students, whether they have failed or passed, are invited to meet with the person who marked their paper to discuss the assessment during scheduled open office hours.

This year I asked colleagues to inform me about the number of students that had shown up. The table below shows the data for those who failed the course. Interestingly one colleague had to do her open office hours via Skype; no less than 7 out of 9 students showed up. Yet, there is some research that suggests that using technology does not make a huge difference.

Number of failed students Number of failed students attending open office hours
Patrick95
Colleague 1 72
Colleague 296
Colleague 394
Colleague 4124
Colleague 594
Colleague 61711
Colleague 7159
Colleague 8174
Colleague 9156
Colleague 1097
Total12862

Why did so few students show up?

I decided to ask some simple questions to the students themselves during a session in our mentor programme. The approximately 100 students who attended (out of nearly 300) might not be representative of the group of students that does not turn up in my office. But I still learned something interesting.

Of the 86 students completing questions via an online survey tool, 36 had failed the course and 29 had attended the open office hours. Those who attended, generally did so to get clarification regarding their paper’s assessment.

Of those who did not attend, some simply stated that they passed the course and saw no need to discuss the feedback. Others referred to having been sick, stressed and/or busy with the new courses – when asked, quite a few of these students did not write to staff to ask for another appointment.

Asked why they thought others had not come, some answered that these must be lazy students or that they missed motivation because they knew what they had done wrong.

But quite a few answers touched upon something that we might all too easily overlook, namely students’ expectations regarding feedback opportunities. These answers did not just concern not knowing what to do with feedback. For instance, one student wrote that students who did not show up might be “insecure and/or uncomfortable with getting feedback”. Another student wrote that “you have limited time with the tutors and tutors often have a lot of work and not much time for you”.

Could it be that low attendance during open office hours is due to barriers to students’ engagement with feedback or, more generally, a lack of feedback literacy?

This is something that I want to explore in more detail. I have already briefly discussed this with our academic writing advisor, and we may want to see whether we can specifically address this issue in a forthcoming curriculum review.

But what about solutions for the here and now? There are many ways in which open office are organised, but what works best?

One colleague suggested changing times. Admittedly, my open office hours are Wednesdays from 08:30-09:30, but this never was a problem – and the feedback open office hours during the aforementioned course were scheduled in the afternoon. Elsewhere in cyberspace people have been suggesting other solutions, including a rethink of faculty office space. I’d love to squeeze in a couch, but my office is rather tiny.

On Twitter someone suggested that the wording ‘open office hours’ is unclear to students and that ‘student drop-in hours’ may make more sense. So, the name plate next to my door now mentions my student drop-in hours and so does the syllabus of an upcoming course.

Let’s see what happens. I hope students will come and talk to me again. The door’s open, simply turn up at the stated time!

The Article Summary

(Photo credit: Joanne H. Lee, Santa Clara University)

Today we have a guest post about teaching the research process by Anne Baker, assistant professor of political science at Santa Clara University. She can be reached at aebaker [at] scu [dot] edu.

Getting students to use academic articles for research papers can be a challenge. In my experience, many students, even those in upper-level courses, are not familiar with search engines such as JSTOR, Lexus Nexus, or Political Science Complete. And if students do happen to use Google Scholar, they frequently rely on excerpts from sources instead of entire articles that they might not have access to. So, what can be done to replace these habits with better practices?

In my advanced writing course on the presidency, I have developed a class activity which provides students with skills they will need if they are going to successfully locate and utilize academic references for their research papers. First, I want them to be able to use the library’s website to access search engines. Second, I want them to understand that research is an iterative process. Sometimes you don’t find what you need for a variety of reasons and you should be able to determine what those reasons are—whether its human error, the need for a wider search net, or that no one has written on the topic (this last possibility always surprises the Google generation). Third, students need to become acquainted with the literature on the presidency, including the subfield’s primary journal, by discovering how research practices in political science have changed overtime, even in a subfield which remains largely qualitative.

I have students work in pairs and I provide them with two search terms related to the institution of the presidency (e.g. signing statements, executive orders, oath of office). I pick the search terms carefully knowing that some topics have no scholarship and represent dead ends and others have later but not earlier scholarship or vice versa. The first step of the activity provides instructions about how to first locate JSTOR on the library’s website and then how to access Presidential Studies Quarterly using JSTOR’s advanced search options. Helpfully, for the purposes of this activity, JSTOR only has copies of the journal until 2000. To access later copies, students have to use the Wiley database, which students have to figure out how to find.

For each search term, I have students locate one article published in the last few years and then another for 1995-2000—a total of four articles. Next, students identify the research question and method the authors used, noting whether it is qualitative or quantitative, the sources of data regardless of method, the type of analysis (e.g. text, interviews, statistical), and the date of publication. After they have their four articles and perform this analysis, I ask them to compare the results of both searches. Finally, we have a class discussion in which we explore road blocks and challenges encountered and review how the field has changed over time.

I have found that this activity makes students more likely to cite academic articles in their final research papers and use them more effectively to support their arguments. Students also exhibit a much better understanding of the subfield and are more likely to use the other search engines that they encountered while on the library’s website. And they learn that research takes time and requires shifting your strategies to find the information you need.

Storytelling with GIS

Another example of why it’s good to consult with librarians:

For the last several years in my globalization course, I’ve had student teams create and deliver presentations on their commodity chain analyses and ethnographies of consumption. Generally students build Powerpoint files for these assignments; occasionally someone uses Prezi. Simple rubrics make grading this work very easy. But the end products aren’t going to make recent graduates stand out from the competition when interviewing with prospective employers. It’s also difficult to convey the content of the entire project in a single presentation without showing a mind-numbing number of slides. Enter the storymap . . .

One of our librarians, a specialist in digital scholarship whom I’ll be working with next semester, introduced me to the digital storytelling tool from Esri,* a.k.a. the Environmental Systems Research Institute, which allows a person to create a multi-media presentation with ArcGIS. Rather than describe what this looks like, I’ll show you:

My task now is to reconfigure the project’s different assignments so that students complete their own storymaps by the end of the course, and figure out how to evaluate them.

*I have no financial interest in this company.

Does the Question Determine the Answer?

Regular readers of this blog know that I sometimes ponder the clarity of my assignment and exam prompts (some past posts on this subject are here, here, and here). Students sometimes don’t hit what, in my mind, the question targets, so I revise in the hopes of creating a prompt that is more transparent. But I don’t want prompts to be answerable with a Jeopardy-like regurgitation of facts. I want students to exert some cognitive effort to figure out how to apply concepts that are relevant to the question at hand.

Usually this situation occurs with my undergraduates, but I’m noticing it more frequently with master’s degree students. A recent example is an assignment from my graduate-level introduction to comparative politics course:

Continue reading

Pay It Forward Assignments

Today we have a guest post from Loleen Berdahl, Professor and Head of Political Studies, University of Saskatchewan. She can be reached at Loleen [dot] berdahl [at] usask [dot] ca.

I am very excited to road test a new ‘Pay It Forward Assignment’ (PIFA) in my classes this fall. The PIFA requires students to create something that will help a student in the same class in subsequent years. It is inspired by David Wiley’s argument in favour of non-disposable assignments. My hope is that this assignment will inspire students to approach the material with creativity, and draw upon their own interests and personal strengths while they engage with class material. Ideally, the assignment will create a library of resources for future classes that can grow over time.

Here are the assignment’s instructions for students:

Have you ever mastered a topic and wanted to share what you have learned with other students? Have you ever wanted to demonstrate your mastery of a topic in a creative manner, rather than the usual format? Have you ever spent a tremendous amount of time on a class assignment and been disappointed to know that the only person who would read it was your professor? If so, the Pay It Forward Assignment (PIFA) is going to be fun and interesting activity for you.

The PIFA is an assignment that a student creates to share with future students. A great way to select the topic for your PIFA is to find something that you initially found difficult in the course. Once you have a good grasp of the topic, create a resource of some sort to help others. Some possibilities for your PIFA:

  • Graphic novella: create a graphic short story that explains a key idea.
  • Mind-map: construct a mind-map for a specific topic or a core idea.
  • Power-point slide deck: create a short series of lecture slides that teach a major point of a lesson.
  • News story video: present a key idea from a lesson as if it were a breaking story for a television news program.
  • Audio: create a short podcast that provides a tutorial on a key idea in a lesson, or that connects a key idea to a current issue in public debate.
  • Photo essay: use a photo or series of photos to illustrate a key idea in a lesson, and then use short amounts of text to explain how the photos illustrate the idea.
  • Checklist: create a critical reading checklist for students to use when assessing a research study, news story, etc., for trustworthiness. Go through an actual article and apply the checklist to demonstrate how it works.
  • Learning activity: outline a new learning activity for a lesson.
  • Lesson table summary: construct a summary table of the lesson’s key points and terms.
  • Linking our class material to another class: using text, audio, or video, explain how something from this class relates to something you learned in another class.
  • Relating lessons summary: using text, audio, or video, explain how the material in a particular lesson relates to the lesson that came immediately before it.
  • Social media activity: create a tweet-thread explaining a core concept covered in a lesson.