How Comprehensible Must Reading Be?

Marcos Benevides’ Slideshare PPT has been floating around for over a year now. It’s a powerful illustration of how unknown words affect reading fluency (speed + accuracy), especially for anyone who thinks students will be OK reading anything that’s less than 98% comprehensible.

Still, the syntactical clues in Marcos’ PPT helped native speakers. In order to simulate a student’s reading experience more accurately, I removed those clues. Here’s the result (download, here, for sharing):

Slide1Slide2Slide3Slide4Slide5

Negligent Motorist = 1, Lance = 0

Yesterday, the following events unfolded while riding my motorcycle:

  1. I notice a car rolling towards towards the road at a TD Bank exit driveway—the driver isn’t looking left (i.e. my direction).
  2. The driver doesn’t look my way, keeps rolling, then suddenly turns left into the road directly in front of me.
  3. I stop short. The momentum sends my motorcycle down on its right side, and me forward, also on my right side.
  4. I’m on the ground now and can’t move, but it’s for an OK reason—I realize that my helmet is stuck between the pavement and bumper of the driver’s car.
  5. The driver gets out of the car and tries to move me (idiot!).
  6. I take some time to watch horrified rubberneckers looking downward at a motorcycle on the ground and its rider partly under a car.
  7. After the disorientation dissipates, I get bored not doing anything under the car, extract myself, take off my gear, and take in the situation.
  8. Motorcycle doesn’t start (it won’t shift BELOW 3rd gear—the one it was in before going down).
  9. I wrap things up with the officer, get the moto towed, start calling insurance companies, and text Bob Patrick. No, Bob is not my emergency contact, but he just happened to have caught a typo in Discipulus Illustris, which led to a nice suggestion (i.e. Quō in annō es? for Quō in gradū es?)

Continue reading

“Classically Attested” & “Latinity”: The Latest Buzz

**Updated 5.19.18** I forgot about this post until a comment came through just now. Was I really not using http://latin.packhum.org in 2016?!?! Sure enough, Perseus is utterly unhelpful, still showing no hits even when “different forms” is checked or if they are there, it’s not in Latin, and it’s buried deep within the pages. Packhum, however, has 60 hits, instantly. 60?!?! There I was this fine night in 2016 thinking I had to defend myself for using “magis placet,” yet look at all those beautiful instances right there! Anyway, the particular phrase, then, is a proxy for any phrase we actually DON’T have. I’m not editing this post.

Teacher-written reading material for Novice & Intermediate language learners is not new, at least for modern languages. TPRS Publishing and TPRS Books frequently add more to their roster, so teachers have their choice of topic. 2015 saw the publication of the first two of these novellas in Latin, [self-]published by Pomegranate Beginnings (i.e. Pluto & Itinera Petri). Since then, there has been a wide range of reception amongst Latin teachers (or Classicists, or Linguists, or Scholars, etc.). The general consensus regarding the positive reception has been something like  “this is helping our students feel successful and have a positive experience in Latin class. If you don’t like it, you don’t have to use it.” Regarding the negative reception, most of it revolves around two buzzwords:

Continue reading

Forget about the Fossa: [Textbook] Embedded Readings Done BETTER

After attending Michele Whaley’s presentation on Embedded Readings at NTPRS, I was convinced that we’ve been playing a game of Telephone since she and Laurie Clarcq began sharing the concept back in 2012. It turns out that it’s “yes and no,” but there is an important distinction that is being made in current Embedded Reading practice. Whereas many of us THINK we’re creating Embedded Readings, most of us might be just adapting authentic texts, class stories, or textbook narratives. Those products are fine, but aren’t necessarily Embedded Readings. Most of us are missing two key features in our adapted readings that make them better:

  • Parallel Stories
  • Withholding New/Tantilizing Information (not just more words)

Continue reading

Grading vs. Reporting Scores: Clarification

In the recent sliding scale scheme, Proficiency is given 0% weight at the start of the year. This doesn’t mean that students see “0” in the gradebook. What this means is that their 95, for example (which they see in the gradebook), holds 0% weight because in the sliding scale scheme we’ve placed all 100% weight on DEA for first quarter in order to set expectations and establish routines. By the fourth quarter, 100% of the weight is on Proficiency, and whenever possible, we manually change the entire course grade to that final Proficiency number/letter so nothing else averages throughout the year.

NTPRS 2016: More Changes, More Thoughts

After attending iFLT, I spent another week in Reno at NTPRS. While iFLT offered more opportunities to observe teachers teaching students, NTPRS offered more opportunities to actually BE a student for those of us in the Experienced track. I appreciated the short demos that most presenters gave, even when the workshops were not titled “___ language demo.” There are some game changes here that warrant their own posts  (e.g. embedded readings straight from the source, Michele, Whaley), but I have much  else to report on. Like last week’s iFLT post, this one includes more of what I intend to think about and/or change for 2016-17. They’re organized by presenter:

Continue reading

iFLT 2016: Immediate Instructional Changes, and Other Thoughts

I just went to my first iFLT conference. I got to chat (live) with Bill VanPatten and Stephen Krashen, saw master teachers teaching with CI, and went to some awesome presentations. I don’t take detailed notes during presentations, but as you can see there’s plenty to take away from a few ideas I emailed to myself over the week. This post includes what I intend to think about and/or change for 2016-17, and would recommend others considering as well. Some of the ideas were ones I’ve seen before but just haven’t gotten around to implementing them, while others were completely new. They’re organized by who inspired me:

Continue reading

Why “do you understand?” is pointless to ask…

Language teachers usually ask this when something indicates that a student didn’t understand (e.g. verbal response “huh?” or non-verbal response deer-in-headlights expression on face, etc.). If this event has already happened, asking the question serves no purpose. In fact, it might even make the matter worse by putting the student on the spot. The student will likely answer “yes, I understand” just to get their teacher to move on to someone else. Here are some comprehension check alternatives:

1) Did I just say/ask X?
2) Did I just say/ask X or Y?
3) I just said/asked ____.
4) What did I just say/ask? -or- Who can tell me what I just said/asked?

The alternatives above are arranged by questioning level from low to high (i.e. yes/no, either/or, fill-in-blank, open-ended). The questions could certainly be asked in the target language, but one popular strategy is to ask, in English/native, “what did I just say/ask?” to a so-called barometer student, who would be one with the slowest processing speed. This popular strategy is interesting because that kind of question is technically harder to answer than “did I just say/ask X?” It’s probably a non-issue because we’re dealing with the native language, but for the sake of variety, or if you find that your barometer students are struggling, you could start asking those lower level comprehension checks in English/native as well the classic “what did I just say/ask?”

Grading Scheme: DEA & Proficiency

**See this post for all other grading schemes*

Here’s a new idea inspired by advice I was giving on various DEA and Proficiency grading weights. In other posts, I’ve written how my DEA weight has been anywhere from 0% to 50% of the grade. You could also try this sliding scale throughout the year…

Quarter 1
DEA = 100%
Proficiency = 0%

Quarter 2
DEA = 50%
Proficiency = 50%

Quarter 3
DEA = 10%
Proficiency = 90%

Quarter 4
DEA = 0%
Proficiency = 100%

A grading scheme like this would establish very clear expectations of how important it is to exhibit behaviors and routines that lead to language acquisition in class (e.g. Look, Listen, Ask). This would work best if you have the admin support to manually override the final grade with just one Proficiency grade from Quarter 4, as suggested in other iterations of my grading systems. Why? We don’t reaaaaally want the 4 quarters to be averaged, but if they are it’s not the end of the world. This kind of grade is far more forgiving so the focus can be on input and not assessments.
N.B. Proficiency is given 0% weight at the start of the year. This doesn’t mean that students see “0” in the gradebook. What this means is that their 95, which they see in the gradebook, holds 0% weight because in the sliding scale scheme we’ve placed all 100% weight on DEA for first quarter in order to set expectations and establish routines.

Alphabet (Don’t Teach It!) Alternative

Months ago, I witnessed a classically ineffective language learning lesson. The good news is that the person in charge wasn’t actually a language teacher, and didn’t have pedagogical training at all. The person was a local substitute who gave the kids something to do, which has its own merits. The truth, however, is that many language teachers spend the first few classes teaching the alphabet. Don’t.

Continue reading