Category Archives: Pedagogy

Why Come to Class?

During a recent consulting session with an instructor about her upcoming blended-delivery course, our discussion turned toward which learning activities would best serve students when delivered online vs. those that would best serve students when delivered in the classroom.

It’s an interesting question: What unique learning attributes are contained in a room that—for a period of time—contains one instructor and a group of students?

When I ask faculty what they do in the classroom, the answer I most often receive first is that they “lecture” or “talk about the content” or “present information.” When pressed further faculty refer to other activities. “I watch students to see if they are paying attention,” or “I ask questions to see if they understand the material.” And then I begin to hear yet another group of activities that includes the following:

  • “We discuss the material.”
  • “Students meet in their project groups and I spend time with each group.”
  • “We do practice problems on the board.”
  • “We review the homework assignment.”

So, here are three things that happen in a face-to-face classroom: delivery of content, assessment of student engagement with the material, and guidance of student learning and performance.

Interestingly, when I ask faculty what they like most about teaching, rarely have I ever heard lecturing top the list. In fact, I’ve never heard that answer. What I usually hear are answers related to that last task. “I enjoy guiding students through the work of learning,” or “watching students get it,” or “seeing the ‘ah-ha’ moments.” I would guess that if we asked students what they enjoy most about learning, these moments would be high on their lists as well.

What if one could actually structure a course to include more of these teaching moments—these moments of guiding student learning?

Barbara Walvoord reported on this type of course in 2003. “I wanted to use my time not to deliver information, but to engage in discussion with students and respond to their writing, encouraging their development of sophisticated analytical skills and creativity.”

To accomplish this, she basically re-created her Shakespeare class as a hybrid course. Information and ideas were delivered via readings and by videos already owned by the library. The rather large course was broken up into several groups of approximately 18 students. Each group met with the instructor once a week for an hour of discussion. Students were held accountable for their weekly short assignments, notes on readings and the video presentations, and on their participation in discussion.

Every student spoke in class every week, wrote every week, and received personalized feedback from the instructor on drafts of essays. The drop rate for this section of the class was the same as that for other sections. All measures indicated a successful and rigorous learning experience for the students.

Walvoord’s framework may have been “faculty productivity”—a term that sets an educator’s teeth on edge, much like nails across a chalkboard. Yet, her process does result in that central joy of instruction. So, what is it that the instructor can uniquely bring to the classroom? Herself? Her feedback? Her ability to guide discussion?

Well, that is actually a large part of what happens when a course is re-designed for hybrid/blended delivery. Faculty time is focused on providing immediate feedback on performance and practice, engaging students in discussions around concepts and ideas, asking guiding questions about group or individual projects, and listening and responding to student presentations.

For more information, see the following resources:

Walvoord, B. (2003). New Modes of Productivity for Student Learning. New Directions for Higher Education, No. 121. 35-49.

Soft (Arts) vs. Hard (Sciences/Technology) Education: Imagination vs. Reason

Both the low marketability of arts degrees and the low salaries of arts educators in our society, when compared to the marketability of degrees and salaries of educators in science or technology topics, reflect an attitude towards the arts that sees them as accessories to our lives, good mainly for entertainment, pleasure, or escape. This attitude frequently undermines arts education funding and is, for some, due to the admitted difficulty non-artists and artists alike face when trying to assess success in arts education and production with measures that make sense to and can be appreciated by “non believers.”

Assessing arts education outcomes (Hanna, 2007)

To this end, Dr. Wendell Hanna (San Francisco State University) recently published a well-written and organized article on the applicability of the new Bloom’s taxonomy to arts education assessment [Hanna, W. (2007). “The New Bloom’s Taxonomy: Implications for Music Education.” Arts Education Policy Review, 108(4): 7-16.]. The first section of the article offers an insightful and concise outline of the significance of assessing music education outcomes and of the history and current state of Bloom’s taxonomy as an education-accomplishment assessment tool. It is followed by a meticulous and convincing (even if a little tedious at times) set of arguments for the way music education activities and national standards fit within the new Bloom’s taxonomy.

Hanna (2007) effectively accomplishes her principle goal, to show that:

Music education functions within and contributes to the same types of knowledge acquisition and cognitive processes, and its outcomes can be assessed using conceptually the same standards and tools as other educational areas that deal with topics traditionally more “respected,” “objective,” and widely accepted as beneficial to individual and social behavior and success.

Does “high assessment” translate to “high value?”

Whether the above conclusion can support claims for the need to keep music education in schools is not as clear to me as it seems to be to the author. Based on her concluding sections, Hanna seems more interested in promoting the usefulness of a new, uniform, and standardized assessment tool than she is in arguing for the general value of musical accomplishments. The goal of this assessment tool is to make communication of musical accomplishments among “music lovers” and between music lovers and non-music administrators easy, efficient, and consistent with concepts non-experts are familiar with. However, defending the value of music education in promoting the individual and social development of students is, in my opinion, a most pressing issue, as the way it is resolved will determine whether or not accomplishing the goal set in Hanna’s paper is of any consequence.

For example, even the process of systematically learning how to knit can be made to fit, to some degree or another, the knowledge acquisition and cognitive processes outlined by Bloom’s taxonomy. This offers us useful ways to assess what processes have been used and to what end and degree of success. Such an exercise, however, will not answer the question of whether the specific “end” in question is “valuable,” “respectable,” and useful to the individual and the society beyond the limited bounds of the activity itself. Algebra, biology, geometry, and all the other “respectable” educational subjects are not respectable simply because students end up learning how to solve equations or properly identify a frog’s internal organs. Rather, they are valued because of what one can contribute to society thanks to her advanced mathematical and scientific skills.

Precisely what these contributions may be is not made explicitly clear, but their value is implicitly accepted as being significant within our culture. On the other hand, what a student can contribute thanks to how well educated she is in music is even less clear and, largely, not accepted as valuable. It seems to me that, before one can appreciate how good a student has become in music and how consistently we can assess her accomplishments based on a standard tool, we must address the question, “Why should anyone become good in music?” The typical response: “for no good reason beyond entertainment and escape,” reveals an attitude that threatens to make efforts like Hanna’s ultimately inconsequential.

The cognitive significance of art & imagination vs. reason

In my opinion, the way to go is to systematically and convincingly argue for the cognitive significance of art in general and music in particular—a non-trivial task that is beyond the scope of the present post. To get things going, however, I would like to briefly assess the longstanding, conventional opposition between imagination and reason, which, I believe, is behind our difficulty to appreciate art’s cognitive significance.

Bear with me for one more paragraph, as I will be tracing an arguably problematic rational consequence of such opposition.

Common sense understands imagination as a mental activity that deals with things that are not really there. It is opposed to reason, which is consequently supposed to be dealing with things that are really there. At the same time, the observation that not all future events can be predicted based solely on past and present observations indicates that future things must include things that do not already belong to the past or present. If the future includes things that are not present (i.e. are not really there) or past (i.e. have never really been there) then reason, by definition, cannot address it. Such a limitation severely undermines the importance of reason to our lives, by stripping from it the power to, in any radical way, influence our outlook. The only way reason can address future things is by making believe that such things—things that are not really there—are present, so that it can subject them to determinate and reflective judgment. In other words, in order for future to be reasoned with it first has to be imagined. The conventional opposition between imagination and reason and the accompanying assumption of reason’s superiority leads, therefore, to a curious and paradoxical “reason” that is superior to imagination, but impotent without it.

Until convinced otherwise, I, for one, will keep imagining.

Avatar photo

What if Confucius Were to Teach Online?

A couple of weeks ago, I attended the Computers and Advanced Technology in Education conference in Beijing, a city where I went to college 20 years ago. To me, it felt a bit strange to hear the familiar subjects—new instructional tools and pedagogical methods—being discussed in an ancient city of China, which has become so modernized that I could barely recognize it. At the conference, the clear divide between the technical and pedagogical tracks reminded me of the disparity between modern technology and the historical cultural roots of China. Almost all of the presentations delivered by Chinese scholars focused on technology, and most of the technologically focused papers seemed to shy away from any deep analysis of the pedagogical impact of these technologies. This made me wonder, ‘Where were those insightful learning theories developed in China thousands years ahead of the west? Where was Confucius and his philosophy of teaching and learning?’

An interesting thought came to my mind as I was wandering through those technology demonstrations: what if Confucius were given the opportunity to teach online? Would he be as resistant to technology as many of our conservative faculty are? Or would he be willing to use the Internet to serve his mission of perfecting human beings through education?

As someone who taught through dialogue and observation, Confucius would very likely to be skeptical about teaching online. Yet, if he were to take advantage of online technology (as many faculty are doing now), I would speculate that his way of teaching would be dramatically different from most online instructors. Since I cannot easily anticipate what Confucius would do, I would just venture to list a few things that I believe would not happen if Confucius were to teach online:

  1. There would be no learning objectives.
  2. There would be no learning modules.
  3. There would be no “instructors” in the class, but “coaches” and “mentors”.
  4. There would be no tests and exams.
  5. There wouldn’t be any “course” or “courses.”
  6. There wouldn’t be a green light allowing every student to study online.

Now, here are my rationales for the would-not-happens.

1. No Learning Objectives
Confucius believed that through “studying,” one proceeded to reach the stage of human excellence. In this regard, the goal of education was to cultivate and facilitate self-improvement. Such improvement, in his opinion, was personal and could only happen through individual engagement in learning. Since humans differ in nature, Confucius suggested that instructions ought to be tailored for each individual student. The Analects of Confucius recorded that once a question was asked about whether the student should immediately put into practice something he was taught. To one student who Confucius thought was particularly zealous, he recommended that the student first consult his father and older brothers. To the other student who Confucius thought lacked enthusiasm, he said yes, put it into practice right away (The Analects of Confucius).

This story makes Confucius the earliest practitioner of constructivism. One could easily infer from his thoughts and practices that he viewed the paths to achieve the ultimate learning goals as being different from one student to another. In this sense, he would oppose putting any milestones on the journey of learning because people were traveling at different routes and even their own routes might change during the process. In developing online learning, we label those milestones as “learning objectives,” which were set by the instructors and given to students as common targets that move them along “a well-trod and clearly marked road.” The main challenge of learning, in this perspective, is to keep students moving down the road on schedule. (Ron Weigel, 2005).

These milestones or learning objectives would certainly be removed by Confucius.

2. No Learning Modules
Break the content into bite-sized chunks to make it easy for students to digest! Carefully label each chunk with instructions and dates for it to be consumed! Place them in a good sequence so that they are completed by the students in the right order!

None of these design tips would have found their place if Confucius were to teach online.
Confucius, even in his own time, never used structured class to teach. Because of this, one can hardly imagine him using modules to break knowledge into discrete pieces.

Confucius saw learning as a process of observation, followed by reflection, and internalization or implementation. His thought infers a clear distinction between information acquisition and the formation of knowledge and wisdom. He would leave it to the students to observe and to acquire information without interfering and would never limit what students could “observe” by storing things in a fixed module.

3. No Instructors but “Coaches” and “Mentors”
Confucius’ methods of teaching were striking. He didn’t discourse at length (no lectures either!). Instead he posed questions, cited passages from classic texts, used apt analogies, or simply just listened and waited for the students to arrive at the right answer. He would not be there to tell students when to finish what assignment or what’s right and what’s wrong. Instead, he was to coach and mentor them on an individual basis.

4. No Tests and Exams
Confucius never use structured exams to assess the progress of student learning. Instead, he seemed to care more about how students reflected upon what they had “observed” and transformed these observations into knowledge and wisdom.

Confucius once asked a student, “Do you think that my way of acquiring knowledge is simply to study many things and remember them?” The student said, “Yes, isn’t that the case?” Confucius replied, “No, I have one principle which I use like a thread, upon which to string them all.”

His answer demonstrated his awareness of his own thinking process, which we now define as “metacognition.” Metacognitive capabilities can never to be mined through pure retrieval of information, nor would they be assessed by tests and exams.

5. No More “Courses”
Although Confucius did have a curriculum for subjects like music, speech, poetry, literature, and history, none of these was offered in isolation. Instead, he took a holistic approach by cross-referencing all subjects.

Without the boundary of walls, online learning has the great potential to enable multidisciplinary curricula, which I believe Confucius would like to use. In this regard, the concept of “a course” or “courses” would disappear.

6. Survival of the Eager and Willing
How many online instructors are struggling with keeping students on task? How many are frustrated by the fact that no matter how hard they try, there are always some that just won’t learn? Well, Confucius would not let himself be bothered by this problem because he would not teach “dullards,” and would “only teach those who were bursting with eagerness for enlightenment” (The Analects of Confucius). He would certainly apply this principle to online learning, which requires more enthusiasm and self-motivation to complete.

In reviewing Confucius theories, I see some limitations in his thinking (such as the unchallengeable acceptance of traditions). However, there are many of his ideas and practices that may still provoke useful thinking about what we do today. I think reviewing his way of teaching would at least make us aware of the overuse of the behaviorist approach in online teaching, which has a tendency to downgrade learning to one or a collection of short-term trainings.

These Kids Today: The 2007 ECAR Study of Students and Information Technology

The Educause Center for Applied Research (ECAR) recently released its fourth annual research study on the role of technology in student life, which describes their findings of the ways college students use technology and the impact this may have on instruction. In case you don’t want to leaf through the 122-page PDF, you can read Andy Guess’s article in Inside Higher Education for an excellent analysis of the study. But the ECAR report is well worth reading. The tables and stats alone will come in handy for you to whip out at any cocktail party when the discussion turns to “these kids today”.

Researchers found that as suspected, college students are using technology like crazy.

Among the interesting statistics:

  • 73% of students have laptops (although half don’t bring them to class)
  • Average hours per week on the Internet: 18
  • 81.6% of students use social networking sites such as Facebook or MySpace
  • 74.7% have music/video devices
  • 85.1% use instant messaging
  • 43.1 % accessed a wiki every week
  • Over 70% use the Internet (including library databases) for research

This study shows that student use of communication tools such as text messaging, IM, and social networking sites has increased significantly—up 11% since the last study in 2006. Students also make frequent use of Blackboard, email, and discussion boards in their academic work. But although this generation of college students has grown up immersed in these new technologies, they are not ready to abandon real-life human interaction quite yet. Researchers found “themes of skepticism and moderation alongside enthusiasm” among the students regarding the use of technology in courses, noting that 59 percent of students preferred a “moderate rather than extensive use of IT in courses.”

One theme that emerged from the study was that many students found that “the poor use (underuse/overuse/inappropriate use) of technology by faculty detracts from the learning experience.” Complaints included time wasted trying to make equipment work, poorly facilitated discussion boards, and poorly-trained faculty. It is good to know the youth of today are discerning customers. Students won’t buy into the use of technology unless a faculty member can use it well and integrate it meaningfully into the curriculum. Students know that technology alone is no substitute for good teaching practices.

Although student opinion seems to be a bit mixed about the use of technology in the classroom, the overall message of the report is clear: the times are changing and instructors must face the reality that this generation of “digital natives” has grown up with higher expectations for the skillful use of technology and has different ways of learning and accessing information. These new technologies aren’t going away and will just evolve into a Web 3.0 and 4.0 and so on. In the introduction, Harvard professor Chris Dede summarizes the entire state of affairs in one sentence: “Our ways of thinking and knowing, teaching and learning are undergoing a sea change and what is emerging is both rich and strange.” Dede recommends that educators work towards a pedagogical model that fuses the old methods and new, but as this is a bit easier said than done.

The Inside Higher Education article posed some interesting questions regarding the report that I’ve adapted a bit: “How can educators adapt their teaching methods to these emerging technologies? And should they? How are you dealing with this “sea change” and navigating through this ocean of wikis, blogs, RSS feeds, social bookmarking, and all things Web 2.0?

Avatar photo

Is Foreign Language Education Stuck in the Stone Age?

I took my first French class in 10th grade at a public school in a small Alabama town. The class was typical of most high-school foreign language courses. We spoke mostly in English and the assessments were designed to ensure that no one would fail the class. Vocabulary tests typically asked us to pair French terms in column A with their English equivalents in column B, like so:

____ Banane A) Apple
____ Pomme B) Orange
____ Orange C) Banana

Grammar tests were only slightly more challenging and usually consisted of simple sentences with missing verbs to conjugate, as shown below:

Demain nous ___________ à la bibliotheque.
  (go)

I often memorized the verb conjugations the night before (or in some cases, just minutes before the start of the test), then filled in blanks feverishly the minute the exam was in my hands. Speed was key, since my mental snapshot of the proper endings for each verb would begin to blur after five to ten minutes. Occasionally, we were subjected to some other form of memorization torture. This usually involved reciting poems or singing French Christmas carols.

That first year, I thought my French was formidable. (Or, as the French would say, “formidable.”) I could rattle off the French names of almost any object in the classroom. I could tell you exactly how to say I go, you go, and we go. (Saying where I, you, or we were going wasn’t always so easy.) The following year, I transferred to the Alabama School of Math and Science (ASMS), a rigorous magnet boarding school in Mobile. I had to take a French placement exam before enrolling at ASMS, and I knew I was in trouble when there wasn’t a single matching question on the test. My horrible score on the placement exam meant I had to start all over again with French I—along with nearly every other student who had taken a year (and in some cases, two years) of traditional high-school French.

On my first day of French class at ASMS, my teacher explained that our lessons would be built around French In Action, a series of videos designed to teach us French through total immersion. (“Videos” really isn’t the right word, since we viewed everything on gigantic laserdiscs.) As we watched the first few episodes, I was completely overwhelmed. I wondered what language I had been studying for the past year in my hometown, because it certainly wasn’t whatever those people on the screen were using to communicate with each other.

French In Action was part soap opera, part Sesame Street, and it wasn’t great at being either. The storylines were bland and the lesson recaps were repetitive. Yet, despite the actors’ dated haircuts, the overacting, the two-dimensional characters, and the ludicrous plot twists (or perhaps because of them), the whole class was hooked. We were so hungry for anything other than the usual verb conjugation tables and vocabulary memorization that we actually felt invested in the simple narratives. We cheered when Mireille’s bratty sister fell in a fountain in the park. We leaned forward with anticipation when it seemed Robert would finally ask Mireille on a date, and we laughed when our teacher tried to explain a new verb or noun through her own unique system of charades. She would flail her arms wildly, run around the room, improvise with props—anything to avoid a direct translation to English. The goal was to make us think in French, and that’s exactly what the class did.

I went on to major in French in undergrad and was the first student at the University of Alabama to participate in a semester-long exchange program with a French university. (There was another student who was supposed to join me for the adventure, but she went home when she discovered the dorm rooms didn’t offer private bathrooms.) After my semester in France, I decided to move to Germany to live with a few German friends I had made in France. I didn’t speak a word of German at the time.

By the time I left Germany five months later, my spoken German was nearly as good as my French. Of course, there were times when I wished I had learned a few basic grammar rules the old-fashioned way. I was forced to rely on my instincts when trying to conjugate a verb in a complex tense or pair the proper article with a masculine, feminine, or neuter noun. And I still couldn’t explain key differences in the four German cases—nominative, accusative, dative, and genitive—if my life depended on it. However, none of that stopped me from understanding and participating in lots of great conversations with my German friends or communicating with strangers at the grocery store or the pharmacy.

My experience learning German made me wonder (more than ever before) why student progress in learning a new language is still assessed through fill-in-the-blank tests and short essays. Before the days of YouTube, I could understand why an immersive approach to foreign language education was easier said than done. I can still recall how my class “oohed” and “aahed” years ago when one of my professors brought in a VHS tape with a few grainy episodes of Friends that she had recorded while in France. She clutched the precious black plastic cartridge tightly, hugging it to her chest as though she feared one of us might snatch it from her before she could insert it into the VCR. She told us that it cost over two hundred dollars to convert the tape to a North American video format, and we all shook our heads to express our disbelief and our gratitude.

Until recently, supplying students with a German episode of The Simpsons or a Japanese news broadcast required about as much planning and sacrifice as a cocaine smuggling operation. A devoted instructor might record a soap opera during a vacation abroad and carry the tape home like a priceless artifact from an archaeological dig. Once the tape was transported safely, the search would begin for the rare translator of foreign media formats—an elusive code breaker who could make the artifact accessible to the instructor’s students. Today, a wealth of foreign media is only a click away.

So, why isn’t everyone leveraging foreign-language media to create more immersive learning experiences? Some instructors might argue that, YouTube or no YouTube, good foreign language education isn’t primarily about learning enough to understand words and phrases used in popular entertainment and carry on an everyday conversation. The argument over what’s really important in language education (grammar, syntax, and spelling vs. general comprehension and diction) is nothing new. Yet, no matter which side you sympathize with, I think most instructors agree that video and audio can go a long way to promote thinking in a foreign language (as opposed to translation), reinforce key concepts, and burn words and phrases into long-term memory. This brings me to the critical, concluding question of my article, which I hope you will respond to by answering the survey below. (And feel free to provide further feedback by posting a comment.)

Which of the factors below best characterizes your feelings on a media-rich, immersive approach to foreign-language education?

View Results

Loading ... Loading ...

Teaching and Learning: Directions in Technology Research

My colleagues have contributed some very practical discussions of tools and theories for teaching and learning. My entry—which I hope will stimulate some conversation—takes a slightly different direction: research in technology and learning.

The Clark/Kozma delivery truck debate1 shifted the focus of educational technology research away from comparing media or delivery systems (i.e. is using a video lecture better than a face-to-face lecture?) to a more systems-based agenda. This approach recognizes that technologies do not stand on their own, but rather teaching strategy, context, and technology/media each play their part in the learning process. This holistic form of analysis encourages educational researchers to address the following questions:

  • What are the particular strengths of a particular tool when used in a particular context to support particular teaching strategies?
  • Conversely, what are the strengths of particular teaching strategies to support students in a particular instructional context, such as an online learning context?

As an example, Roblyer and Knezek (2003) suggest one research focus might be on ways to increase the learning impact of technologies already in common use. PowerPoint is an existing, commonly used technology. How might one examine the instructional benefits of PowerPoint in a particular learning situation?

We start—as always—with the learning objective: what will the students know, or be able to do? Consider the following example: “Students will be able to describe the overall concept behind and components of the Kreb’s cycle.” To assess this objective, we create standardized grading guidelines (a.k.a. rubrics) for an essay question assessment. Armed with a rubric and a specific assessment method, we can attempt to determine which of the following approaches will best help students grasp this concept quickly and retain it longer.

  1. The concept is described by the instructor in a classroom lecture using a text-only PowerPoint that is then placed on Blackboard for 24/7 access.
  2. The concept is described by the instructor in the classroom using a graphical PowerPoint that is then placed on Blackboard for 24/7 access.
  3. Students collaborate to create a graphical PowerPoint describing the cycle and share it on Blackboard 24/7 with their fellow students.

In each case, the PowerPoint is posted on Blackboard. However, there are differences in the media—one is textual, one is graphical—between instances 1 and 2. The difference between these two cases and case three is, of course, the active-learning instructional strategy of having students collaboratively create the PowerPoint representation of their learning.

I’d be interested in knowing your thoughts on directions in research on teaching, learning and technology. What technologies and teaching strategies work well in your disciplinary context? Let’s start a conversation!


1Roblyer, M.D. (2003). New millennium research for educational technology: A call for a national research agenda. Journal of Research on Technology in Education 36(1) 60-1.

Avatar photo

Seven Teaching, Learning and Instructional Design Myths

Let me start by asking you some questions:

  • Do you take notes while listening to a lecture?
  • Do you multi-task thinking that you can get more things done with less time?
  • Do you try to address as many learning styles as possible in the learning material you’re developing?
  • Do ask your students to practice again and again, thinking that more practice will ensure greater learning?
  • Do you use various media when designing learning materials in order to meet the needs of visual learners, auditory learners, and students with disabilities?
  • Do you use the teaching cycle of giving students an example and the asking them to solve a similar problem by themselves? Do you repeat such a cycle as a way to build on students’ skills?
  • Do you think helping faculty master Blackboard would prepare them to learn another course management system?

If you answered yes to any of these questions, read on to get a glimpse of the cognitive load theory and Ruth Clark’s book Efficiency in Learning. If you are as open and “persuadable” as I am, you may join me in identifying some more myths.

To me, cognitive load theory is like an analogy of the computer vs. the human brain: the hard drive is your long-term memory while the RAM is your working memory. The RAM of your brain or the working memory is a determinant in how quickly you can learn. As your brain gets “booted up” for a learning process, there are three kinds of workloads it has to take:

  1. the necessary load or the “intrinsic cognitive load,”
  2. the unnecessary load or the “extraneous cognitive load,” (This includes distractions or confusion caused by bad instruction.)
  3. and the load that helps connect learning with your own experiences or the “germane cognitive load.” The goal of instructional design is to manage the intrinsic load, minimize the extraneous, and promote the germane load.

I believe that instructional design is a science of “common sense”—the most direct channel to make people “get it.” However, there are a few practices that seem to be based on common sense, but are merely myths when you plug in the cognitive load theory. Here are seven of them I’d like to share with you:

Myth 1: Taking notes helps reinforce learning.

Research cited in Ruth Clark’s book indicated that the effort of capturing what the lecturer is saying and recording it in a written format will only add more extraneous cognitive load to your brain. This leaves less room for your brain to process the content of the lecture. So, as a learner, it would be more beneficial to replace note-taking with jotting the key points during or after the lecture. As an instructor, you should prepare lecture notes for your students and tell them that instead of writing down what you’re saying, they should think about it and reflect upon it.

Myth 2: Multi-tasking makes the learner accomplish more.

Multi-tasking will work effectively if all or the majority of the tasks have become automatic behaviors. In other words, the tasks should be something that you have been doing repeatedly, making them embedded into your long-term memory so that they require minimal working memory to process. However, if you really are to learn something, you better free up your working memory as much as possible. So, tell your students that they can listen to the radio while driving, but stop text-messaging while doing their homework.

Myth 3: In designing learning materials, we should address as many learning styles as possible.

Now let me share with you the only note I’ve taken from Ruth Clark’s workshop: “… cognitive commonality overrules individuality.” I jotted it down because it is such a brave statement. As Ruth Clark said, “I am expecting rotten tomatoes thrown at me for saying this.” Not from me, Ruth. Actually, I am in concert with her in disclosing this myth. When you design a learning object or training tutorial, the attempt to ensure “no child is left behind” may actually leave everybody behind. Instead of accommodating every student’s learning preference, designers should focus their energy on addressing the cognitive commonality that has been scientifically approved, such as letting users control the pace of the learning. process. As for individuality, leave the options that can be controlled by the learner. For example, allow the learner to turn on the audio to hear narration instead of leaving it on by default.

Myth 4: Practice makes perfect.

Does this rule apply to things beyond playing musical instrument? Well, research conducted by learning scientists indicates that more errors are introduced when practice goes beyond certain timeframe. I think this is the way that your mind tells you “O-kay, I got it. Stop exhausting me!” So, think about the amount of practice you want to assign to the students. By the way, I wish I could send this information to my elementary, middle, and high-school teachers in China. If I had a dollar for every hour I spent writing each Chinese character 100 times, I’d be a millionaire!

Myth 5: Multimedia improves learning by addressing multiple learning styles.

Insert the images for visual learners, add the audio narration for the auditory ones, and don’t forget the text for people who like to read. (And be sure to make the text big and bold for the visually impaired). Oh, and what about some background music for today’s multi-sensory learners? Thanks to today’s technology, all these requests can be easily accomplished through multimedia. The question is: will this piece of multimedia help people to learn better? The answer is likely to be no because media redundancy distracts learners. It adds unnecessary workload (extraneous cognitive load) to the brain, leaving less room for it to process the information. So, instructional designers who are constantly tempted by various fancy tech tools ought to remember that making things simple and direct remains the rule of thumb.

Myth 6: Giving examples followed by practice, followed by another example, followed by more practice (“example + practice + example + practice + example + practice”) is an effective way to teach.

Although the effectiveness of this method is questionable, this was a method often used by my teachers. Today, I still rely on this method from time to time when I teach. When compared with this formula [(example + practice) x N], research found that another formula which provides various examples before asking students to practice (example + example + example +example…+ practice) proved to be a more efficient way to learn. However, providing students with examples on which to model their own work can be problematic in its own right. I often debate whether or not I should give students a completed sample of an assignment. On the one hand, it may help clarify my expectations. Yet, on the other hand, it may also lock students to a pattern and constrain their creativity. If you share the same concern, let’s try giving students a diverse variety of examples so that they don’t follow one pattern and can practice better and learn more.

Myth 7: Being an expert of one course management system makes it easier to learn another.

Research shared by Ruth Clark shows that when expert and novice chess players were given a random chessboard, the novice group actually remembered it better than the experts. Why? Because the experts are confused by the meaningless layout of the board and are subconsciously going through a process of differentiating the “new look” with the one that they are familiar with. Those who are familiar with the layout and navigation scheme of Blackboard face the same frustration when they have to learn another system that is structured differently. This finding brings an alert to decision makers to think more carefully in selecting or changing technology solutions for users, especially those who are already comfortable using one particular application. (This is true even if they also complain about it.) The finding also echoes Gerry McGovern’s call to end web redesign in his article “Web Redesign is a Bad Strategy.” McGovern advocates that designers put more energy into improving content and simplifying the existing structure instead of building a new one.

What’s next?

Feel free to share any teaching myths or miracles by leaving a comment.