Newsela – A Follow-Up Look

From time to time, I take a look at a topic I was interested in, say 5 or so years ago, and ask what has happened since. Have classroom strategies that seemed to be enjoyable and productive survived and how have they matured? Here is an example of what I mean.

A decade ago, I was interested in the potential of technology for allowing greater individualization of instruction. My primary interest was in technology that allowed ideas from the 1970s-80s called mastery learning to become practical. Mastery learning proposed that group-based instruction largely ignored differences in aptitude and background knowledge, leading to frustration and learning challenges because the group advanced whether individuals were ready or not. To relate this to a widely recognized alternative based in technology consider the approach now allowed by the Kahn Academy. 

A different approach, maintaining more of a group-based strategy, was proposed by Newsela. This company argued that reading content (individual articles) could be presented at different reading levels allowing a class to read versions of the same material maintaining the opportunity for social opportunities such as class discussions. This approach made sense to me especially when applied to reading tasks that might be described as reading to learn – e.g., assignments in science, social studies, etc. The focus on informative content rather than fiction had obvious implications for present student learning and for the future. The following two images contrast the same content presented at different reading levels.

I wrote multiple posts describing Newsela and how it might be implemented. Others were offering similar observations.

Individualizing literacy instruction with Newsela (2015)

Layering Newsela (2017)

Not all good ideas work or are practical so I decided to follow up what is now a decade later and see how the company and the product seem to be doing. 

Adoption Level

It is difficult to get accurate information about student use. Newsela currently reports that it is used by 3.3 million teachers and 40 million students, the exact total it reported in 2016. Newsela has a lite and pro level and the lite level has attracted a lot of attention and occasional use. Occasional is a guess as I could not find stats on the level of activity. For some classrooms and individuals reading an occasional story would be a productive activity. I am assuming that the combination of those using the lite and the paid levels accounts for the differences in usage statistics that are reported.  

The paid version is better suited to using the tool as part of the curriculum. The startup’s paid product is between $6 to $14 per student. Newsela is sold at a rate of $6000 per school or $1000 per grade. Newsela estimates that gross bookings have grown 115% over the years of the pandemic, and that revenue grew 81%. More than 11 million students were using Newsela under a licensing agreement by the end of 2021.

The version of Newsela I described in my late 2010s posts has changed substantially. Newsela has significantly evolved in recent years to become more AI-driven, expanding both its suite of educational products and the ways users interact with its content and assessment tools. A secondary emphasis on writing has emerged. Usage trends reflect a shift toward greater integration of artificial intelligence and differentiated instruction, as well as changes in accessibility and assessment features for teachers and students.

Efficacy Studies

My tendency when advocating, or at least describing, an instructional strategy implemented through a commercially available tool or product is to search for published research that evaluates the approach I want to describe. The following are descriptions of two studies I located. 

WestEd (2018) Newsela efficacy study: Building comprehension through leveled nonfiction content.

Classes of fifth-grade students from two districts were randomly assigned to a Newsela or a Control condition. Reading instruction in the Newsela classes was modified to include at least two Newsela articles per week – one in class and at least one at home. Students in the Control condition relied on their normal reading curriculum. The study ran for 14 weeks and used the difference in STAR pre and post-performance scores as the dependent variable. Student compliance with the Newsela homework expectation varied widely, with 55% meeting the one-per-week expectation. When those meeting the expected level of engagement were compared with the control group, their achievement gains were significantly greater.

Literacy gains from weekly Newsela ELA use

This year-long study made use of differences in pre and post-MAP reading assessments as the dependent variable. The classes of third and fourth grade educators participated as Newsela and control conditions. The Newsela classes were asked to read at least two stories and take one multiple-choice test per week. Teachers in the control condition relied on their own selection of reading material with the largest source being content they had found through Google sources. Fourth-grade students, but not third-grade students, achieved at a significantly higher level in the Newsela condition. 

Why can’t I find peer-reviewed published studies

Often, I am frustrated when I cannot find studies that directly support the strategy I want to describe. This is the case with Newsela and I have been thinking about why this is the case.

Newsela has engaged outside agencies (e.g., WestEd) to conduct research using their products, but these studies are available as what I would describe as technical reports and don’t seem to appear in scholarly journals. After reading these reports I can see that if I had been asked to review the research for publication, I would also identify issues that would cause me to suggest that the study not be published. In the studies I will describe here, I see flaws in the research design that allow alternate explanations for the positive results. 

Applied research is often very difficult because those implementing the research have their own issues and priorities. Sometimes a methodology does involve tight controls from the beginning and sometimes it seems that original design is allowed to slip as unanticipated issues come up. 

For example, in the first study I describe, the plan was to have a control group and a Newsela group with one in-class and one homework reading assignment a week. It turned out that the homework assignment was ignored in many cases and to generate significant evidence that Newsela was productive the researchers compared those who did the homework against the control group. This may not be important, but it could also mean that the Newsela group now consists of more motivated readers than the control group, and this interest in reading, rather than the Newsela content and approach, was what created the difference in the development of reading skill. It is unclear to me from reading the description why expectations for completing the homework such as including completion as part of the grading scheme was not implemented. I can imagine a different controversy if what I propose was implemented as you would then extra reading required in the Newsela group and not in the control group. Perhaps the most ideal approach would be to maintain control of all of the reading assignments within the classroom setting so that the time allocated could be matched. 

The second study I have described is limited by what I would describe as clear identification of what is the intended independent variable. What has always attracted my interest in Newsela was the group-based, but individualized approach the content allows. Each Newsela document is available at multiple level (5) of complexity. This allows those readers at different levels of aptitude and skill development to read a variant of the same content so that discussion and a social element of instruction can be maintained. My personal interest in technology-supported learning has always been based on the potential of individualization. One argument some make about many technology applications that allow for differences in rate of learning is that students are isolated and miss out on the social benefits of a classroom setting. Newsela offer an alternative approach that maintains the social setting. 

This study creates a different or at least an added difference when comparing the Newsela group and the control group. The authors report that when teachers select the reading content for the control condition this material differs in category from the Newsela treatment. Teachers in the control condition were described as relying on Google searches to find content fitting with the topics that they covered and this content contained significantly less “nonfiction” content. A cleaner approach more consistent with what I think is the unique Newsela content would be to compare the impact of a single version of articles versus multiple versions of the same articles. 

Summary

As a commercial venture Newsela seems to be doing well. It has a solid base of schools that have committed to purchasing the program. My criticism of the weak methodologies used in evaluation efforts is mostly a function of my interest in the impact of the individualization efforts the resources provide. Having current and nonfiction content is important, but the strategy on which the company originally made its name has not been rigorously evaluated.

It now seems educators could use any of several AI tools to create similar content. Prompts such as rewrite this content at a level appropriate to fifth grade students could be applied to any content a teacher could upload. Given this option, the value to a district would depend on the time savings to teachers and the constant access to new content would be the advantages Newsela provides. 

Loading

Where is the thinking in the AI classroom?

The concept of generative activities has consistently shaped my thinking and teaching on learning. I admit that such activities are not ultimately necessary. Still, they represent ways for learners and those who try to help them grow to understand and imagine how skills and knowledge might be applied. Focusing on generative activities was particularly useful in my interest in studying – the work an individual does to make experiences personally informative and useful.

Basic Definitions:

Studying – the mental and external activities a learner engages in after exposure to potentially useful experiences that are intended to store a representation of these and create meaning.

Generative activities – external tasks intended to encourage productive cognitive (mental) behaviors 

Why is this perspective important at this time? My concern is that certain uses of AI are frequently being substituted for generative activities allowing individuals to accomplish tasks without achieving the cognitive benefits (i.e., retention, understanding) engagement with the generative tasks make more likely. 

Why would learners substitute AI for generative activities? It seems likely they see AI as producing an equal or even superior product without the effort required to create such products on their own.  This reflects both a focus on short-term benefits over long-term benefits and probably a lack of understanding of how personal knowledge and skills are developed, or perhaps even a disinterest in developing these personal attributes.

Some background:

When I explain the concept of generative activities I like to start with Rothkopf’s concept of a mathemagenic task because this researcher’s focus tends to make intuitive sense to most people. Rothkopf was interested in questions and variations in how questions might be associated with written material. 

Questions presented before you read. 

Questions presented after you read. 

Inserted questions – questions added within text. 

Different types of questions – application questions, factual questions. 

Mathemagenic tasks

The made-up word mathemagenic translates roughly as giving birth to knowledge, implying that in attempting to answer questions, you might accomplish something else – a better likelihood of future retention, greater likelihood that you would recognize possible applications – that would not have occurred without exposure to the questions. My favorite example relates to the challenge educators often face in encouraging students to see the relevance of general concepts they have been taught. This translates as connecting new ideas with what you already know. The examples and the relevance are potentially there if you can make connections. So, what not ask students directly – provide an example of XXX? If personal examples exist, but learners have not made the effort to make the connections, perhaps the request will encourage that specific cognitive effort. 

There is a huge body of research on all aspects of questioning. Questions are an everyday classroom activity, but the insight is just why do we spend the time, and could a more careful use of questions result in improved results? My favorite example here is what is called wait time – the average delay after asking a question (silence to allow thinking) is a little over a second. If we want students to think, typical behavior in classroom discussions is not particularly rationale. There is reason to examine and challenge typical behavior.

Anyway, questions are an external task that can be used to manipulate – change the odds of – productive cognitive behaviors. I suggest adding one important final point: a learner can ask herself questions, e.g., using flashcards. So various ways in which questions can be generated and used are an aspect of what those interested in study behavior investigate. 

Generative Tasks

The concept of generative activities is simply an expansion of this same idea and asking questions would be one of many generative strategies. The idea of generative activities is not new (Wittrock, 1974, 2010) and to many educators may seem obvious and a reflection of common classroom practices. While true, researchers have attempted to understand the underlying mechanisms and to consider just how efficient different activities were especially in the comparison of one to others ( Fiorella & Mayer, 2016). A personal interest and one clearly relevant to the topic of how AI is applied in classrooms is writing to learn. I have always felt through self awareness that requires careful examination of existing ideas and integration of ideas from a variety of experiences to produce a product. There is a substantial body of research to support such perceptions (e.g., Graham et al., 2020). To be clear, researchers consider a variety of writing activities under the umbrella of writing to learn. The product need not be a massive, semester summarizing paper, but perhaps also notes and short, five-minute end of class descriptions related to the content just presented. 

Caveat

One issue I think is important that may not be apparent in the notions that generative activities are intended to encourage productive cognitive skills is that such skills may occur without this external requirement and guidance and there is always the possibility that for some motivated and capable of thinking deeply, without such tasks, the task represents a form of “busy work”. In other words, the task adds little beyond annoyance. Of course, the reality is that educators in actual classrooms typically do not feel that they can arbitrarily assign tasks to some students and not others, so they must always deal with reactions to assignments, both legitimate and resulting from laziness. 

AI and Generative Tasks

AI discussions related to education always seem to generate a good news / bad news situation. There seem to be several examples that apply to this general topic.

AI can be applied to render the potential benefits of a generative strategy useless. For example, if AI is used to respond wholly to a writing-to-learn assignment, the learner completes the assignment without engaging in much cognitive work. The educator is then in a position of assigning a task that takes valuable learning time and adds a commitment to the effort to provide feedback, but has little impact. 

In contrast, AI can be used to formulate questions (both objective and open-ended) related to assigned material and to respond to a learner’s responses to such questions. Learners can even generate such activities on their own.  It seems to me that the use of what might be described as short essay questions offer a unique advantage that would be difficult or at least very time consuming for the educator to administer. AI tools are very flexible and can ask and react to the answers for different types of questions. Short answer questions are a form of writing to learn and involve greater “retrieval practice” benefits than formats such as multiple choice that are useful, but less demanding of retrieval. 

Summary

My effort here was intended as a way educations might frame their way of thinking about AI in classrooms using  examples I assume are familiar. I hope this approach can be generalized. Of course, the challenge is in manipulating AI-based and any assigned activities so that productive thinking activities are encouraged and also that students gain insight into the importance of the mental work that is required of certain task. I understand this may seem obvious, but the work of adjusting to the advantages and disadvantages of AI tools will take some time and careful study. For example, I wonder if writing and organizing notes may accomplish much the same benefits as creating a writing to learn product. Learning to write is somewhat different than writing to learn although writing across the curriculum offers a secondary benefit of practicing writing skills. There are plenty of options to consider. We presently do little to teach advanced note making skills and note using skills even though these topics have received a great deal of attention as benefits to out of school functioning. 

Citations

Fiorella, L., & Mayer, R. E. (2016). Eight ways to promote generative learning. Educational Psychology Review, 28(4), 717-741.

Graham, S., Kiuhara, S. A., & MacKay, M. (2020). The effects of writing on learning in science, social studies, and mathematics: A meta-analysis. Review of Educational Research, 90(2), 179-226.

Rothkopf, E. Z. (1970). The concept of mathemagenic activities. Review of educational research. 40(3), 325-336.

Wittrock, M. C. (1974). Learning as a generative process . Educational Psychologist, 11(2), 87–95. https://doi.org/10.1080/00461527409529129

Wittrock, M. C. (2010). Learning as a generative process. Educational Psychologist, 45(1), 40-45.

Loading

AI and the Writing Process

I found the Writing Process Model (Flower & Hayes, 1981; Hayes, 2012) helps me think about the development of writing skills and the specific application typically described as writing to learn. This model identifies the processes and the interaction among the processes involved in writing and has been used to guide both writing researchers and the development of instructional tactics. 

The model provides researchers and instructional designers with a concrete framework to work with by identifying specific skills that can be studied as the source of individual differences in writing skills or targeted for development, assuming that greater proficiency with these skills will lead to more effective writing.

I first used this model to speculate how specific technology tools could support writers. For example, technology offers powerful ways to take and review notes, a method for planning by creating an outline or “mind map”, and a way to record text that allows for easy manipulation and revision. I have begun to think about the model in a different way brought on by easy access to AI for writing in general and in classrooms more specifically. Given the general goals of learning to write and writing to learn, when do specific uses of AI facilitate and when do these applications harm the development of the writing subprocesses as involved in each category of writing experience? Put another way, instead of thinking of AI as an all-or-nothing approach to creating written content, would it make more sense to evaluate the use of AI in impacting writing subprocesses and perhaps have students use AI more selectively?

The Writing Process Model

The model identifies three general components a) planning, b) translation, and c) reviewing (see the following illustration). Planning involves setting a goal for the project, gathering information related to this goal, which we refer to as research, and organizing this information so that the product generated makes sense. The goal may be self-determined or the result of an assignment. Research may involve remembering what the author knows about a topic or acquiring new information. Research should also include identifying the characteristics of the audience. What do they already know? How should I explain things so that they will understand? Finally, the process of organization involves establishing a sequence of ideas in memory or externally to represent the intended flow of logic or ideas.

What many of us think of as writing is what Flower and Hayes describe as translation. Translation is the process of getting our ideas from the mind to the screen and this externalization process is typically expected to conform to conventions of expression such as spelling and grammar.

Finally, authors read what they have written and make adjustments. This review may occur at the end of a project or at the end of a sentence. Authors may also solicit advice from others rather than relying solely on their own review.

One additional aspect of the model that should not be overlooked is its iterative nature. This is illustrated in the figure, which presents the model using arrows. We may be tempted, even after an initial examination of this model, to view writing as a mostly linear process – we think a bit and jot down a few ideas, we use these ideas to craft a draft, and we edit this draft to address grammatical issues. However, the path to a quality finished product is often more circuitous. We do more than make adjustments in spelling and grammar. As we translate our initial ideas, we may discover that we are vague on a point we thought we understood and need to conduct further research. We may decide that a different organizational scheme makes more sense. This reality interpreted using our tool metaphor would suggest that within a given project we seldom can be certain we have finished the use of a given tool and the opportunity to move back and forth among tools is quite valuable.

This model describes the processes identified by Flower and Hayes, but ignores two other components. The first is the writing task, which consists of the assignment and any writing completed at a given moment. The other missing element is the long-term memory of the writer. The long term memory or existing knowledge provides a source of information and strategies that the writer can use without resorting to new research. 

Hayes (2012) updated this model to incorporate additional research and comments from colleagues and his own laboratory. Later work placed a strong emphasis on both revising the planning component and subsequent drafts, as well as on the role of working memory. Working memory recognizes that cognitive capacity is limited, meaning activities must fit within the existing capacity or perhaps be ignored. Processes become less demanding less capacity as a function of practice and aptitude and among other topics, is vital in understanding changes that could occur for learners across the grade levels. For example, Hayes noted that keyboarding is more difficult for younger learners than handwriting and products produced on a computer would thus appear of lower quality. Differences in cognitive demands do change with experience with keyboarding, eventually becoming less demanding than handwriting. Other factors can also impact cognitive demands and in one example, Hayes noted that studies requiring adults to write in all caps reduced the quality of the final product because transcription in this form is less practiced. As an aside, I could not help recognizing the tendency of some to write in all caps on social media and my perception of the quality of what is produced in that format. Anyway. Finally, Hayes recognized that writing was a motivated activity and differences in motivation could emphasize the processes.  

One of the significant patterns in writing proficiency Hayes and others (Bereiter & Scardamalia, 1987) associated with the limited capacity of working memory and the gradual development of proficiency in the cognitive writing subskills, was the tendency to move from what Hayes and others describe as knowledge telling to knowledge transforming. The first approach results in a dump of ideas from long term memory triggered by the assignment and information that has already been recalled. In other words, a product can be generated without changing what is stored in memory or how this content is organized. Most instructors have a feeling for how this works. They have asked students to reply to an essay question and received at least some responses that seem to be everything students knew about the topic rather than a specific answer to the question. It can be hard to know if the student thinks they have answered the question or if this is just a frequently productive ploy. It is the first situation that shares characteristics with the idea of knowledge telling.

In contrast, writing classified as knowledge transforming requires that previously stored information be reorganized, reinterpreted, or extended based on speculation or insights. The writing task sometimes determines the difference, but transforming is more demanding and, when expected, is more likely to be produced by more capable writers. Writing to learn could involve either output, but the greater manipulation of ideas in knowledge transformation reflects the most significant benefit. 

Klein (1999) offers a review organized around the writing processes hypothesized to be responsible for learning. A brief summary of some insights from this paper provides examples of tasks and skills that differentiate knowledge telling from knowledge transforming, also relating the distinction to the processes involved. 

  1. Point of utterance (no revision and limited planning). This explanation assumes that learning occurs in the attempt to generate comments on the topic (no revision or planning is expected). This category might be described as spontaneous writing – the learner is asked or personally commits to writing on a topic. An activity fitting within this category would be the five minute writing tasks some college lecturers assign at the end of a class. 
  2. Genre-related – This explanation focuses on the benefits of transforming ideas to the structure demanded of a specific genre. For example, in a “position paper,” a writer is expected to take on of several possible positions are generate an argument supporting this selection. In selecting, organizing, and connecting knowledge to fit the demands of a formal writing task, the learner creates an understanding that would not exist without the imposed task.
  3. Backward search – This explanation assumes that skilled writers formulate complex goals (characteristics of the desired product, audience needs, etc.) and then rework existing knowledge in terms of these goals. This seems a more generalized version of the “genre-related” explanation, requiring more sophisticated and complex problem-solving.

My own writing is generally linear with backtracking depending my assessment of how well things are going. This impression is based on my own behavior and may not accurately reflect the ideal recursive approach. When I take on a project, I assess what I already know and have notes on a topic and then read and take notes on additional material. I generate a rough idea of how these ideas could be organized (some would create a formal outline) and begin developing a draft. Often, I realize I need to fill a hole in my mental outline or find a reference in support of what I am trying to accomplish and have to take a closer look at my large collection of digital notes or read and take notes on another source or two. Upon completion of this initial draft, I reread what I have down to take care of lower level deficiencies (spelling, grammar) and often to add a little more material here or there so the document makes more sense or hangs together better. When writing for myself and without an external reviewer, I tend not to make major structural revisions. However, when I submit a paper for review, I do sometimes have to make larger adjustments even if the content is deemed useful. 

Where does AI play a role? Where should AI play a role? I have worked through my thoughts on these questions based on my own circumstances and also on what should be the circumstances for those in more formal learning settings. I want to produce content that meets a reasonable standard of quality, but I am not that interested in becoming a more accomplished writer. I want what I write to be a reflection of my experiences and what I have learned, and I assume I can learn from the process of writing. I am willing to invest time in this objective. I propose that others generate a similar analysis for themselves based on personal goals or goals for others they may be responsible for educating. 

How do I use AI? I primarily use AI to facilitate the research and planning components of my writing process and to perform some of the revision tasks. These emphases are consistent with my desire to learn from writing and my lesser interest in improving my writing skills. I am not suggesting everyone apply my priorities, I am suggesting it is possible to identify priorities and use AI strategically and efficiently.

Here is how this works. I have developed a large collection of notes consistent with methods of personal knowledge management and a technology-enabled second brain. This is a multidecade-long process that involves reading widely in the educational technology and educational psychology literature. When I get an idea for a writing project, I use AI to query this body of content for ideas related to my intended project. Based on the content I produce and queries for related information (some generated by AI), I review this new content and supplement my notes on the topic. I then often use AI (typically NotebookLM) to provide a structure for the intended project based on my collection of notes. I consider this proposed approach when generating my initial draft.

I use Grammarly heavily when I write. I use the pro version so I get constant feedback not only on spelling and basic grammar, but also more substantive recommended changes at the paragraph level. Grammarly identifies issues I should consider and offers suggestions. I suppose this could be a valuable learning opportunity, but I admit I just select the recommend changes that sound good. Grammarly has just pushed a major update that offers even more capabilities, but I have yet to explore which might be useful to me. 

The use of AI for research, planning, and text revisions do not limit my writing to learn. I think it is quite reasonable to associate AI capabilities with specific components of the writing process model, as applied to writing and writing to learn, in order to accomplish tasks that are not essential to developing writing skills or learning opportunities. The stickier problem presents itself when you must consider how to control the use of AI. 

Sources

Bereiter.C. & Scardamalia, M. (1987). Two models of composing processes (pp. 1-30). In C. Bereiter & M. Scardamalia (Eds) The psychology of written Composition. Erlbaum. 

Flower, L., & Hayes, J. R. (1981). A cognitive process theory of writing. College composition and communication, 32(4), 365-387.

Hayes, J. R. (2012). Modeling and Remodeling Writing. Written Communication, 29(3), 369-388. https://doi.org/10.1177/0741088312451260

Klein, P. D. (1999). Reopening inquiry into cognitive processes in writing-to-learn. Educational Psychology Review, 11, 203-270.

Loading

NotebookLM Enhancements

I was excited when Google announced its investment in developing NotebookLM as a tool for learners. First, I encountered the description of its new program Learning Your Way, which promised to personalize textbooks in various ways, including explanations based on personal interests. Examples of what this would look like were provided, but to submit your own content for the “learn your way” treatment, you first had to indicate your interest and then wait to be selected. I am still waiting. However, Google has also announced enhancements for the existing NotebookLM that make several proven study strategies more straightforward to implement, as well as a couple of unique ways to repurpose instructional text as a podcast or something resembling a slide-based lecture. The updates available to all are what I will describe here.

Before anyone gets overly excited, what Google can presently make available requires that an instructor or a learner must be able to load content into NotebookLM. There is no legal way in which what I will describe can be done with your existing commercial paper or even digital textbook. Companies producing instructional materials may either collaborate with tech companies or develop similar capabilities, but they are not doing so at present. I wrote a couple of textbooks earlier in my career. When my collaboration with a textbook company ended, I was given complete control of the copyrights and now I have full-length chapters I can use for demonstrations. This is what I am doing here. NotebookLM notebooks can be shared and I will make the content on which this tutorial is based available for you to explore

For those unfamiliar with the newest updates, NotebookLM presents as three vertical panels. The leftmost panel shows the content you have uploaded to NotebookLM as well as provides the opportunity to upload additional content and open individual files for viewing. The middle panel is where AI prompts are applied and the output from such prompts are displayed. The rightmost panel is where the older and new AI tools are activated and output from these tools can be accessed for later display. 

As soon as NotebookLM is launched, the middle panel displays a summary of the source files that have been added. Loading source files does not include them in this summary unless they have been selected (see check box). NotebookLM was developed as an RAG (retrieval-augmented generation), which means the prompts are applied to the content you provide rather than applied to a massive corpus available to the model. The new “Discover” tool available from the top of the sources panel breaks with this approach and allows a user to request additional information intended to be related to the purpose for a particular notebook that can be reviewed and added as an additional source. 

Before I continue, I want to identify one frustration I have with the way NotebookLM fits into my workflow. Developers have specific applications in mind when creating applications, and their vision may not include what an individual user prefers. My preference is to read, highlight, and annotate content before using an AI tool. While NotebookLM allows access to sources, it does not support highlighting or annotation, nor does it display such added layers from uploaded material. The workaround is to export personal additions and add this content to NotebookLM separately. 

Mindmap

New capabilities appear in the right-most column. The audio overview (podcast) and Mindmap capabilities were already available and will not be covered here. 

Quiz

The new tool that first caught my attention was the quiz. I have explored various ways to have LLMs ask questions about designated content, but having a built-in tool for doing so is unique. The tool generates a double-digit set of multiple-choice questions that are then stored as an associated resource. I have no idea why the number of items varies. When selected, the MC questions are displayed one at a time (see the following image) and then, no matter which alternative is selected, a comment is provided. The “explain” button (see red box in the following image) shows the content from the text associated with that question. If you want more questions, the set generated can be deleted and a new set generated. 

I think it is valuable when using this tool to be aware of your reaction and recognize you are working with an AI tool. As someone who had to generate multiple-choice questions for actual student exams, knowing what I know about student reactions, I would not use some of the AI-generated questions on an examination. I thought the stems of some AI questions were kind of vague or misleading, but the explanation clarified the logic justifying a given answer. If you understand this and are using the questions as a study experience, the ambiguity is probably fine. As a final comment to this post, I will explain why I prefer to ask an AI tool to generate open-ended (essay) questions on a specific topic and then wait to evaluate my response. Interacting with content in this way is more helpful. 

Flash Card Tool

Flash cards are a useful tool for implementing memory enhancement strategies such as retrieval practice and interleaving. While relying completely on flashcards would not be wise, this approach is effective and useful for quick reviews spaced over time. Again, the “Explain” button (see the second of the following images) offers a connection to the original content and this provides potential opportunities to review beyond simply verifying that a response is correct. Recognize that NoteBookLM generated 69 cards for the targeted chapter, so this tool covers the content in detail. 

Video Overview

The video overview offers what seems like an extension of the existing audio option. It strikes me as the equivalent of a lecture based on a series of slides. So imagine a large lecture environment in which you see the screen on which an instructor is projecting her PowerPoint and you are hearing the audio but not seeing the presenter. Having written this chapter, it is an interesting experience watching a lecture based on your words. It is possible to download the video so I decided to offer this download through YouTube if want to see an example. 

Just Use Chat

NotebookLM, before these built-in tools, was a way to interact with uploaded content using AI chat prompts. I find this capability more powerful than the tools that seem to recreate standard study experiences. The familiar tools are probably a way to engage less experienced learners, but I would argue that concepts such as debating and tutoring offer opportunities that take advantage of the language capabilities of LLMs. The following is an example of a chat I had with the content.

[Prompt] Ask me a short-answer question about generative tasks and wait to evaluate my reply

According to the sources, what two specific categories of generative activities are highlighted for their value in improving understanding, retention, or transfer, and what general type of content processing is involved in these activities?

——————————————————————————–

I will now wait to evaluate your reply.

[Learner reply] I remember the chapter talked about responding to questions and I can’t remember another example from the book. How about writing to learn, but I don’t think that was in the chapter.

[AI response] That was a strong attempt, and your intuition about “writing to learn” was very close to one of the specified terms!

Here is the answer to the question, based on the specific terminology used in the sources:

The two specific categories of research emphasizing generative activities that improve understanding, retention, or transfer are labeled as:

1. “Authoring to learn”

2. “Teaching to learn”

These activities involve a “generative processing of content”.

Your suggestion of “writing to learn” is directly supported by the text, as “authoring to learn” is described as involving the preparation and implementation of writing tasks. This generative processing of content is suggested to be superior in many ways to studying only for personal use.

While responding to questions is certainly a critical activity discussed in the sources—often serving as an external strategy for evaluation and regulation that compensates for weak metacognition—it was not one of the two specific research categories given the label “generative activities” in this context.

[ Added comment] What is not visible in my copies of this text from this interaction are the inserted numbered buttons that provide access to sections of text from the original text.


Notebooks can be shared. You are invited to explore the Notebook that was the basis for this prompt.

Loading

The Role of Writing in K-12 Education: Challenges and Opportunities

In K-12 education, writing often takes a backseat to other academic priorities. Elementary education is dominated by math and reading, while secondary education spreads its focus across a wide range of subjects. Unfortunately, this distribution of attention has left the development of writing proficiency hindered by several factors.

Limited Time for Writing Practice

One significant issue is the limited time students spend on writing activities, both in and out of the classroom. Research shows that only about 25% of middle school students and 30% of high school students meet the recommended minimum of 30 minutes of daily writing practice. Writing assignments at these levels are often brief, typically a paragraph or a short essay, with few opportunities for more complex projects that require synthesizing ideas from multiple sources.

While initiatives like “writing across the curriculum” aim to increase writing opportunities and integrate writing into other subjects, many teachers outside of language arts lack the training to effectively incorporate writing into their instruction (Picou, 2020). This lack of preparation, combined with time constraints, contributes to disparities in writing proficiency, as reflected in differences in NAEP scores across schools (Mo & Troia, 2017).

The Broader Benefits of Writing

Writing is not just a skill—it is a multifaceted process that integrates numerous subskills and offers significant cognitive and academic benefits. However, these benefits are only fully realized when students receive meaningful feedback on their work. Unlike other disciplines, evaluating written work is particularly time-intensive for educators, which may discourage frequent and substantive writing assignments.

Beyond skill development, writing also serves as a powerful tool for learning. Often referred to as “writing to learn,” this process involves cognitive demands that enhance understanding and retention (previous post). Writing tasks act as generative activities, externalizing thought processes and encouraging deeper engagement with the material. For example, when students are asked to provide personal examples of a concept, they connect prior knowledge to new ideas, fostering meaningful learning.

Two types of writing tasks stand out in their educational value: writing to explain and writing to persuade.

  1. Writing to Explain
    This task requires students to learn something and then articulate their understanding through writing. The act of externalizing knowledge serves as a form of self-assessment, revealing gaps in understanding and prompting further learning. This process, often linked to metacognition, helps students refine their knowledge as they work to organize and express their ideas. As educational psychologist Graham and colleagues (2020) note, writing to learn has consistently been shown to enhance academic outcomes.
  2. Writing to Persuade
    Persuasive writing involves crafting a position, supporting it with evidence, and addressing counterarguments. Despite its importance in developing critical thinking and reasoning skills, persuasive writing accounts for only about 20% of writing-to-learn tasks. This is a missed opportunity, as persuasive writing offers concentrated practice in analysis and argumentation, much like debate, but in a more efficient format for classroom use.

The Impact of AI on Writing Development

While writing is already underutilized in education, the rise of AI tools presents a new challenge. Educators are grappling with how to integrate AI productively without further reducing the time students spend writing. If AI tools are used to complete writing tasks for students, the generative benefits of writing—such as critical thinking and cognitive engagement—may be lost.

As someone who uses AI tools daily, I recognize their potential to enhance productivity and creativity. However, I also understand the risks. For example, I use AI to interact with a personal corpus of notes, allowing me to explore ideas before writing. This approach complements my writing process rather than replacing it. Students, however, may lack the motivation or understanding to use AI in similarly constructive ways. Under time pressure, they may rely on AI to bypass the cognitive effort required for writing, undermining the development of essential skills.

Addressing the Challenges

One strategy that I think would address both writing challenges would require an increase in supervised classroom writing. Such tasks could be improved with collaborative writing activities that included peer editing and revision. The peer responsibilities would include attention to both writing quality and content accuracy when the task is a writing across the curriculum task.

Moving Forward

The challenges facing educators are undeniably complex, and the rise of AI adds another layer of difficulty. However, ignoring these realities will not improve the situation. Writing remains a critical skill, both as a standalone competency and as a tool for learning across disciplines. By increasing classroom writing opportunities and leveraging collaborative approaches, educators can help students develop the skills they need to succeed in an AI-driven world.

I welcome your thoughts on this analysis and any ideas you might have for addressing the interconnected issues of writing development, AI integration, and educational priorities.

Sources

  • Graham, S., Kiuhara, S. A., & MacKay, M. (2020). The effects of writing on learning in science, social studies, and mathematics: A meta-analysis. Review of Educational Research, 90(2), 179-226.
  • Mo, Y., & Troia, G. A. (2017). Predicting students’ writing performance on the NAEP from student- and state-level variables. Reading and Writing, 30(4), 739–770.
  • Picou, A. (2020). Are schools making writing a priority? New study shows students are not spending enough time. The Learning Agency Lab

Loading

AI and Confirmation Bias

Everyone needs to be aware of confirmation bias. This cognitive tendency is responsible for our focusing more on those ideas and facts that support our existing way of understanding the world. It applies to what information we select when given a choice and to the information that “sticks” when perceived. When I try to explain the second issue, I often describe what we already know as a model or perhaps an outline and new information that fits with this model or outline is easier to store and recall. This characteristic of human information processing is described as a bias because it argues that we find information that does not fit our existing way of seeing the world as less likely to influence us. We might argue we are open minded, but we still must work to identify and process challenges to existing ways of thinking.

Technology has long been argued to work in ways that support confirmation bias. Back in the early days of online search, there was a concern for what was called a filter bubble. This was the notion that search engines tracked our preferences and would bias hits appearing early in the list of possible sources based on our past choices as we responded to search results. Did we select results from the top of the response list or not? I read about this “problem” and struggled to find a way to test it. After reading accounts of the problem, I decided I could use my own online history as a test. I spent a lot of time searching and reading about technology and had made use of Apple equipment for years. It occurred to me that the word Apple was ambiguous — it could mean a computer company or a fruit. I reasoned that if I searched anonymously and when using a browser aware of my identity and searched for “apple” I should get different hits at the top of the search results. The results were fairly inconclusive and always favored computers. I may have been wrong about anonymous searchers, or perhaps just the notion that more people would want to know about apple the fruit.

AI and Confirmation Bias

I have tried to identify ways in which AI prompts might generated personally biased responses and started with some searches. The first warning I discovered argued that the prompts themselves could be an unrecognized problem. Psychology Today reported that Confirmation Bias could already exist at the stage of writing a prompt, so of course, the response from the AI tool would attempt to provide the information requested.

A recent announcement from ChatGPT led me to explore a different possibility. The new capability involved the opportunity to turn history on and off. I am still not certain of the intent of this control, but it got me thinking about the filter bubble issue I had tried to investigate in the early 2010s.

Before trying my original internet history strategy, I thought I would just ask ChatGPT about potential bias. If you just read the following, you may not note what I did, but it was clear that the AI tool knew quite a bit about me and used this information in responding.

Here is the reply to my prompt:

Great question, Mark.

When history is turned on, ChatGPT uses your prior conversations and context across sessions to tailor responses more personally. When history is off, each prompt is treated in isolation, with no memory of past conversations or preferences. Here’s how that difference plays out:

? With History ON

  • Personalization: ChatGPT may remember your preferences (e.g., your interest in educational psychology, your writing style, your teaching context) and adapt replies accordingly.
  • Continuity: It can refer to earlier parts of a conversation or even previous sessions if memory is active (e.g., remembering that you’re revising a specific slide deck).
  • Efficiency: You don’t have to reintroduce background details repeatedly; the system can “pick up where you left off.”
  • Persistent memory (if enabled): You can allow ChatGPT to remember things long-term, like your goals or recurring projects.

? With History OFF

  • No personalization: ChatGPT won’t recall your identity, preferences, or past queries.
  • Each prompt is standalone: Even within a single session, it treats each message without prior context unless you include it explicitly.
  • Privacy-first: The content of your chats won’t be stored or used to improve models.
  • More generic answers: Responses will tend to be more neutral or general-purpose since there’s no way to reference or build on prior knowledge.

? Same Prompt, Different Responses

Let’s say you prompt:

“Give me three ways to explain working memory to pre-service teachers.”

  • With history on, ChatGPT might recall that you’re preparing a university-level course for future educators and tailor examples accordingly (perhaps referencing your slide deck or prior discussions of models like Baddeley & Hitch).
  • With history off, it might assume a general audience and provide broader educational examples (e.g., working memory in classroom tasks like copying from the board or doing mental arithmetic).

Aside from the information itself, note several indicators it was tailoring the response to me. It identified my vocation as an educational psychologist. It used examples that would be helpful to someone who wanted to understand a general topic based on their background. So, it used educational examples.

Back to the “Apple” test.

I wanted to generate my own test of biased feedback and tried a version of the approach I had tried with search results. I tried to think of a topic I approach in a unique way and that might bias an AI response based on my personal history.

I am interested in the use of digital technology that seems different from more popular representation. For example, should class notes be taken by hand or on a laptop? Should people read from paper or a tablet screen? I think I can make a case for a digital approach because it is easier to store and utilize by-products of experiences over extended period of time. So digital notes and highlights can be exported and retained in ways that are easier to search, interrelate, update, etc.. Yet, this is not position taken in public sources offering advice to students, teachers, and decision makers.

My history of prompting AI to offer content are biased by my way of viewing such issues. For example, I might use the following prompts:

I am interested in comparing note-taking by hand and on a digital device. What are the advantages of digital notes when those notes are intended for use over an extended period of time

I am interested in comparing note-taking by hand and on a digital device. What are the advantages of digital notes when the goal is to discover connections among ideas recorded on different systems

Now for my test. What if I create a more neutral prompt and submit this prompt with my history turned on and off?

Summarize the research comparing note-taking by hand and on a laptop. What should a college instructor recommend?

With memory on

Research comparing handwritten and laptop notetaking focuses on how each method affects learning outcomes such as comprehension, retention, and conceptual understanding. Below is a summary of the key findings and instructional implications.

Key Research Findings

1. Encoding vs. External Storage

  • Notetaking supports learning through encoding (processing while writing) and external storage (referring back to notes).
  • The method of note-taking influences encoding quality more than external storage.

2. Handwriting Promotes Deeper Processing

  • Mueller & Oppenheimer (2014): Students taking notes by hand performed better on conceptual questions than those using laptops.
  • Explanation: Typing leads to verbatim transcription, while handwriting encourages summarizing and processing.
  • Follow-up studies confirmed that handwritten notes promote deeper understanding, especially for material requiring synthesis or inference.

3. Laptop Notetaking and Verbatim Transcription

  • Laptop users tend to transcribe lectures verbatim, which is a shallower form of processing.
  • Even when told to avoid verbatim notes, laptop users often struggle to summarize effectively.

4. Distraction and Multitasking

  • Laptops introduce distractions: Web browsing, messaging, and multitasking impair learning for both the note-taker and nearby students.
  • Sana, Weston, & Cepeda (2013): Nearby students also performed worse when seated near multitasking laptop users.

5. Volume vs. Depth

  • Laptops allow more content to be recorded, but more content doesn’t always equal better learning.
  • Handwritten notes tend to be shorter, more selective, and encourage rephrasing, which benefits comprehension.

6. Review and Revision

  • Laptop notes are easier to organize, search, and revise, which can aid long-term study.
  • Handwritten notes may encourage memory consolidation, especially when rewritten or summarized later.

Recommendations for College Instructors

1. Promote Purposeful Notetaking

  • Encourage students to summarize and paraphrase regardless of method.
  • Teach effective notetaking strategies such as the Cornell method or concept mapping.

2. Consider a Handwriting Default

  • Suggest handwriting for lectures emphasizing conceptual understanding or discussion.
  • Provide explicit rationale (e.g., “This will help you understand and remember better”).

3. Allow Laptops with Conditions

  • If laptops are allowed, encourage use of apps that limit distractions.
  • Consider designating laptop zones to reduce spillover distraction.

4. Support Note Review

  • Promote post-lecture activities: reviewing notes, rewriting, or using notes to answer practice questions.
  • Share outlines or slides to reduce pressure to capture everything.

5. Adapt to Context

  • For highly detailed or technical lectures (e.g., coding or statistics), typing may be more efficient.
  • Consider hybrid approaches (e.g., typed notes with handwritten annotations).

Final Takeaway

Recommendation: Encourage handwritten note-taking when feasible, especially for classes involving complex reasoning or discussion. If laptops are used, guide to avoid verbatim transcription and minimize distractions. Whatever the method, emphasize that note-taking is a learning activity — not just a record-keeping task.

Would you like citations or slides to support this summary in a teaching context?

— — — — — —

With memory off

Summarize the research comparing note-taking by hand and on a laptop. What should a college instructor recommend?

Sure! Here’s a concise summary of the research comparing hand note-taking and laptop note-taking, plus a recommendation for college instructors:

Research Summary:

  • Handwritten notes tend to enhance learning and retention better than typing notes on laptops.
  • Writing by hand encourages processing and summarizing information, leading to deeper understanding.
  • Laptop note-taking often leads to verbatim transcription, which can reduce comprehension and critical thinking.
  • Some studies find that laptop users are more easily distracted by unrelated activities (social media, browsing).
  • However, laptops allow for faster note-taking and easier organization or review, which can benefit some students, especially those with certain disabilities or faster speaking lectures.

Recommendation for College Instructors:

  • Encourage students to take handwritten notes, especially for conceptual understanding and long-term retention.
  • If laptops are allowed, instructors should coach students to avoid verbatim typing and focus on summarizing.
  • Consider a hybrid approach: use laptops for occasional digital resources, but emphasize active listening and synthesis in notes.
  • For accessibility, allow laptop use when needed but promote best practices to maximize learning.

Would you like me to draft a brief instructor note or syllabus statement based on this?

Comparison Summary

You can examine these two responses and reach your own conclusions. My interpretation is very possibly biased by my effort to demonstrate an impact of history. My analysis follows.

  1. Without access to personal history, the prompt response is significantly shorter. This was predicted when I asked ChatGPT directly about what to expect with history on and off.
  2. I would say the response with personal history on still favors nondigital notes, but there is more recognition of differences and some concessions to situations in which digital content may provide an advantage.

So what?

I think it possible to argue both approaches demonstrate a type of bias. Perhaps there is value in understanding this and recognizing that the individual prompting the AI must still be sensitive to personal bias. Simple prompts may also elicit simple responses, and more sophisticated issues end up being ignored.

Loading

Learning Now That We Have AI

I have been exploring and writing about AI now since the release of ChatGPT. I think about AI both as a user and as an educator and believe I have enough experience and a perspective from which to offer an overview. The tools have improved to a level at which I find them quite valuable as a research and writing assistant. My main approach is to use a tool that allows me to interact with my own digital notes collected over maybe a decade to assist in many of my writing projects. I count myself among those who advocate for the professional value of AI.

My thoughts on AI in education are more complex and it is this perspective I will try to share here. Let me start with what I think is an essential assumption and that is students will find ways to use AI. Whatever the perspective of individual educators is, I argue it is necessary to begin accepting this assumption. Trying to think across many different content areas and skills, it seems reasonable to stipulate that there are certain skills that must be practiced directly to develop (e.g., writing, reasoning, problem solving) and elements of information that may not be life changing whether retained in an individual’s memory or not, but that the general benefit of existing knowledge which is about stored information and the connections that exist within this information offer advantages in understanding and reasoning. Any given fact can certainly now be searched when needed, but this option does not account for the general benefits of what I would describe as general knowledge. We accumulate general knowledge by interacting with our world, but the purposeful accumulation of important information is more efficient through the process we commonly call education. Let me add one more assumption to this position statement. We cannot learn for others nor can we make them learn. We can at best provide access to information and provide external tasks that have the potential to influence the processes of learning. Ultimate responsibility must be placed on individual learners and this is often a requirement at a time when individual learners lack the background and perhaps cannot make decisions understanding how learning works and how skills and knowledge may influence their futures. 

Here is my thinking about AI. Educators have a limited amount of time during which they can directly influence learners. They must depend on the cooperation of learners and perhaps their parents when attempting to influence learning during other times. I would describe this reality as important in making decisions about how this time of maximum influence will be spent. For example, I write a lot about study behavior. Educators and sometimes do use class time for studying. When they make this commitment, they are also reducing the time available for other experiences – presentation of information, experiences such as science labs, peer interactions such as guided discussions and debate. Some reactions to AI suggest that class time be used to some extent to control the use of AI. For example, writing a theme during class rather than at home or study hall or completing math problems during class rather than as homework. If the limitation of AI is determined to be significant enough, this can be done, but this will then replace other activities.

So, I believe that the development of some skills and a general knowledge base cannot be eliminated because of AI and this development can only be guaranteed during the time during which an entire class would have to be prevented from using AI. To be clear, I am not advocating for this option. I am trying to identify the benefits and costs of options which I believe cannot be individualized; e.g., educators cannot really differentiate what is required of different individuals in a classroom situation.

Much of what we are playing with involves decisions about when to attempt to exert control over personal goals and motivations. I was a university prof and there is a common approach at this level that differentiates the requirements for a major from general education requirements. We don’t allow students to decide if they want to develop basic writing skills because we require a couple of semesters of composition. We expect a basic level of function in mathematics, but allow individuals to make decisions as to whether the basic course will be what amounts to a high school level course in algebra or the Introduction to The Calculus. 

How strongly do we as educators believe we should ignore personal goals and motivation? This is a question for us and for other stakeholders in the educational process. We certainly cannot control learners, but we can arrange evaluation processes to recognize when some mandatory proficiency has not been achieved. Politicians and the general public already tend to blame educators when basic proficiencies do not match those existing in other countries or when graduates seem unprepared for vocations or for civic responsibilities. What consequences do those who are critical suggest for educators or what are they willing to tolerate for the learners who are ultimately responsible? 

When I write about this topic it becomes clear to me that the issues I address are very complex and perhaps that is a useful message for others who have simplistic positions on the process of education or the issues educators face. I am a big fan of research informing practice. One challenge with the type of issues described here is that most involve cumulative effects over an extended period of time. Longitudinal studies may eventually provide useful insights, but the downsides could impact an entire generation before the research makes this outcome clear.

Is a summary possible?

I am willing to say that AI offers great benefits to supplement human actions. We all should be prepared to take advantage and guided experience in developing AI-related knowledge and skills should now be a component of what we teach. 

Reliance on AI in place of tasks that develop skills is detrimental. You cannot learn to write if AI replaces your attempts to write. You cannot develop critical thinking or reasoning skills if you do not struggle with tasks that require these skills. The issue then is whether the skills are important to the individual and when is the optimal time to make this decision. Perhaps even this is too narrow of a perspective. What are the commitments each of us owes to each other when it comes to basic knowledge and skills?

If forced to take a position, I would suggest that individuals be required to learn and be able to develop knowledge and skills AI-unaided and be able to demonstrate they can apply AI in ways appropriate to the tasks they presently must accomplish. The notion of tool or augmentation seems useful here and it would seem curriculum developers could differentiate cognitive skills from tool proficiency accordingly. 

Note: I find that as I write about this topic I encounter the complexities that I think are important to consider. I certainly welcome comments that address these complexities and possibly provide me when ideas I can address in response.

Loading