Tag: Digital

  • Most students, educators use AI–but opinions differ on ethical use

    Most students, educators use AI–but opinions differ on ethical use

    Key points:

    As generative AI continues to gain momentum in education each year, both its adoption and the attitudes toward its use have steadily grown more positive, according to a new report from Quizlet.

    The How America Learns report explores U.S. student, teacher, and parent perspectives on AI implementation, digital learning and engagement, and success beyond the classroom.

    “At Quizlet, we’ve spent nearly two decades putting students at the center of everything we do,” said Quizlet CEO Kurt Beidler. “We fielded this research to better understand the evolving study habits of today’s students and ensure we’re building tools that not only help our tens of millions of monthly learners succeed, but also reflect what they truly need from their learning experience.”

    AI becomes ubiquitous in education
    As generative AI solutions gain traction in education year over year, adoption and attitudes towards the technology have increased and improved. Quizlet’s survey found that 85 percent of respondents–including high school and college teachers, as well as students aged 14-22–said they used AI technology, a significant increase from 66 percent in 2024. Of those respondents using AI, teachers now outpace students in AI adoption (87 percent vs 84 percent), compared to 2024 findings when students slightly outpaced teachers.

    Among the 89 percent of all students who say they use AI technology for school (up from 77 percent in 2024), the top three use cases are summarizing or synthesizing information (56%), research (46 percent), and generating study guides or materials (45 percent). The top uses of AI technology among teachers remained the same but saw significant growth YoY: research (54 percent vs. 33 percent), summarizing or synthesizing information (48 percent vs. 30 percent), and generating classroom materials like tests and assignments (45 percent vs. 31 percent).

    While the emergence of AI has presented new challenges related to academic integrity, 40 percent of respondents believe that AI is used ethically and effectively in the classroom. However, students are significantly less likely to feel this way (29 percent) compared to parents (46 percent) and teachers (57 percent), signaling a continued need for education and guidelines on responsible use of AI technology for learning.

    “Like any new technology, AI brings incredible opportunities, but also a responsibility to use it thoughtfully,” said Maureen Lamb, AI Task Force Chair and Language Department Chair at Miss Porter’s School. “As adoption in education grows, we need clear guidelines that help mitigate risk and unlock the full potential of AI.  Everyone–students, educators, and parents–has a role to play in understanding not just how to use AI, but when and why it should be used.”

    Digital learning demands growth while equity gap persists
    Just as AI is becoming a staple in education, survey results also found that digital learning is growing in popularity, with 64 percent of respondents expressing that digital learning methods should be equal or greater than traditional education methods, especially teachers (71 percent).

    Respondents indicated that flexibility (56 percent), personalized learning (53 percent), and accessibility (49 percent) were the most beneficial aspects of digital learning. And with 77 percent of students making sacrifices, including loss of sleep, personal time, and missed extracurriculars due to homework, digital learning offers a promising path toward a more accommodating approach. 

    While the majority of respondents agreed on the importance and benefits of digital learning, results also pointed to a disparity in access to these tools. Despite nearly half (49 percent) of respondents agreeing that all students in their community have equal access to learning materials, technology, and support to succeed academically, that percentage drops to 43 percent for respondents with diagnosed or self-identified learning differences, neurodivergent traits, or accessibility needs.

    Maximizing success for academic and real-world learning
    While discussion around AI and education has largely focused on use cases for academic learning, the report also uncovered an opportunity for greater support to help drive success beyond the classroom and provide needed resources for real-world learning.

    Nearly 60 percent of respondents believe a four-year college degree is of high importance for achieving professional success (58 percent). However, more than one-third of students, teachers, and parents surveyed believe schools are not adequately preparing students for success beyond the classroom.

    “As we drive the next era of AI-powered learning, it’s our mission to give every student and lifelong learner the tools and confidence to succeed, no matter their motivation or what they’re striving to achieve,” said Beidler. “As we’ve seen in the data, there’s immense opportunity when it comes to career-connected learning, from life skills development to improving job readiness, that goes well beyond the classroom and addresses what we’re hearing from students and teachers alike.”

    The top five skills respondents indicated should be prioritized more in schools are critical thinking and problem solving (66 percent), financial literacy (64 percent), mental health management (58 percent), leadership skills (52 percent), and creativity and innovation (50 percent).

    This press release originally appeared online.

    eSchool News Staff
    Latest posts by eSchool News Staff (see all)

    Source link

  • 4 ways AI is empowering the next generation of great teachers

    4 ways AI is empowering the next generation of great teachers

    Key points:

    In education, we often talk about “meeting the moment.” Our current moment presents us with both a challenge and an opportunity: How can we best prepare and support our teachers as they navigate increasingly complex classrooms while also dealing with unprecedented burnout and shortages within the profession?

    One answer could lie in the thoughtful integration of artificial intelligence to help share feedback with educators during training. Timely, actionable feedback can support teacher development and self-efficacy, which is an educator’s belief that they will make a positive impact on student learning. Research shows that self-efficacy, in turn, reduces burnout, increases job satisfaction, and supports student achievement. 

    As someone who has spent nearly two decades supporting new teachers, I’ve witnessed firsthand how practical feedback delivered quickly and efficiently can transform teaching practice, improve self-efficacy, and support teacher retention and student learning.

    AI gives us the chance to deliver this feedback faster and at scale.

    A crisis demanding new solutions

    Teacher shortages continue to reach critical levels across the country, with burnout cited as a primary factor. A recent University of Missouri study found that 78 percent of public school teachers have considered quitting their profession since the pandemic. 

    Many educators feel overwhelmed and under-supported, particularly in their formative years. This crisis demands innovative solutions that address both the quality and sustainability of teaching careers.

    What’s often missing in teacher development and training programs is the same element that drives improvement in other high-performance fields: immediate, data-driven feedback. While surgeons review recordings of procedures and athletes get to analyze game footage, teachers often receive subjective observations weeks after teaching a lesson, if they receive feedback at all. Giving teachers the ability to efficiently reflect on AI-generated feedback–instead of examining hours of footage–will save time and potentially help reduce burnout.

    The transformative potential of AI-enhanced feedback

    Recently, Relay Graduate School of Education completed a pilot program with TeachFX using AI-powered feedback tools that showed remarkable promise for our teacher prep work. Our cohort of first- and second-year teachers more than doubled student response opportunities, improved their use of wait time, and asked more open-ended questions. Relay also gained access to objective data on student and teacher talk time, which enhanced our faculty’s coaching sessions.

    Program participants described the experience as “transformative,” and most importantly, they found the tools both accessible and effective.

    Here are four ways AI can support teacher preparation through effective feedback:

    1. Improving student engagement through real-time feedback

    Research reveals that teachers typically dominate classroom discourse, speaking for 70-80 percent of class time. This imbalance leaves little room for student voices and engagement. AI tools can track metrics such as student-versus-teacher talk time in real time, helping educators identify patterns and adjust their instruction to create more interactive, student-centered classrooms.

    One participant in the TeachFX pilot said, “I was surprised to learn that I engage my students more than I thought. The data helped me build on what was working and identify opportunities for deeper student discourse.”

    2. Freeing up faculty to focus on high-impact coaching

    AI can generate detailed transcripts and visualize classroom interactions, allowing teachers to reflect independently on their practice. This continuous feedback loop accelerates growth without adding to workloads.

    For faculty, the impact is equally powerful. In our recent pilot with TeachFX, grading time on formative observation assignments dropped by 60 percent, saving up to 30 hours per term. This reclaimed time was redirected to what matters most: meaningful mentoring and modeling of best practices with aspiring teachers.

    With AI handling routine analysis, faculty could consider full class sessions rather than brief segments, identifying strategic moments throughout lessons for targeted coaching. 

    The human touch remains essential, but AI amplifies its reach and impact.

    3. Scaling high-quality feedback across programs

    What began as a small experiment has grown to include nearly 800 aspiring teachers. This scalability can more quickly reduce equity issues in teacher preparation.

    Whether a teaching candidate is placed in a rural school or urban district, AI can ensure consistent access to meaningful, personalized feedback. This scalable approach helps reduce the geographic disparities that often plague teacher development programs.

    Although AI output must be checked so that any potential biases that come through from the underlying datasets can be removed, AI tools also show promise for reducing bias when used thoughtfully. For example, AI can provide concrete analysis of classroom dynamics based on observable actions such as talk time, wait time, and types of questions asked. While human review and interpretation remains essential–to spot check for AI hallucinations or other inaccuracies and interpret patterns in context–purpose-built tools with appropriate guardrails can help deliver more equitable support.

    4. Helping teachers recognize and build on their strengths

    Harvard researchers found that while AI tools excel at using supportive language to appreciate classroom projects–and recognize the work that goes into each project–students who self-reported high levels of stress or low levels of enjoyment said the feedback was often unhelpful or insensitive. We must be thoughtful and intentional about the AI-powered feedback we share with students.

    AI can also help teachers see what they themselves are doing well, which is something many educators struggle with. This strength-based approach builds confidence and resilience. As one TeachFX pilot participant noted, “I was surprised at the focus on my strengths as well and how to improve on them. I think it did a good job of getting good details on my conversation and the intent behind it. ”

    I often tell new teachers: “You’ll never see me teach a perfect lesson because perfect lessons don’t exist. I strive to improve each time I teach, and those incremental gains add up for students.” AI helps teachers embrace this growth mindset by making improvement tangible and achievable.

    The moment is now

    The current teacher shortage is a crisis, but it’s also an opportunity to reimagine how we support teachers.

    Every student deserves a teacher who knows how to meaningfully engage them. And every teacher deserves timely, actionable feedback.  The moment to shape AI’s role in teacher preparation is now. Let’s leverage these tools to help develop confident, effective teachers who will inspire the next generation of learners.

    Latest posts by eSchool Media Contributors (see all)

    Source link

  • Weaving digital citizenship into edtech innovation

    Weaving digital citizenship into edtech innovation

    Key points:

    What happens when over 100 passionate educators converge in Chicago to celebrate two decades of educational innovation? A few weeks ago, I had the thrilling opportunity to immerse myself in the 20th anniversary of the Discovery Educator Network (the DEN), a week-long journey that reignited my passion for transforming classrooms.

    From sunrise to past sunset, my days at Loyola University were a whirlwind of learning, laughter, and relentless exploration. Living the dorm life, forging new connections, and rekindling old friendships, we collectively dove deep into the future of learning, creating experiences that went far beyond the typical professional development.

    As an inaugural DEN member, the professional learning community supported by Discovery Education, I was incredibly excited to return 20 years after its founding to guide a small group of educators through the bountiful innovations of the DEN Summer Institute (DENSI). Think scavenger hunts, enlightening workshops, and collaborative creations–every moment was packed with cutting-edge ideas and practical strategies for weaving technology seamlessly into our teaching, ensuring our students are truly future-ready.

    During my time at DENSI, I learned a lot of new tips and tricks that I will pass on to the educators I collaborate with. From AI’s potential to the various new ways to work together online, participants in this unique event learned a number of ways to weave digital citizenship into edtech innovation. I’ve narrowed them down to five core concepts; each a powerful step toward building future-ready classrooms and fostering truly responsible digital citizens.

    Use of artificial intelligence

    Technology integration: When modeling responsible AI use, key technology tools could include generative platforms like Gemini, NotebookLM, Magic School AI, and Brisk, acting as ‘thought partners’ for brainstorming, summarizing, and drafting. Integration also covers AI grammar/spell-checkers, data visualization tools, and feedback tools for refining writing, presenting information, and self-assessment, enhancing digital content interaction and production.

    Learning & application: Teaching students to ethically use AI is key. This involves modeling critical evaluation of AI content for bias and inaccuracies. For instance, providing students with an AI summary of a historical event to fact-check with credible sources. Students learn to apply AI as a thought partner, boosting creativity and collaboration, not replacing their own thinking. Fact-checking and integrating their unique voices are essential. An English class could use AI to brainstorm plot ideas, but students develop characters and write the narrative. Application includes using AI for writing refinement and data exploration, fostering understanding of AI’s academic capabilities and limitations.

    Connection to digital citizenship: This example predominantly connects to digital citizenship. Teaching responsible AI use promotes intellectual honesty and information literacy. Students can grasp ethical considerations like plagiarism and proper attribution. The “red, yellow, green” stoplight method provides a framework for AI use, teaching students when to use AI as a collaborator, editor, or thought partner–or not at all.This approach cultivates critical thinking and empowers students to navigate the digital landscape with integrity, preparing them as responsible digital citizens understanding AI’s implications.

    Digital communication

    Technology integration: Creating digital communication norms should focus on clarity with visuals like infographics, screenshots, and video clips. Canva is a key tool for a visual “Digital Communication Agreement” defining online interaction expectations. Include student voice by the integration and use of pictures and graphics to illustrate behaviors and potentially collaborative presentation / polling tools for student involvement in norm-setting.

    Learning & application: Establishing clear online interaction norms is the focus of digital communication. Applying clear principles teaches the importance of visuals and setting communication goals. Creating a visual “Digital Communication Agreement” with Canva is a practical application where students define respectful online language and netiquette. An elementary class might design a virtual classroom rules poster, showing chat emojis and explaining “think before you post.” Using screenshots and “SMART goals” for online discussions reinforces learning, teaching constructive feedback and respectful debate. In a middle school science discussion board, the teacher could model a respectful response like “I understand your point, but I’m wondering if…” This helps students apply effective digital communication principles.

    Connection to digital citizenship: This example fosters respectful communication, empathy, and understanding of online social norms. By creating and adhering to a “Digital Communication Agreement,” students develop responsibility for online interactions. Emphasizing respectful language and netiquette cultivates empathy and awareness of their words’ impact. This prepares them as considerate digital citizens, contributing positively to inclusive online communities.

    Content curation

    Technology integration: For understanding digital footprints, one primary tool is Google Drive when used as a digital folder to curate students’ content. The “Tech Toolbox” concept implies interaction with various digital platforms where online presence exists. Use of many tools to curate content allows students to leave traces on a range of technologies forming their collective digital footprint.

    Learning & application: This centers on educating students about their online presence’s permanence and nature. Teaching them to curate digital content in a structured way, like using a Google Drive folder, is key. A student could create a “Digital Portfolio” in Google Drive with online projects, proud social media posts, and reflections on their public identity. By collecting and reviewing online artifacts, students visualize their current “digital footprint.” The classroom “listening tour” encourages critical self-reflection, prompting students to think about why they share online and how to be intentional about their online identity. This might involve students reviewing anonymized social media profiles, discussing the impression given to future employers.

    Connection to digital citizenship: This example cultivates awareness of online permanence, privacy, responsible self-presentation, and reputation management. Understanding lasting digital traces empowers students to make informed decisions. The reflection process encourages the consideration of their footprint’s impact, fostering ownership and accountability for online behavior. This helps them become mindful, capable digital citizens.

    Promoting media literacy

    Technology integration: One way to promote media literacy is by using “Paperslides” for engaging content creation, leveraging cameras and simple video recording. This concept gained popularity at the beginning of the DEN through Dr. Lodge McCammon. Dr. Lodge’s popular 1-Take Paperslide Video strategy is to “hit record, present your material, then hit stop, and your product is done” style of video creation is something that anyone can start using tomorrow. Integration uses real-life examples (likely digital media) to share a variety of topics for any audience. Additionally, to apply “Pay Full Attention” in a digital context implies online viewing platforms and communication tools for modeling digital eye contact and verbal cues.

    Learning & application: Integrating critical media consumption with engaging content creation is the focus. Students learn to leverage “Paperslides” or another video creation method to explain topics or present research, moving beyond passive consumption. For a history project, students could create “Paperslides” explaining World War II causes, sourcing information and depicting events. Learning involves using real-life examples to discern credible online sources, understanding misinformation and bias. A lesson might show a satirical news article, guiding students to verify sources and claims through their storyboard portion. Applying “Pay Full Attention” teaches active, critical viewing, minimizing distractions. During a class viewing of an educational video, students could pause to discuss presenter credentials or unsupported claims, mimicking active listening. This fosters practical media literacy in creating and consuming digital content.

    Connection to digital citizenship: This example enhances media literacy, critical online information evaluation, and understanding persuasive techniques. Learning to create and critically consume content makes students informed, responsible digital participants. They identify and question sources, essential for navigating a digital information-saturated world. This empowers them as discerning digital citizens, contributing thoughtfully to online content.

    Collaborative problem-solving

    Technology integration: For practicing digital empathy and support, key tools are collaborative online documents like Google Docs and Google Slides. Integration extends to online discussion forums (Google Classroom, Flip) for empathetic dialogue, and project management tools (Trello, Asana) for transparent organization. 

    Learning & application: This focuses on developing effective collaborative skills and empathetic communication in digital spaces. Students learn to work together on shared documents, applying a “Co-Teacher or Model Lessons” approach where they “co-teach” each other new tools or concepts. In a group science experiment, students might use a shared Google Doc to plan methodology, with one “co-teaching” data table insertion from Google Sheets. They practice constructive feedback and model active listening in digital settings, using chat for clarification or emojis for feelings. The “red, yellow, green” policy provides a clear framework for online group work, teaching when to seek help, proceed cautiously, or move forward confidently. For a research project, “red” means needing a group huddle, “yellow” is proceeding with caution, and “green” is ready for review.

    Connection to digital citizenship: This example is central to digital citizenship, developing empathy, respectful collaboration, and responsible problem-solving in digital environments. Structured online group work teaches how to navigate disagreements and offers supportive feedback. Emphasis on active listening and empathetic responses helps internalize civility, preparing students as considerate digital citizens contributing positively to online communities.

    These examples offer a powerful roadmap for cultivating essential digital citizenship skills and preparing all learners to be future-ready. The collective impact of thoughtfully utilizing these or similar approaches , or even grab and go resources from programs such as Discovery Education’s Digital Citizenship Initiative, can provide the foundation for a strong academic and empathetic school year, empowering educators and students alike to navigate the digital world with confidence, integrity, and a deep understanding of their role as responsible digital citizens.

    In addition, this event reminded me of the power of professional learning communities.  Every educator needs and deserves a supportive community that will share ideas, push their thinking, and support their professional development. One of my long-standing communities is the Discovery Educator Network (which is currently accepting applications for membership). 

    Latest posts by eSchool Media Contributors (see all)

    Source link

  • Integrating AI into education is not as daunting as it seems

    Integrating AI into education is not as daunting as it seems

    Key points:

    Forty-some years ago, students sat in straight rows with books, papers, and pencils neatly lined up on their desks. But beginning in the 1990s, educators faced very different classrooms as computers found their way into schools.

    For most teachers, it felt daunting to figure out how to integrate new tools into curriculum requirements–and how to find the time to make it happen. To help this digital transformation then, I joined the South Dakota Department of Education to lead summer immersion teacher training on technology integration, traveling the state to help schools understand how to use new tools like video systems. I was one of many who helped educators overcome that initial learning curve–and now tools like computers are an integral part of the education system.

    Let’s face it: The advent of new technologies can be overwhelming. Adjusting to them takes time. Now, with the coming of age of AI, teachers, administrators, students, and parents have endless questions and ideas on how it might positively or negatively influence education. I’ve seen it in my current role, in which I continue to empower educators and states to use modern technology to support student learning. And while concerns about AI are valid, there are many positive potential outcomes. For educators in particular, AI can be a huge value-add, automating certain administrative tasks, helping understand and predict student success and struggles, and even helping tailor instruction for individual students.

    The upside is huge. As schools embark on their AI journeys, it’s important to remember that we’ve been here before–from the introduction of the internet in classrooms to the abrupt shift to e-learning at the outset of COVID-19. Superintendents, boards of education, and other education leaders can draw on important lessons from prior technological transformations to fully take advantage of this one.

    Here are some rules of the road for navigating the integration of disruptive technologies:

    1. Choose the right tools. The AI tool(s) you choose can have varying results. School districts should prioritize proven technologies with a track record in education. For students, this includes adaptive learning platforms or virtual tutors. Some of the best tools are those that are specifically designed by and for educators to expedite administrative tasks such as grading and lesson planning. Even more valuable is the ability to support education-specific issues such as identifying struggling students with early warning systems and using AI to provide projections for student futures.

      2. Training is everything. With proper training, AI can be less intimidating. We don’t expect students to understand a new concept by reading a few paragraphs in a textbook, and we shouldn’t expect teachers to figure out how to best use AI on their own. President Trump’s recent executive order prioritizes the use of AI in discretionary grant programs for teacher training, which is an important step in the right direction.

      3. Engage parents. Moms and dads may be concerned if they hear–without a deeper explanation–that a school board is rolling out an AI tool to help with teaching or administrative tasks in their children’s education. Keep an open line of communication with the guardians of students about how and why AI is being used. Point parents to resources to help them improve their own AI literacy. To a reasonable degree, invite feedback. This two-way communication helps build trust, allay fears and clarify any misconceptions, to the benefit of everyone involved, including, most importantly, the students.

      4. Humans must be involved. The stakes are high. AI is not perfect. Administrators must ensure they and the educators using AI tools are double checking the work. In the parlance of responsible AI, this is known as having a “human in the loop,” and it’s especially important when the outcomes involve children’s futures. This important backstop instills confidence in the parents, students and educators.

      5. Regularly evaluate if the tools are living up to expectations. The point of integrating AI into teachers’ and administrators’ workstreams is to lighten their load so they can spend more time and energy on students. Over time, AI models can decay and bias can be introduced, reducing the effectiveness of AI outputs. So, regular monitoring and evaluating is important. Educators and administrators should regularly check in to determine if the integration of AI is supporting their goals.

      6. The learning curve may create more work at first–but the payoff is exponential. Early adoption is important. I worked with school districts that pushed off integrating digital technologies–ultimately, it put the educators behind their peers. AI can make a difference in educators’ lives by freeing them up from administrative burdens to focus on what really matters–the students.

      This is the start of a journey–one that I believe is truly exciting! It’s not the first nor the last time educators adopt new technologies. Don’t let AI overwhelm or distract you from tried-and-true integration techniques. Yes, the technology is different–but educators are always adapting, and it will be the same with AI, to the benefit of educators and students.

      Latest posts by eSchool Media Contributors (see all)

    Source link

  • Beyond Digital Literacy: Cultivating “Meta AI” Skills in Students and Faculty – Faculty Focus

    Beyond Digital Literacy: Cultivating “Meta AI” Skills in Students and Faculty – Faculty Focus

    Source link

  • What we lose when AI replaces teachers

    What we lose when AI replaces teachers

    Key points:

    A colleague of ours recently attended an AI training where the opening slide featured a list of all the ways AI can revolutionize our classrooms. Grading was listed at the top. Sure, AI can grade papers in mere seconds, but should it?

    As one of our students, Jane, stated: “It has a rubric and can quantify it. It has benchmarks. But that is not what actually goes into writing.” Our students recognize that AI cannot replace the empathy and deep understanding that recognizes the growth, effort, and development of their voice. What concerns us most about grading our students’ written work with AI is the transformation of their audience from human to robot.

    If we teach our students throughout their writing lives that what the grading robot says matters most, then we are teaching them that their audience doesn’t matter. As Wyatt, another student, put it: “If you can use AI to grade me, I can use AI to write.” NCTE, in its position statements for Generative AI, reminds us that writing is a human act, not a mechanical one. Reducing it to automated scores undermines its value and teaches students, like Wyatt and Jane, that the only time we write is for a grade. That is a future of teaching writing we hope to never see.

    We need to pause when tech companies tout AI as the grader of student writing. This isn’t a question of capability. AI can score essays. It can be calibrated to rubrics. It can, as Jane

    said, provide students with encouragement and feedback specific to their developing skills. And we have no doubt it has the potential to make a teacher’s grading life easier. But just because we can outsource some educational functions to technology doesn’t mean we should.

    It is bad enough how many students already see their teacher as their only audience. Or worse, when students are writing for teachers who see their written work strictly through the lens of a rubric, their audience is limited to the rubric. Even those options are better than writing for a bot. Instead, let’s question how often our students write to a broader audience of their peers, parents, community, or a panel of judges for a writing contest. We need to reengage with writing as a process and implement AI as a guide or aide rather than a judge with the last word on an essay score.

    Our best foot forward is to put AI in its place. The use of AI in the writing process is better served in the developing stages of writing. AI is excellent as a guide for brainstorming. It can help in a variety of ways when a student is struggling and looking for five alternatives to their current ending or an idea for a metaphor. And if you or your students like AI’s grading feature, they can paste their work into a bot for feedback prior to handing it in as a final draft.

    We need to recognize that there are grave consequences if we let a bot do all the grading. As teachers, we should recognize bot grading for what it is: automated education. We can and should leave the promises of hundreds of essays graded in an hour for the standardized test providers. Our classrooms are alive with people who have stories to tell, arguments to make, and research to conduct. We see our students beyond the raw data of their work. We recognize that the poem our student has written for their sick grandparent might be a little flawed, but it matters a whole lot to the person writing it and to the person they are writing it for. We see the excitement or determination in our students’ eyes when they’ve chosen a research topic that is important to them. They want their cause to be known and understood by others, not processed and graded by a bot.

    The adoption of AI into education should be conducted with caution. Many educators are experimenting with using AI tools in thoughtful and student-centered ways. In a recent article, David Cutler describes his experience using an AI-assisted platform to provide feedback on his students’ essays. While Cutler found the tool surprisingly accurate and helpful, the true value lies in the feedback being used as part of the revision process. As this article reinforces, the role of a teacher is not just to grade, but to support and guide learning. When used intentionally (and we emphasize, as in-process feedback) AI can enhance that learning, but the final word, and the relationship behind it, must still come from a human being.

    When we hand over grading to AI, we risk handing over something much bigger–our students’ belief that their words matter and deserve an audience. Our students don’t write to impress a rubric, they write to be heard. And when we replace the reader with a robot, we risk teaching our students that their voices only matter to the machine. We need to let AI support the writing process, not define the product. Let it offer ideas, not deliver grades. When we use it at the right moments and for the right reasons, it can make us better teachers and help our students grow. But let’s never confuse efficiency with empathy. Or algorithms with understanding.

    Latest posts by eSchool Media Contributors (see all)

    Source link

  • Data, privacy, and cybersecurity in schools: A 2025 wake-up call

    Data, privacy, and cybersecurity in schools: A 2025 wake-up call

    Key points:

    In 2025, schools are sitting on more data than ever before. Student records, attendance, health information, behavioral logs, and digital footprints generated by edtech tools have turned K-12 institutions into data-rich environments. As artificial intelligence becomes a central part of the learning experience, these data streams are being processed in increasingly complex ways. But with this complexity comes a critical question: Are schools doing enough to protect that data?

    The answer, in many cases, is no.

    The rise of shadow AI

    According to CoSN’s May 2025 State of EdTech District Leadership report, a significant portion of districts, specifically 43 percent, lack formal policies or guidance for AI use. While 80 percent of districts have generative AI initiatives underway, this policy gap is a major concern. At the same time, Common Sense Media’s Teens, Trust and Technology in the Age of AI highlights that many teens have been misled by fake content and struggle to discern truth from misinformation, underscoring the broad adoption and potential risks of generative AI.

    This lack of visibility and control has led to the rise of what many experts call “shadow AI”: unapproved apps and browser extensions that process student inputs, store them indefinitely, or reuse them to train commercial models. These tools are often free, widely adopted, and nearly invisible to IT teams. Shadow AI expands the district’s digital footprint in ways that often escape policy enforcement, opening the door to data leakage and compliance violations. CoSN’s 2025 report specifically notes that “free tools that are downloaded in an ad hoc manner put district data at risk.”

    Data protection: The first pillar under pressure

    The U.S. Department of Education’s AI Toolkit for Schools urges districts to treat student data with the same care as medical or financial records. However, many AI tools used in classrooms today are not inherently FERPA-compliant and do not always disclose where or how student data is stored. Teachers experimenting with AI-generated lesson plans or feedback may unknowingly input student work into platforms that retain or share that data. In the absence of vendor transparency, there is no way to verify how long data is stored, whether it is shared with third parties, or how it might be reused. FERPA requires that if third-party vendors handle student data on behalf of the institution, they must comply with FERPA. This includes ensuring data is not used for unintended purposes or retained for AI training.

    Some tools, marketed as “free classroom assistants,” require login credentials tied to student emails or learning platforms. This creates additional risks if authentication mechanisms are not protected or monitored. Even widely-used generative tools may include language in their privacy policies allowing them to use uploaded content for system training or performance optimization.

     

    Data processing and the consent gap

    Generative AI models are trained on large datasets, and many free tools continue learning from user prompts. If a student pastes an essay or a teacher includes student identifiers in a prompt, that information could enter a commercial model’s training loop. This creates a scenario where data is being processed without explicit consent, potentially in violation of COPPA (Children’s Online Privacy Protection Act) and FERPA. While the FTC’s December 2023 update to the COPPA Rule did not codify school consent provisions, existing guidance still allows schools to consent to technology use on behalf of parents in educational contexts. However, the onus remains on schools to understand and manage these consent implications, especially with the rule’s new amendments becoming effective June 21, 2025, which strengthen protections and require separate parental consent for third-party disclosures for targeted advertising.

    Moreover, many educators and students are unaware of what constitutes “personally identifiable information” (PII) in these contexts. A name combined with a school ID number, disability status, or even a writing sample could easily identify a student, especially in small districts. Without proper training, well-intentioned AI use can cross legal lines unknowingly.

    Cybersecurity risks multiply

    AI tools have also increased the attack surface of K-12 networks. According to ThreatDown’s 2024 State of Ransomware in Education report, ransomware attacks on K-12 schools increased by 92 percent between 2022 and 2023, with 98 total attacks in 2023. This trend is projected to continue as cybercriminals use AI to create more targeted phishing campaigns and detect system vulnerabilities faster. AI-assisted attacks can mimic human language and tone, making them harder to detect. Some attackers now use large language models to craft personalized emails that appear to come from school administrators.

    Many schools lack endpoint protection for student devices, and third-party integrations often bypass internal firewalls. Free AI browser extensions may collect keystrokes or enable unauthorized access to browser sessions. The more tools that are introduced without IT oversight, the harder it becomes to isolate and contain incidents when they occur. CoSN’s 2025 report indicates that 60 percent of edtech leaders are “very concerned about AI-enabled cyberattacks,” yet 61 percent still rely on general funds for cybersecurity efforts, not dedicated funding.

    Building a responsible framework

    To mitigate these risks, school leaders need to:

    • Audit tool usage using platforms like Lightspeed Digital Insight to identify AI tools being accessed without approval. Districts should maintain a living inventory of all digital tools. Lightspeed Digital Insight, for example, is vetted by 1EdTech for data privacy.
    • Develop and publish AI use policies that clarify acceptable practices, define data handling expectations, and outline consequences for misuse. Policies should distinguish between tools approved for instructional use and those requiring further evaluation.
    • Train educators and students to understand how AI tools collect and process data, how to interpret AI outputs critically, and how to avoid inputting sensitive information. AI literacy should be embedded in digital citizenship curricula, with resources available from organizations like Common Sense Media and aiEDU.
    • Vet all third-party apps through standards like the 1EdTech TrustEd Apps program. Contracts should specify data deletion timelines and limit secondary data use. The TrustEd Apps program has vetted over 12,000 products, providing a valuable resource for districts.
    • Simulate phishing attacks and test breach response protocols regularly. Cybersecurity training should be required for staff, and recovery plans must be reviewed annually.

    Trust starts with transparency

    In the rush to embrace AI, schools must not lose sight of their responsibility to protect students’ data and privacy. Transparency with parents, clarity for educators, and secure digital infrastructure are not optional. They are the baseline for trust in the age of algorithmic learning.

    AI can support personalized learning, but only if we put safety and privacy first. The time to act is now. Districts that move early to build policies, offer training, and coordinate oversight will be better prepared to lead AI adoption with confidence and care.

    Latest posts by eSchool Media Contributors (see all)

    Source link

  • Digital learning in a new age

    Digital learning in a new age

    Key points:

    Digital learning–in the form of online, hybrid, and blended schools and courses–is growing steadily in U.S. schools. These learning options can transform education because they allow for learning, teaching, and student engagement outside the confines of traditional physical schools.

    Students no longer have to show up at a school building every morning, and millions of students and families are demonstrating their preference for more flexible learning options by choosing their district’s online schools, charter schools, and private schools.

    Digital learning meets the needs of today’s students, who are seeking flexibility in their scheduling. Many high school students want to pursue sports, arts, and career interests in the form os jobs, internships, and other program. Others simply crave the control an innovative school gives them over the time, place, and pace at which they learn. Digital learning also meets the needs of teachers, who, just like knowledge workers around the world, are interested in employment that allows them to choose their schedules.

    Online and hybrid learning is becoming easier to implement as technology grows and improves. Unlike just a few years ago, when teachers were concerned about using multiple technology tools, much-improved integration and interoperability between platforms is making adoption of multiple tools far easier.

    While relatively few students and families prefer their education to be 100 percent online, many students are selecting hybrid options that combine online and face-to-face interactions. Much like young knowledge workers who are increasingly blending home offices with corporate headquarters, digital learning is showing up in unexpected places as well. Let’s take a closer look at two examples: career and technical education (CTE) and physical education (PE).

    CTE is often perceived as being “hands on” in ways that casual observers might expect would not align well with digital learning–but the truth is exactly the opposite.

    Digital learning is broadening the world of CTE for students. Online and hybrid schools provide CTE programs by offering a combination of online career courses and by partnering with businesses, state and regional training centers, and other
    organizations to combine online learning with on-the-ground, real-world jobs, internships, and learning opportunities.

    Hybrid schools and programs, including those run by mainstream districts, provide academic scheduling flexibility to students who seek to prioritize their time in jobs, internships, or career training. No longer do these students have to fit in their career interests after regular school hours or on weekends–when many companies and high-value jobs are not open or available.

    For example, a student interested in a veterinary career can work at a vet’s office during the regular week and school hours, completing some of their online coursework after normal work hours.

    Virtual Arkansas, a state-supported course provider supporting districts across Arkansas, has made digital CTE a central element of its offerings.

    “CTE is a key part of our value to students and schools across Arkansas. Students, teachers, counselors, and the business community, all appreciate that we are providing flexible options for students to gain real-world expertise and experience via our online and hybrid programs,” said John Ashworth, the programs’ executive director.

    Perhaps even more surprising than CTE shifting to digital is the idea that next generation physical education is based on online tools, adept teachers, and student voice and choice.

    Today’s students are accustomed to going into a coffee shop and ordering their drink with a dozen customized features. And yet, in traditional PE classes, we expect students to all want to learn the same sport, activity, or exercise, at the same time and pace. That’s how too many traditional gym classes operate–based on the factory model of education in which all students do the same thing at the same time.

    There’s a better way, which is being embraced by online schools, hybrid schools, and traditional districts. Online and hybrid PE classes shift exercise, activity, and wellness to match student interests and timing. A student chooses from hundreds of detailed instructional videos in dozens of categories, from aquatics to basketball to yoga, trains using the videos combined with instruction provided by a teacher, and tracks her progress.

    This doesn’t sound like a traditional gym class; instead, it mimics the ways that young adults are active in gyms, yoga studios, and sports leagues all around the country. Consider fitness clubs from the local YMCA to the most high-end club–they are all offering a wide variety of classes, on varied schedules to fit busy lifestyles, and at different levels of expertise. No school can match this, of course, by the traditional approach to gym class. But Joe Titus, founder and CEO of Hiveclass, which offers online physical education courses, points out that student agency to
    choose from a wide variety of PE options is possible–when schools are ready to make the leap.

    Online schools and district programs are already doing so, with fantastic outcomes as students lean into their choices and options. As futurist William Gibson said decades ago, “the future is here, it’s just not evenly distributed.”

    Online and hybrid CTE, physical education, and other options prove the point. The next step is to make these options widely available to all the students who are seeking a better alternative.

    Latest posts by eSchool Media Contributors (see all)

    Source link

  • Your Digital and Analog Life with Jo Davis

    Your Digital and Analog Life with Jo Davis

    As a writer, Jo Davis is used to sharing through her online presence. In this episode of The Social Academic, we talk about her life online such as her digital portfolio. And, offline through the coloring books she designed, the Starseed Panic Pages, and journaling. What does it mean to be intentional about your digital and analog life as an academic? We talk about focus and what it can do for your brain to be on paper.

    I’ve admired Jo Davis’ writing for years. I followed her on X after reading one of her movie critiques. When she shared a recent podcast appearance on the Moments that Define Us, I thought she was perfect to come on The Social Academic to talk about her life online and on paper. And, what it means to be her authentic self.

    P.S. Jo’s workshop, Writing Through The Noise: Letters to the Inner Child is coming up. There’s 2 session to choose from: August 8 and 9. You may find it helpful, here’s info about her workshops for you.

    Jo Davis is a professor, author, freelance writer, film critic, artist, and a beacon of creativity. She teaches writing and rhetoric at the University of Denver.

    Here’s her episode of the Moments That Define Us podcast, “You need to find calm in the noise”.

    Source link

  • Digital learning is different

    Digital learning is different

    Key points:

    In the animated film Up, the character Dug is a talking dog with an interesting mannerism. Each time he sees a movement off to the side, he stops whatever he is doing, stares off in that direction and shouts, “Squirrel!” I feel that this is a perfect representation of how schools often deal with new and emerging technologies. They can be working hard to provide the best instruction for their students but become immediately distracted anytime a new technology is introduced.

    From the internet and computers to cell phones and artificial intelligence, schools continue to invest a lot of time and money into figuring out how best to use these new technologies. Overall, schools have done a good job adapting to the numerous digital tools introduced in classrooms and offices–and often, these tools are introduced as standalone initiatives. Why do school districts feel the need to ‘reinvent the wheel’ every time a new technology is released? Instead of looking at each new technology as a tool that must be integrated in the curriculum, why not determine what is missing from current instruction and identify what prevents integration from occurring naturally?

    Schools need to recognize that it is not just learning how to use these new digital tools that is important. They must learn how to interpret and use the incredible variety of resources that accompany these tools–resources that provide perspectives that students would never have access to when using physical resources.

    Digital is different

    For centuries, learning material has come from a variety of physical resources. These include human-made items (i.e. textbooks, documents, paintings, audio recordings, and movies) as well as one of the most commonly used physical resources: teachers. In traditional instruction, teachers spend a great deal of class time teaching students information from these physical resources. But the physical nature of these resources limits their availability to students. To ensure that students have long-term access to the information provided by these physical resources, most traditional instruction emphasizes memorization, summarizing, and note taking. 

    With digital resources, students can access information at any time from anywhere, which means learning how to retain information is less important than learning how to effectively find credible information. The authenticity of the information is important because the same tools that are used to access digital resources can just as easily be used to create new digital resources. This means there is a lot of misinformation available online, often consisting of nothing more than personal opinions. Students need to not only be able to search for information online, but they also need to be able to verify the authenticity of online information. The ability to identify misleading or false information is a skill that will benefit them in their personal and academic lives.

    Learning

    While it is fairly easy to find information online, especially with the inclusion of AI in search engines, there are some search techniques that will reduce the amount of misinformation found in simple search requests. By teaching students how to refine their searches and discussing the impact of these search skills, students will be more discerning when it comes to reviewing search results. They need to be aware that the most helpful sites do not always appear at the top of the search list. Some sites are sponsored and thus automatically placed at the beginning of the search list. Other sites will tweak their web search parameters to ensure a higher priority in the search list.  A better understanding of how online searching works will result in more effective searches. 

    Once information is found, the authenticity of the resource and the information itself needs to be established. Fortunately, there are standard practices that can be utilized to teach verification. In the early 2000’s, a popular checklist method called CRAAP (Currency [timeliness], Relevance, Authority, Accuracy, Purpose) emerged. While this method was effective in evaluating the authenticity of the website, it did not ensure the accuracy of the information on the website. In 2019, the SIFT (Stop, Investigate, Find confirming resources, and Trace claims) methodology was introduced.  This methodology focuses on determining if online content is credible. These are not the only tools available to teachers. Librarians and media specialists are a good place to start when determining age-appropriate lessons and material to teach verification.

    Students need to have access to some high-quality digital resources starting in elementary school. Teaching website verification at an early age will help students understand, from the beginning, that there is a lot of misinformation available online. At the same time, schools need to ensure that they provide access to digital resources that are age appropriate. Today’s network technology provides many ways for schools to monitor and control what information or sites are available to students at different grade levels. While these network tools are effective, they should be used in conjunction with well-trained teachers who understand how to safely navigate digital resources and students who are expected to practice responsible internet behavior. Introducing a select number of digital resources in elementary classes is the first step toward creating discerning researchers who will gain the ability to effectively judge a website’s appropriateness and usefulness.

    Teaching

    In order to create opportunities for students to experience learning with digital resources, instructional practices need to be less reliant on teacher-directed instruction. The use of physical resources requires the teacher to be the primary distributor of the information. Typically, this is done through lecture or whole-class presentations. With digital resources, students have direct access to the information, so whole-class distribution is not necessary. Instead, instructional practices need to provide lessons that emphasize finding and verifying information, which can be done by shifting to a learner-centered instructional model. In a learner-centered lesson, the onus falls on the student to determine what information is needed, and if the found information is credible for a given task. The class time that previously would have been spent on lecture becomes time for students to practice finding and authenticating online information. Initially, these learning experiences would be designed as guided practice for finding specific information. As students become more proficient with their search skills, the lesson can shift toward project-based lessons.

    Project-based lessons will help students learn how to apply the information they find, as well as determine what unknown information they need to complete the work. Unlike lesson design for practicing information searching and verification, project-based lessons provide opportunities for students to decide what information is needed and how best to use it. Instead of directing the student’s information-gathering, the teacher provides guidance to ensure they are accessing information that will allow the students to complete the project.

    This shift in instruction does not necessarily mean there will be a significant curricular change. The curricular content will remain the same, but the resources could be different. Because students control what resources they use, it is possible that they could find resources different from the ones specified in the curriculum. Teachers will need to be aware of the resources students are using and may have to spend time checking the credibility of the resource. Given the varying formats (text, audio, video, graphic) available with digital resources, students will be able to determine which format(s) best supports their learning style. Because most digital tools utilize the same digital resources and formats, teaching students how to learn with digital resources will prepare them for adapting to the next new digital tool. It is simply a matter of learning how to use the tool–after all, they already know how to use the resource.

    When creating units of study, teachers should consider the type of resources students will be using. To simplify matters, some units should be designed to utilize digital resources only and include lessons that teach students how to find and verify information. Students still need to develop skills to work with physical resources as well. It may be helpful to start off with units that utilize only physical or digital resources. That way teachers can focus on the specific skills needed for each type of resource. As students gain proficiency with these skills, they will learn to use the appropriate skills for the given resources.

    The amount of information available to the public today is staggering. Unfortunately, too much of it is unverified and even purposely misleading. Trying to stop misinformation from being created and distributed is not realistic. But teaching students how to validate online information can make the distribution of and exposure to misinformation much less impactful. The open nature of the internet allows for many divergent opinions and perspectives. We need to ensure that when students graduate, they have the skills necessary to determine the authenticity of online information and to be able to determine its merit.

    Teaching and learning with digital resources is different, and traditional instruction does not meet the learning needs of today’s students. Giving students the opportunity to master learning with digital resources will prepare them for the next technology “squirrel” and will enable them to determine how best to use it on their own.

    Latest posts by eSchool Media Contributors (see all)

    Source link