These institutions are the backbone of American higher education. They serve the largest share of students by far, and state-supported colleges and universities play an outsize role in providing economic mobility for Americans of all backgrounds. I’ve spent my entire career working on behalf of public universities, most recently as president of the Association of Public and Land-grant Universities. I know the enormous good they do for their students and for society at large. We have the best publicly supported system of higher education in the world. We can and must continue to improve it.
I also understand why our public institutions will benefit from an accreditor that aligns with their mission and their public obligations. They need an accreditor that offers true peer review and a disciplined focus on improving student outcomes. They need an accreditor familiar with the mechanics of state oversight, able to promote academic quality while also being more efficient by eliminating redundant bureaucracy in the accreditation process.
The Commission for Public Higher Education was formed earlier this year to answer those needs. Established by a consortium of six public university systems—the State University System of Florida, the University System of Georgia, the University of North Carolina System, the University of South Carolina System, the University of Tennessee System and the Texas A&M University System—the aim of CPHE is to offer public universities across the country an alternative to the regional accreditors that have long dominated higher education, each claiming a geographical monopoly that lumped together for-profit schools, bespoke private colleges and open-access public institutions under the same set of rules and regulations.
I agreed to serve as chair of the Board of Directors for CPHE because I believe there’s a need for innovation in accreditation. We are seizing the opportunity to improve institutional accreditation by focusing on outcomes, as well as streamlining the process by taking advantage of the considerable oversight that public institutions are subject to at the state level. An accreditor purpose-built by public institutions, for public institutions, can promote academic quality while driving innovation in student success and eliminating unnecessary costs in the legacy model of accreditation.
There is clearly enthusiasm for the vision behind CPHE. Ten diverse institutions have already signed on to join CPHE’s initial cohort (full list below), and the commission is fielding additional inquiries from across the country. We’ve just issued a call for public university faculty and administrators to join our first group of peer-review teams, and we look forward to pioneering a new model of more straightforward and more transparent accreditation review.
CPHE Initial Cohort
Appalachian State University
Chipola College
Columbus State University
Florida Atlantic University
Florida Polytechnic University
North Carolina Central University
Texas A&M–Kingsville
Texas A&M–Texarkana
University of North Carolina at Charlotte
University of South Georgia
University leaders and state policymakers nationwide see the value in a streamlined approach to accreditation that shifts the focus from inputs and operational minutiae to meaningful outcomes for students and taxpayers.
The legacy approach to accreditation is plagued by the need for each accreditor to serve the huge diversity of institutional missions and governing structures that underlie the American system of higher education. Trying to impose the same set of criteria and procedures on every institution, from small private colleges to huge public flagships, has led to decades of ineffective oversight and wasted effort. There is little or no evidence that institutional accreditation has driven quality improvements across the sector, while it is abundantly clear that it has imposed arbitrary and opaque regulatory demands on institutions that already are subjected to multiple layers of oversight as public agencies.
Institutions like Georgia State University, where I served more than a decade as president, are closely scrutinized by their governing boards, by state regulators and legislative bodies, by auditors and bond ratings agencies. They have public disclosure and consumer protection requirements above and beyond what is demanded of private and for-profit colleges. I have firsthand experience with how costly and cumbersome accreditation reviews divert institutional resources that would be better spent supporting student success, and I am confident a public-focused accreditor can streamline reporting and compliance costs without compromising oversight.
An accreditor attuned to the nuances of public oversight can add value by focusing on academic quality and student success, using a process of peer review to promote continuous improvement through the dissemination of best practices and innovations. That’s why CPHE’s accreditation standards are tailored toward public purpose and academic excellence, with provisions for measuring student learning, promoting academic freedom and intellectual diversity, and driving continuous improvement of student outcomes.
At core, the purpose of accreditation is to reassure students and taxpayers that universities are delivering on their promise to provide a quality education that leaves students better off. An accreditor tightly focused on that public mission can go a long way in shoring up the trust that higher education needs to thrive.
Mark Becker is the chair of the Board of Directors of the Commission for Public Higher Education. He formerly served as president of the Association of Public and Land-grant Universities from 2022 to 2025, and before that he was president of Georgia State University from 2009 to 2021.
Since 2022, there’s been a surge in the number and types of applications using generative AI, but not all tools are the same. So how can faculty, staff and students learn to identify the differences and determine when it’s appropriate to leverage these tools?
Colby College developed a platform, called Mule Chat, that allows users to explore several large language models, including ChatGPT, Gemini, Claude and LLaMA. The platform provides a safe on-ramp into generative AI usage and relies on student tutors to disseminate information to peers.
In the latest episode of Voices of Student Success, host Ashley Mowreader speaks with David Watts, the director of Colby College’s Davis Institute for Artificial Intelligence, and Michael Yankoski, Davis AI research and teaching scientist, to learn about the college’s AI institute and how Mule Chat works.
An edited version of the podcast appears below.
Inside Higher Ed: Can we start the conversation by talking a bit about what AI at Colby College looks like? What is the landscape you’re working with and how are you thinking about AI when it comes to teaching and learning?
David Watts: I am new to Davis AI, as we call it at Colby, but the [Davis AI] Institute has actually been around since before ChatGPT, so Colby kind of had a pioneering approach.
David Watts, director of the Davis Institute for AI at Colby College
Colby is a small liberal arts college, and they had the vision that this was going to be around for a while. And rather than, as most institutions were doing, sort of keep it at bay or ban it from campus, Colby dove in and wanted to engage with it and understand how it is going to impact education.
I spent most of my career in industry, mostly in research and development, and so I when I wanted to make the jump over to academia, I wasn’t expecting to find that small liberal arts colleges had done this, and when I saw what Colby had done, I was really drawn to it and came over. So I’ve really loved what has been going on and what continues to go on at Colby with the Davis Institute for Artificial Intelligence.
Inside Higher Ed: Michael, your role puts you directly in connection with faculty when it comes to integrating AI into their classrooms or into their programs. Can you talk about what that looks like and how maybe that looks different at a liberal arts institution?
Michael Yankoski, research and teaching scientist, Davis Institute for AI at Colby College
Michael Yankoski: One of the most amazing aspects of the Davis Institute for Artificial Intelligence here at a place like Colby is the liberal arts approach that the institution as a whole is able to engage with.
That means that we’re able to facilitate conversations from a multiplicity of different disciplines and bring faculty together from different approaches across the divisions in the college—from the STEM fields to the humanities to the social sciences. And have really productive, very generative conversations around ways to engage with artificial intelligence and the shared learning and shared knowledge of people who have been really pioneering in the area. To able to say, “How can I integrate generative artificial intelligence with my pedagogy? How can I help think with students about how to engage these technologies in a way that is beneficial for their education, help empower students in their education and then on the research side?”
Many faculty with whom we work at the Davis Institute are exploring ways to integrate artificial intelligence in their research program, and to say, “Is there a way that artificial intelligence can help me accelerate my research or take my research in new directions?” The opportunity to bring people together to discuss that and to facilitate those conversations across the disciplines is one of the best aspects of the liberal arts approach to artificial intelligence.
Inside Higher Ed: Does Colby have an institutional policy for AI use, or what appropriate AI use looks like?
Watts: It’s a moving target. Anyone who tells you they have it all figured out is probably embellishing. It is a moving target, but one of the things we did was make sure we engage faculty, and in fact, we started with faculty, then we engaged administrators, we engaged students and we engaged general counsel, and evaluated what the challenges are, what the downsides are. And we made sure that we built what we call guidelines rather than policy.
The guidelines talk through the dos and don’ts but also leave enough flexibility for our faculty to think through how they want to engage with AI, especially since AI is a moving target, too. As we grow and learn with our faculty, we adapt and adjust our guidelines and so they’re out there for everyone to see, and we will continue to evolve them as we move forward.
Inside Higher Ed: Can you introduce our listeners to Mule Chat? What is it and how does it work on campus?
Watts: Michael has been here and was one of the originators of creating Mule Chat on campus. And so he can tell you a lot of the details and how it’s been working.
But what I loved about what Michael and the team did, and it was a collaborative effort, was to create, I’ll call it an on-ramp. We were working towards moving the needle from banning AI, as one extreme, to engaging with AI and creating a tool that allowed faculty, students and staff to all easily engage with multiple tools through Mule Chat.
It lowered the activation barrier to entry to AI and allowed us to have an on-ramp for people to come in and start seeing what the possibilities are, and it has worked brilliantly.
Yankoski: The idea behind Mule Chat originally was to provide a place for students, faculty and staff to begin to get experience with and understanding around generative AI. To provide a space where folks could come and understand a bit more about, what are these tools? How do they work? What are they capable of? What are some of the areas we need to be aware of, the risks and the best practices, and how can we provide this on-ramp, as David described, for people to be able to engage with generative artificial intelligence?
This is about student success, empowering students to understand what these technologies are, what they’re good at, what they’re not good at. And then also, one of the key principles here was equity of access. We wanted to ensure that anybody on Colby’s campus, regardless of whether they could afford one of the premium subscription services, was able to get access to these frontier models and to understand how to then do the prompt engineering work, and to then compare the kinds of outputs and capabilities of some of the frontier models. And so really, the core sort of genesis and driving desire for the creation of Mule Chat was to provide this on-ramp that would empower student success, allow equity of access, and also would provide a safe and secure place for people to be able to engage these technologies and to learn.
Inside Higher Ed: Can you describe the functionality of Mule Chat? For someone who has never experimented with LLMs, what does it look like or feel like to engage with Mule Chat?
Watts: You touched on something really great there, because that was part of the idea. We introduced multiple models into Mule Chat so that people could compare and get an idea of what it’s capable of and what it’s not capable of.
I’ll give an example of a faculty member who we are working with right now who started with Mule Chat, engaged with it in their preparation—this is a professor of East Asian studies—how they prepare their classes, realized what the capabilities were, started doing more with it, with their students. The students then brought interesting ideas about what else we can do and pushed beyond even the limits of Mule Chat. And then Davis AI can go help them bring in, for example, they were looking at—not only just looking at old archives and using that in their teaching of East Asian studies, but also bringing in video capability, for example, and in fact, even creating new videos or some of the research that they’re doing now, bringing in more capabilities above and beyond Mule Chat. So it is exactly what Michael was saying, an on-ramp that then opens up the possibilities of what we can do with AI in higher education.
Yankoski: I think the real value of the Mule Chat interface is that it allows people to compare the different models.
Folks can use prompt engineering to compare the outputs of one model and then put that alongside the outputs of another model and be able to observe the way that different models might reason or might do their inference in different kinds of ways.
That side-by-side comparison is a really powerful opportunity for people to engage with the different models and to experience the different kinds of outputs that they create. To build on what David was saying, the ability to then put other tools [like videos] inside of the Mule Chat platform, that allows for deeper research into particular areas. For example, we have a tool that we built, which is called Echo Bot.
The Colby student newspaper is called the Colby Echo, so we’ve been able to bring all the archives of the Echo into a tool that allows students and faculty researchers to engage with those archives and chat with the entire archive of the Colby Echo. We’ve been working closely—and this goes back to the liberal arts approach—with different faculty across campus, as well as the college libraries, to bring this tool online and make it available within the Mule Chat system.
Inside Higher Ed: Let me know if you can build me an IHE bot, because I can never find anything in our archives. I could really benefit from something.
Watts: We can brainstorm on that.
Inside Higher Ed: Great, we’ll talk about licensing later.
I wanted to ask, it seems there’s a new AI tool that pops every other day. So when you’re talking about comparing different tools and thinking about what might be most relevant for students, how often are you scouting out the landscape to understand what’s out there and relevant?
Watts: That’s a great question, and actually extremely important that we do that.
Not only are we reaching out and finding, reading, learning, attending conferences, helping to create conferences ourselves that bring in people and experts who are different perspectives, but we also then have lots of people on campus who have their own ideas. People come to us regularly with, “Oh, look at this cool tool. We should use it for this thing on campus.”
And that’s when we use that for educating people about some of the potential pitfalls that we have to watch out for, talking about guardrails and when you’re bringing in new capability, just like you had to think about when you’re bringing in new software. But I think it’s even more imperative that we’re very careful about what AI tools we bring into campus. You’re absolutely right that there are tons of them that all have different capabilities. But one of the things we try to teach is that there’s a full spectrum: the great, the good, the bad and the ugly. You have to think about that entire spectrum. And that’s one of the beauties of what I loved about coming to a liberal arts college was that you have multiple perspectives, and coming from all forms of disciplines in the humanities, the arts, the natural sciences, the social sciences, and all are engaged and can be engaged across AI.
Yankoski: I think that’s what’s so unique and really powerful about the Davis Institute for Artificial Intelligence approach. When we work with faculty and students and really, if some faculty member or student has an idea that they want to explore, we have structures that allow for technology grants, for faculty to be able to come and to propose the use of a new tool, or to advance their teaching or to advance their research.
Then that’s a great opportunity to engage with that faculty member and perhaps their research assistants, and work with those students and that faculty member to explore the possibility of using that tool. Each faculty member knows their domain so much better than we do. As the core Davis AI team, we’re able to work with that faculty and those students to better understand the use case, better understand the tools that they want to engage, and then work with them to consult and to create a pathway forward. That’s an incredible opportunity as well for the students to understand, how do we think about the security of the data? How do we think about the processing pipeline? How do we think about the best practices with regards to utilizing artificial intelligence in this particular domain?
Really that’s about student empowerment and student success as they get ready to transition out of college into an economy where increasingly expectations around knowledge and the ability to utilize and to vet artificial intelligence are only going to increase.
Inside Higher Ed: How are students engaged in this work?
Yankoski: One the most intriguing aspects of Mule Chat has been that students have been really leading in teaching and empowering other students to utilize the tool and to understand the quantum engineering aspects and to understand the different models.
The student leaders have been working with Mule Chat and then actually teaching other students, teaching faculty and helping lead the sessions, as well as working on their own projects within Google Chat. So it’s been a really strong and quite incredible platform for student engagement and student empowerment as students learn from one another and then are able to learn how to teach about these tools to their peers.
Watts: That’s absolutely a huge part of what we did, and I mentioned that, even though students come first, we started working to move the needle with faculty first on purpose, with students in mind. And then we branched out into, now we can engage the students. Once you have enough buy-in from faculty, start engaging the students, and we’ve been doing a lot of that.
Then what’s beautiful, the magic happens when the students start coming up with thoughts and ideas that grow in ways that faculty haven’t thought of. Because remember that a lot of this is new to faculty as well.
So we actually then will identify key students that we have been working with and actually hire them on board as Davis AI research associates that then help us continue to move the needle, because there’s nothing better for students than to hear from other students about what’s possible. And the same goes for faculty, by the way. So, you know, Michael was mentioning a little bit about our strategy with faculty and how we engage them. But a part of what we do is faculty sessions. We give them creative names like “Bagels and bots,” and we include food and then we have those sessions where faculty talk to faculty. We do the same with the students, so students can talk to students. And it’s just wonderful to see the magic that happens when that begins to grow organically.
Inside Higher Ed: What has the reception been to Mule Chat?
Watts: Most people were skeptical [of AI] early on; most were in the mode of “push it away.” I think that drove some interesting behaviors in faculty and students.
So a big part of what we’ve been trying to do is essentially drive towards AI literacy for all. And when I say all, it’s an interdisciplinary approach. We’re looking across the entire campus, and so all students in all departments are what we’re driving towards. Now, you correctly point out that there will always be skeptics. I will strive for 100 percent, but if we asymptotically approach that into the future, I’ll live with that.
The goal is to prepare students, and that’s who we need to make sure that we’re preparing for the life they’re going to go into that’s been transformed by AI, that touches everybody. One of the cool things is we’re giving out grants to faculty to engage with AI and come up with ideas, and we’re doing that on multiple levels, and those faculty are now coming from all. We have art professors. We have writing professors. We have East Asian studies. We have professors from government, we have all of them engaging and so we’ve been able to, therefore, move the needle quite a bit so that a lot more people are a lot more receptive and open to it on campus, which is great.
Inside Higher Ed: You mentioned that Colby has a faculty-led approach, but sometimes that means that students from specific majors or disciplines might be less exposed to AI than others, depending on who their faculty are. It seems like you all are taking a balanced approach, not only encouraging enthusiastic AI entrepreneurs but also working with the skeptics.
Watts: It’s absolutely critical that we work on both ends of that spectrum, if that makes sense. We’re driving great innovation, and there’s great examples of research right here on campus that are doing wonderful things in an interdisciplinary way.
We just won an NSF grant for ARIA, an NSF institute looking at AI assistance in mental health, because that’s one of the most challenging spaces for how the models interact with people with mental and behavioral health challenges. It’s a perfect example of our interdisciplinary approach, with a professor from psychology working with a professor from computer science to go tackle these challenging areas. And I think that’s one of the things that Colby has done well, is to take that broader, interdisciplinary approach. Many people say that word now, but I think the liberal arts are primed for leading the charge on what that’s going to look like, because AI, by its nature, is interdisciplinary.
Inside Higher Ed: What’s next on campus? Is there any area that you’re all exploring or looking to do some more research in, or new tools and initiatives that our listeners should know about for the future?
Watts: We’re consistently evaluating that and bringing them in. What we’re trying to do is let it grow based on need as people explore and come up with ideas.
I mentioned the video; we’re now enabling video capability so we can do some of that research. It also opens up more multimodal approaches.
One of the approaches to the ARIA research, for example, is we want to be able to detect and therefore build context-aware assistance to have better results for everyone. So if we can solve the mental and behavioral health challenges, it’s probably one of the most difficult ones. It can also solve some of the other areas of underrepresented people who are left out or underrepresented groups who are left out of training, for example, which can lead to challenging behaviors.
I’m really excited about all of those possibilities and the areas that allow us to enable. We talked about access, we can also talk about accessibility.
We have on campus the Colby College Museum of Art; one of the faculty in computer science is exploring accessibility options using AI with a robotic seeing-eye dog. If someone wanted to visit the museum who was blind or visually impaired, they could interact with a seeing-eye dog that they’re used to, but this seeing-eye dog now might have more capability to communicate with people about what they’re seeing and in a museum setting, for example.
So really excited about that type of research: how do we really benefit humanity with these types of tools.
Inside Higher Ed: One thing I wanted to ask about is resources allocated from the university to be able to access all these tools. What investment is the college making to ensure that students are able to stay on the cutting edge of AI initiatives?
Watts: That’s absolutely critical. We want to make it no cost to our students and accessible to our students, but it still costs. So [it’s vital to] make sure that we have funding.
We were very lucky that we got a Davis endowment that enabled us to build the Davis Institute. That was huge because, and you can think about some of the challenges with federal funding and all of that stuff, but to have an endowment that allowed us to draw on that and really build strong capabilities at Colby College was critical. But you’re touching on the fact that we’re going to need to continue to do that. And that’s where, for example, the NSF grant and other grants that we will continue to explore will help us with how we continue to grow our impact and grow our value as we head into the future.
Rating colleges against each other is a tricky enterprise on a good day. For community colleges it’s particularly vexed, given how intensely local they are and the simple fact that most of them don’t compete with each other. If, say, a community college in Illinois or Arizona does something terrific, I don’t feel threatened by it; our students in Pennsylvania aren’t going to move there in large numbers based on ratings.
Still, the lure of lists is powerful. The new Carnegie classifications, as outlined by Inside Higher Ed, rate community colleges largely by the subsequent earnings of their students compared to local labor markets. The article outlines one key objection based on economic geography: In some parts of the country, the median wages and cost of living are so high that even students coming out of very successful vocational programs will struggle economically at first.
It’s similar to the objection I noted a few years ago to the “social mobility” ratings that Washington Monthly offered, in which colleges were graded based on how many quartiles of income their students jumped. To score really well on that metric, you’d better have most of your students start in the bottom quartile. A college located in an area with more students in the second quartile simply couldn’t compete, no matter how well it did its job.
The measurement error in this case is more well-meaning than in many others, but it’s still an error. And I’m still unconvinced that it adequately captures the value of students who transfer, whether with a degree or just with a bunch of classes.
Presumably, those objections could be incorporated into a more refined effort. But even the objections implicitly concede that the only relevant scale on which to measure education is income. Postcollege income matters, of course, but it’s not the only thing that matters. If it were, we would stop training early-childhood teachers and social workers immediately.
Part of the attraction of measuring income is that it’s quantifiable. Even I sometimes get twitchy when academics refer to the “ineffable” benefits of something; it can be hard to disentangle idealism from wishful thinking. But some noneconomic benefits of higher education are relatively easy to quantify in the short term.
What if we measured colleges on the voting rates of new graduates?
Voting is quantifiable, at least for now. It’s a basic form of adult civic engagement. It doesn’t rely on economic cycles that can wreak havoc with starting salaries. And we know from decades of political science that on average, people who vote are more knowledgeable about politics and social issues than people who don’t. (Contrary to popular myth, people who consistently vote a party line are more informed on average than ticket-splitters, but that’s another article.)
Voting rates also wouldn’t be distorted by the low earnings of students who transferred and are in their junior or senior years of college when surveyed. Yes, voting rates are higher in presidential years, but presidential years happen at the same time for every college in the country, so they wouldn’t affect comparisons between institutions.
If we took postgraduation voting rates seriously, colleges would be incentivized to improve the civic literacy and involvement of their students. That strikes me as an excellent outcome. That’s especially true for community colleges, given that their student bodies are much more representative of America than the elite selective universities.
Community colleges train, yes, but they also educate. Why not educate for democracy? And why not support—with funding and publicity—the colleges that do a particularly good job of that?
Donald Trump’s second term in office continues to confound onlookers. Yet a growing number of universities around the world are offering courses for students to try to make sense of the mercurial president.
Universities outside the U.S. are also involved. First run in 2017, Trumpism: An American Biography is an optional module for second-year history students at the University of Sheffield, which explores how U.S. history can shed light on the present.
Andrew Heath, lecturer in U.S. history at Sheffield, told Times Higher Education that part of the module’s purpose was to get students thinking about the history of terms such as populism that are “often thrown around in the media to make sense of Trump and Trumpism,” and to encourage them to think critically about the way that comparisons are invoked.
But teaching about such a fast-moving political situation is not easy. “It’s a module that always poses challenges—readings can quickly feel dated; teaching it in an election year last time around was harder. Every iteration of the unit needs significant updating,” added Heath.
Christopher Breem, managing director of the McCourtney Institute for Democracy at Pennsylvania State University, said it is always hard to teach about something going on in the present. But this is often what students are most interested in because they recognize that it is important to them and their future to understand it, he said.
“I think if you are up front with students that there are unavoidable risks associated with teaching any subject in real time, they accept that.”
During the first Trump presidency, some academics came under an intense national spotlight for their courses that explicitly referenced him. One professor who previously taught a course mentioning Trump said the whole experience was “unpleasant,” with staff and the university receiving numerous phone calls and emails.
“The university took my information off the website, and we had a police officer outside of the classroom,” the professor said. “I turned on my house alarm during the day. Frightened, I turned down opportunities for press interviews.”
The academic, who wished to remain anonymous, said it was hard to keep up with the constant change and disruption of the Trump administration but that students were very engaged.
“We are not advocates who use our classes to tell students what action they should take. We are teachers and scholars who inform our students, give them the skills to think in a rigorous, disciplined way, and with integrity. They then decide how to use their skills.”
Since 2022, there’s been a surge in the number and types of applications using generative AI, but not all tools are the same. So how can faculty, staff and students learn to identify the differences and determine when it’s appropriate to leverage these tools?
Colby College developed a platform, called Mule Chat, that allows users to explore several large language models, including ChatGPT, Gemini, Claude and LLaMA. The platform provides a safe on-ramp into generative AI usage and relies on student tutors to disseminate information to peers.
In the latest episode of Voices of Student Success, host Ashley Mowreader speaks with David Watts, the director of Colby College’s Davis Institute for Artificial Intelligence, and Michael Yankoski, Davis AI research and teaching scientist, to learn about the college’s AI institute and how Mule Chat works.
An edited version of the podcast appears below.
Inside Higher Ed: Can we start the conversation by talking a bit about what AI at Colby College looks like? What is the landscape you’re working with and how are you thinking about AI when it comes to teaching and learning?
David Watts: I am new to Davis AI, as we call it at Colby, but the [Davis AI] Institute has actually been around since before ChatGPT, so Colby kind of had a pioneering approach.
David Watts, director of the Davis Institute for AI at Colby College
Colby is a small liberal arts college, and they had the vision that this was going to be around for a while. And rather than, as most institutions were doing, sort of keep it at bay or ban it from campus, Colby dove in and wanted to engage with it and understand how it is going to impact education.
I spent most of my career in industry, mostly in research and development, and so I when I wanted to make the jump over to academia, I wasn’t expecting to find that small liberal arts colleges had done this, and when I saw what Colby had done, I was really drawn to it and came over. So I’ve really loved what has been going on and what continues to go on at Colby with the Davis Institute for Artificial Intelligence.
Inside Higher Ed: Michael, your role puts you directly in connection with faculty when it comes to integrating AI into their classrooms or into their programs. Can you talk about what that looks like and how maybe that looks different at a liberal arts institution?
Michael Yankoski, research and teaching scientist, Davis Institute for AI at Colby College
Michael Yankoski: One of the most amazing aspects of the Davis Institute for Artificial Intelligence here at a place like Colby is the liberal arts approach that the institution as a whole is able to engage with.
That means that we’re able to facilitate conversations from a multiplicity of different disciplines and bring faculty together from different approaches across the divisions in the college—from the STEM fields to the humanities to the social sciences. And have really productive, very generative conversations around ways to engage with artificial intelligence and the shared learning and shared knowledge of people who have been really pioneering in the area. To able to say, “How can I integrate generative artificial intelligence with my pedagogy? How can I help think with students about how to engage these technologies in a way that is beneficial for their education, help empower students in their education and then on the research side?”
Many faculty with whom we work at the Davis Institute are exploring ways to integrate artificial intelligence in their research program, and to say, “Is there a way that artificial intelligence can help me accelerate my research or take my research in new directions?” The opportunity to bring people together to discuss that and to facilitate those conversations across the disciplines is one of the best aspects of the liberal arts approach to artificial intelligence.
Inside Higher Ed: Does Colby have an institutional policy for AI use, or what appropriate AI use looks like?
Watts: It’s a moving target. Anyone who tells you they have it all figured out is probably embellishing. It is a moving target, but one of the things we did was make sure we engage faculty, and in fact, we started with faculty, then we engaged administrators, we engaged students and we engaged general counsel, and evaluated what the challenges are, what the downsides are. And we made sure that we built what we call guidelines rather than policy.
The guidelines talk through the dos and don’ts but also leave enough flexibility for our faculty to think through how they want to engage with AI, especially since AI is a moving target, too. As we grow and learn with our faculty, we adapt and adjust our guidelines and so they’re out there for everyone to see, and we will continue to evolve them as we move forward.
Inside Higher Ed: Can you introduce our listeners to Mule Chat? What is it and how does it work on campus?
Watts: Michael has been here and was one of the originators of creating Mule Chat on campus. And so he can tell you a lot of the details and how it’s been working.
But what I loved about what Michael and the team did, and it was a collaborative effort, was to create, I’ll call it an on-ramp. We were working towards moving the needle from banning AI, as one extreme, to engaging with AI and creating a tool that allowed faculty, students and staff to all easily engage with multiple tools through Mule Chat.
It lowered the activation barrier to entry to AI and allowed us to have an on-ramp for people to come in and start seeing what the possibilities are, and it has worked brilliantly.
Yankoski: The idea behind Mule Chat originally was to provide a place for students, faculty and staff to begin to get experience with and understanding around generative AI. To provide a space where folks could come and understand a bit more about, what are these tools? How do they work? What are they capable of? What are some of the areas we need to be aware of, the risks and the best practices, and how can we provide this on-ramp, as David described, for people to be able to engage with generative artificial intelligence?
This is about student success, empowering students to understand what these technologies are, what they’re good at, what they’re not good at. And then also, one of the key principles here was equity of access. We wanted to ensure that anybody on Colby’s campus, regardless of whether they could afford one of the premium subscription services, was able to get access to these frontier models and to understand how to then do the prompt engineering work, and to then compare the kinds of outputs and capabilities of some of the frontier models. And so really, the core sort of genesis and driving desire for the creation of Mule Chat was to provide this on-ramp that would empower student success, allow equity of access, and also would provide a safe and secure place for people to be able to engage these technologies and to learn.
Inside Higher Ed: Can you describe the functionality of Mule Chat? For someone who has never experimented with LLMs, what does it look like or feel like to engage with Mule Chat?
Watts: You touched on something really great there, because that was part of the idea. We introduced multiple models into Mule Chat so that people could compare and get an idea of what it’s capable of and what it’s not capable of.
I’ll give an example of a faculty member who we are working with right now who started with Mule Chat, engaged with it in their preparation—this is a professor of East Asian studies—how they prepare their classes, realized what the capabilities were, started doing more with it, with their students. The students then brought interesting ideas about what else we can do and pushed beyond even the limits of Mule Chat. And then Davis AI can go help them bring in, for example, they were looking at—not only just looking at old archives and using that in their teaching of East Asian studies, but also bringing in video capability, for example, and in fact, even creating new videos or some of the research that they’re doing now, bringing in more capabilities above and beyond Mule Chat. So it is exactly what Michael was saying, an on-ramp that then opens up the possibilities of what we can do with AI in higher education.
Yankoski: I think the real value of the Mule Chat interface is that it allows people to compare the different models.
Folks can use prompt engineering to compare the outputs of one model and then put that alongside the outputs of another model and be able to observe the way that different models might reason or might do their inference in different kinds of ways.
That side-by-side comparison is a really powerful opportunity for people to engage with the different models and to experience the different kinds of outputs that they create. To build on what David was saying, the ability to then put other tools [like videos] inside of the Mule Chat platform, that allows for deeper research into particular areas. For example, we have a tool that we built, which is called Echo Bot.
The Colby student newspaper is called the Colby Echo, so we’ve been able to bring all the archives of the Echo into a tool that allows students and faculty researchers to engage with those archives and chat with the entire archive of the Colby Echo. We’ve been working closely—and this goes back to the liberal arts approach—with different faculty across campus, as well as the college libraries, to bring this tool online and make it available within the Mule Chat system.
Inside Higher Ed: Let me know if you can build me an IHE bot, because I can never find anything in our archives. I could really benefit from something.
Watts: We can brainstorm on that.
Inside Higher Ed: Great, we’ll talk about licensing later.
I wanted to ask, it seems there’s a new AI tool that pops every other day. So when you’re talking about comparing different tools and thinking about what might be most relevant for students, how often are you scouting out the landscape to understand what’s out there and relevant?
Watts: That’s a great question, and actually extremely important that we do that.
Not only are we reaching out and finding, reading, learning, attending conferences, helping to create conferences ourselves that bring in people and experts who are different perspectives, but we also then have lots of people on campus who have their own ideas. People come to us regularly with, “Oh, look at this cool tool. We should use it for this thing on campus.”
And that’s when we use that for educating people about some of the potential pitfalls that we have to watch out for, talking about guardrails and when you’re bringing in new capability, just like you had to think about when you’re bringing in new software. But I think it’s even more imperative that we’re very careful about what AI tools we bring into campus. You’re absolutely right that there are tons of them that all have different capabilities. But one of the things we try to teach is that there’s a full spectrum: the great, the good, the bad and the ugly. You have to think about that entire spectrum. And that’s one of the beauties of what I loved about coming to a liberal arts college was that you have multiple perspectives, and coming from all forms of disciplines in the humanities, the arts, the natural sciences, the social sciences, and all are engaged and can be engaged across AI.
Yankoski: I think that’s what’s so unique and really powerful about the Davis Institute for Artificial Intelligence approach. When we work with faculty and students and really, if some faculty member or student has an idea that they want to explore, we have structures that allow for technology grants, for faculty to be able to come and to propose the use of a new tool, or to advance their teaching or to advance their research.
Then that’s a great opportunity to engage with that faculty member and perhaps their research assistants, and work with those students and that faculty member to explore the possibility of using that tool. Each faculty member knows their domain so much better than we do. As the core Davis AI team, we’re able to work with that faculty and those students to better understand the use case, better understand the tools that they want to engage, and then work with them to consult and to create a pathway forward. That’s an incredible opportunity as well for the students to understand, how do we think about the security of the data? How do we think about the processing pipeline? How do we think about the best practices with regards to utilizing artificial intelligence in this particular domain?
Really that’s about student empowerment and student success as they get ready to transition out of college into an economy where increasingly expectations around knowledge and the ability to utilize and to vet artificial intelligence are only going to increase.
Inside Higher Ed: How are students engaged in this work?
Yankoski: One the most intriguing aspects of Mule Chat has been that students have been really leading in teaching and empowering other students to utilize the tool and to understand the quantum engineering aspects and to understand the different models.
The student leaders have been working with Mule Chat and then actually teaching other students, teaching faculty and helping lead the sessions, as well as working on their own projects within Google Chat. So it’s been a really strong and quite incredible platform for student engagement and student empowerment as students learn from one another and then are able to learn how to teach about these tools to their peers.
Watts: That’s absolutely a huge part of what we did, and I mentioned that, even though students come first, we started working to move the needle with faculty first on purpose, with students in mind. And then we branched out into, now we can engage the students. Once you have enough buy-in from faculty, start engaging the students, and we’ve been doing a lot of that.
Then what’s beautiful, the magic happens when the students start coming up with thoughts and ideas that grow in ways that faculty haven’t thought of. Because remember that a lot of this is new to faculty as well.
So we actually then will identify key students that we have been working with and actually hire them on board as Davis AI research associates that then help us continue to move the needle, because there’s nothing better for students than to hear from other students about what’s possible. And the same goes for faculty, by the way. So, you know, Michael was mentioning a little bit about our strategy with faculty and how we engage them. But a part of what we do is faculty sessions. We give them creative names like “Bagels and bots,” and we include food and then we have those sessions where faculty talk to faculty. We do the same with the students, so students can talk to students. And it’s just wonderful to see the magic that happens when that begins to grow organically.
Inside Higher Ed: What has the reception been to Mule Chat?
Watts: Most people were skeptical [of AI] early on; most were in the mode of “push it away.” I think that drove some interesting behaviors in faculty and students.
So a big part of what we’ve been trying to do is essentially drive towards AI literacy for all. And when I say all, it’s an interdisciplinary approach. We’re looking across the entire campus, and so all students in all departments are what we’re driving towards. Now, you correctly point out that there will always be skeptics. I will strive for 100 percent, but if we asymptotically approach that into the future, I’ll live with that.
The goal is to prepare students, and that’s who we need to make sure that we’re preparing for the life they’re going to go into that’s been transformed by AI, that touches everybody. One of the cool things is we’re giving out grants to faculty to engage with AI and come up with ideas, and we’re doing that on multiple levels, and those faculty are now coming from all. We have art professors. We have writing professors. We have East Asian studies. We have professors from government, we have all of them engaging and so we’ve been able to, therefore, move the needle quite a bit so that a lot more people are a lot more receptive and open to it on campus, which is great.
Inside Higher Ed: You mentioned that Colby has a faculty-led approach, but sometimes that means that students from specific majors or disciplines might be less exposed to AI than others, depending on who their faculty are. It seems like you all are taking a balanced approach, not only encouraging enthusiastic AI entrepreneurs but also working with the skeptics.
Watts: It’s absolutely critical that we work on both ends of that spectrum, if that makes sense. We’re driving great innovation, and there’s great examples of research right here on campus that are doing wonderful things in an interdisciplinary way.
We just won an NSF grant for ARIA, an NSF institute looking at AI assistance in mental health, because that’s one of the most challenging spaces for how the models interact with people with mental and behavioral health challenges. It’s a perfect example of our interdisciplinary approach, with a professor from psychology working with a professor from computer science to go tackle these challenging areas. And I think that’s one of the things that Colby has done well, is to take that broader, interdisciplinary approach. Many people say that word now, but I think the liberal arts are primed for leading the charge on what that’s going to look like, because AI, by its nature, is interdisciplinary.
Inside Higher Ed: What’s next on campus? Is there any area that you’re all exploring or looking to do some more research in, or new tools and initiatives that our listeners should know about for the future?
Watts: We’re consistently evaluating that and bringing them in. What we’re trying to do is let it grow based on need as people explore and come up with ideas.
I mentioned the video; we’re now enabling video capability so we can do some of that research. It also opens up more multimodal approaches.
One of the approaches to the ARIA research, for example, is we want to be able to detect and therefore build context-aware assistance to have better results for everyone. So if we can solve the mental and behavioral health challenges, it’s probably one of the most difficult ones. It can also solve some of the other areas of underrepresented people who are left out or underrepresented groups who are left out of training, for example, which can lead to challenging behaviors.
I’m really excited about all of those possibilities and the areas that allow us to enable. We talked about access, we can also talk about accessibility.
We have on campus the Colby College Museum of Art; one of the faculty in computer science is exploring accessibility options using AI with a robotic seeing-eye dog. If someone wanted to visit the museum who was blind or visually impaired, they could interact with a seeing-eye dog that they’re used to, but this seeing-eye dog now might have more capability to communicate with people about what they’re seeing and in a museum setting, for example.
So really excited about that type of research: how do we really benefit humanity with these types of tools.
Inside Higher Ed: One thing I wanted to ask about is resources allocated from the university to be able to access all these tools. What investment is the college making to ensure that students are able to stay on the cutting edge of AI initiatives?
Watts: That’s absolutely critical. We want to make it no cost to our students and accessible to our students, but it still costs. So [it’s vital to] make sure that we have funding.
We were very lucky that we got a Davis endowment that enabled us to build the Davis Institute. That was huge because, and you can think about some of the challenges with federal funding and all of that stuff, but to have an endowment that allowed us to draw on that and really build strong capabilities at Colby College was critical. But you’re touching on the fact that we’re going to need to continue to do that. And that’s where, for example, the NSF grant and other grants that we will continue to explore will help us with how we continue to grow our impact and grow our value as we head into the future.
New early-applicant data from the Common App found that applications from Black, low-income, first-generation and rural potential students are all up compared to this point last year. However, international applications dipped, and the most selective institutions are experiencing the smallest application growth compared to other types of institutions. Applicants are also increasingly choosing to submit standardized test scores.
The Common App report, released Thursday, is the first in a series of monthly research briefs on college applicant trends typically released between November and March. The November brief showed that applicants, and applications, rose over all compared to this time last year, with notable growth among particular groups.
For example, applications from those who identified as Black or African American increased 16 percent and multiracial applicants rose 11 percent compared to the same time last application season. The report also found that applicants who identified as first-generation grew by 12 percent, while low-income applicants, who qualified for a Common App fee waiver, increased at more than twice the rate of other applicants. Rural applicants grew by 15 percent compared to last year, while thosefrom metropolitan areas grew only 6 percent.
But the number of international students applying dropped 9 percent compared to this point last year, driven by a 14 percent drop in applicants from India, which has historically been the second-biggest source of international applicants on the Common App platformafter China. Applicants from Asia broadly and from Africa also dropped significantly, 9 percent and 18 percent respectively, with a whopping 43 percent decline in applicants from Ghana. These trends suggest theTrump administration’s policies, including international student visa delays and denials, may be deterring students.
At a time when highly selective institutions are under new political pressures, the report found that colleges and universities with admit rates of 25 percent or below had the slowest application growth, at 4 percent. Applications to other types of institutions grew at two or three times that rate.
The return of standardized test requirements at some institutions is also driving more applicants to submit test scores. Notably, applications reporting scores rose 11 percent compared to this time last year. However, students who identify as underrepresented minorities or first-generation or who qualify for a Common App waiver were less likely to share their scores.
Johns Hopkins University announced Thursday that it’s eliminating tuition, fees and living expenses for its Homewood campus undergraduates whose families make less than $100,000 a year; students whose families earn up to $200,000 will pay no tuition. It joins a wave of other institutions—especially private, selective ones—that have announced tuition guarantees.
In a news release, the university said the change “means students from a majority of American families, including middle-class families earning above the national median household income of $87,730, can attend Hopkins at no expense.”
Further, Hopkins said, “Most families with incomes up to $250,000 will continue to qualify for significant financial aid. Even those with annual incomes exceeding $250,000 may qualify, especially when there are multiple children in college at the same time.”
Most of the university’s undergrads study on the Homewood campus, in North Baltimore. The release said the new aid levels “will go into effect for eligible current students in the spring 2026 semester and for new, incoming students next fall.”
In a message to the university community, JHU president Ron Daniels said that since businessman and former New York mayor Michael Bloomberg donated $1.8 billion to the university in 2018, Hopkins’s share of Pell Grant–eligible students rose from 15.4 percent to 24.1 percent, the highest proportion in university history.
“Our financial aid investment has continued to grow, inspired by Mayor Bloomberg’s transformative gift, with generous contributions by more than 1,200 donors who have given $240 million for financial aid at Hopkins over the last several years,” Daniels wrote. “We are in their collective debt.”
Sterling College will close at the end of the spring semester, officials announced Wednesday.
The small college in Craftsbury Common, Vt., will cease operations in May due to “persistent financial and enrollment challenges,” according to a statement posted on its website.
“We understand that this news is difficult and deeply personal for every member of our community. Sterling College has always been more than a place of learning; it has been a home where curiosity, creativity, and compassion thrived,” officials wrote in the closure announcement.
Sterling, which offered “transdisciplinary, experiential, competency-assessed educational programs,” according to its website, historically capped enrollment at 125 students. Founded in 1958, Sterling is one of a few U.S. work colleges, a model that allows students to keep tuition down via campus labor. Residential students at Sterling work five hours per week in different roles.
Federal data shows that Sterling only had a head count of 78 students in fall 2023.
While the college managed to eke out modest surpluses in recent years, it had a meager endowment of just over $1.1 million, much of that restricted, according to financial documents.
Sterling is now the second institution to announce a closure this month, following Trinity Christian College in Illinois, which is shutting down next year due to similar challenges.
In August, the Trump administration issued an executive action ordering colleges and universities to submit disaggregated data about their applicants and prove they are following the letter of the law when it comes to race in admissions. But a new notice, published to the Federal Register Wednesday, clarifies that the mandate only applies to four-year institutions.
“We posed a directed question to the public to seek their feedback … [and] based both upon our initial thinking and public comment, we propose limit[ing] eligibility of [the new IPEDS Admissions and Consumer Transparency Supplement] to the four-year sector,” the notice stated.
Colleges that are obligated to comply must still submit six years’ worth of application and admissions data, disaggregated by student race and sex, during the next survey cycle, it said. But any college that admits 100 percent of its applicants and does not award merit or identity-based aid will be exempt.
Since the action was first published, institutions across the sector have warned the Trump administration that collecting and reporting such data would be a difficult task and place an undue burden on admissions offices. But with smaller staff sizes and limited resources, community colleges were particularly adamant about the challenge the requirement posed.
“It’s not just as easy as collecting data,” Paul Schroeder, the executive director of the Council of Professional Associations on Federal Statistics, toldInside Higher Ed in August. “It’s not just asking a couple questions about the race and ethnicity of those who were admitted versus those who applied. It’s a lot of work. It’s a lot of hours. It’s not going to be fast.”
When I first began teaching Islam, there was no road map. In 2001, I was a visiting assistant professor of Islamic and Middle Eastern studies at the University of Iowa—the first full-time professor of Islam in the history of the state. I was in my 20s, still finishing my dissertation, when the attacks of Sept. 11 unfolded. Suddenly, I found myself trying to explain a 1,400-year-old religion to students who had watched the Twin Towers fall on live television.
Teaching Islam in American universities has never been more widespread, more diverse or more embattled. That is the story of the past two decades: a field that has grown dramatically, transformed in terms of who teaches it, and now finds itself under intensifying political scrutiny.
That experience in Iowa shaped everything that came after. I discovered that my task was not only to introduce students to the theological, historical and cultural breadth of Islam but also to help them unlearn the simplistic caricatures they had absorbed from media and politics. Islam was not a monolith. It was not synonymous with terror. It was, like Christianity or Judaism, a faith defined by argument, diversity and adaptation.
Those class lectures eventually became the foundation for No god but God: The Origins, Evolution and Future of Islam, first published in 2005. I hoped the book would serve both general readers and university classrooms. To my surprise, it quickly became a popular text for teaching Islam in the United States and far beyond. It has been translated into dozens of languages, adopted in seminaries and world religion courses, and read in mosques, churches and synagogues.
Two decades later, the landscape of Islamic studies in American universities looks profoundly different. In 2001, very few institutions offered dedicated courses on Islam outside of theology departments. Today, there are hundreds of such courses, spanning history, political science, gender studies and literature. The proliferation has been remarkable—though uneven. Some courses are rigorous, rooted in language and text, while others are more ad hoc, responding to student demand and global events.
Another profound shift has been in who is teaching Islam. For most of the modern history of religious studies in America, Christian professors taught Christianity, Jewish professors taught Judaism—but it was rare to find Muslim professors teaching Islam. In nearly two decades of studying the subject, I had only one Muslim professor. That has changed dramatically. Today, Muslim scholars occupy faculty positions across the country, and new professional associations—such as the International Quranic Studies Association, of which I am a member—are fostering networks of Muslim academics who bring both scholarly expertise and lived experience into the classroom. This diversification has expanded the kinds of questions and perspectives that shape the field, though it has also forced universities to confront new debates over authority, representation and bias.
Meanwhile, the teaching of Islam—like so many fields in the humanities—is now buffeted by unprecedented political pressure. Across the country, state governments have moved to limit what can and cannot be taught in universities and ban diversity, equity and inclusion programs. More recently, elite universities such as Columbia and Harvard have faced political scrutiny from the Trump administration and Congress into their Middle East studies programs, accused by some lawmakers of being biased. In today’s climate, teaching Islam can feel like an act of defiance. Professors often self-censor, conscious that a stray lecture note could trigger outside campaigns or even threats. The irony is that in a moment when greater understanding of Islam is needed more than ever, the very institutions best equipped to provide that education are being undermined.
Yet this is precisely why teaching Islam in universities matters more than ever. At a time when Islam has faded from the headlines but remains entangled in the debates that define our era—from authoritarianism to surveillance to religious pluralism—the classroom is one of the few places where the faith can be encountered on its own terms. The role of professors is not to sanitize or defend Islam, but to present it in all its richness, contradictions and ongoing transformations.
The fully updated 20th-anniversary edition of No god but God is my attempt to support that task for another generation of teachers and students. The new preface reflects on what has changed since 2005—the Arab Spring, the rise of digital Islam, the ebb of the “war on terror”—and what has not: Islam’s enduring struggle to reconcile tradition and modernity, authority and pluralism.
More than two decades of teaching have convinced me that education about Islam cannot be episodic, tied only to moments of crisis or headlines of violence. It must be sustained, interdisciplinary and grounded in serious scholarship. It must expand beyond political science courses on terrorism and foreign policy, and beyond theology seminars comparing sacred texts, into the wider humanities and social sciences. And it must center the lived experiences of Muslims themselves.
The classroom is not a mosque. But it is one of the few spaces where young people can confront their assumptions, wrestle with complexity, and imagine new ways of understanding the role of religion in the world. That was my conviction in 2001, when I walked into a lecture hall in Iowa just days after Sept. 11. It remains my conviction today.
The classroom may not be a mosque, but it remains one of the few places where Islam can be encountered in all its richness, contradictions and humanity.
Reza Aslan is a writer and scholar of religion. His books includeZealot: The Life and Times of Jesus of Nazareth and No god but God: The Origins, Evolution, and Future of Islam, now available in an updated 20th-anniversary edition from Random House. He is a professor of creative writing at the University of California, Riverside.