State is planning to oust 38 universities from a partnership, The Guardian reported.
Drew Angerer/AFP/Getty Images | Lance King, Mario Tama and Justin Sullivan/Getty Images | Liz Albro Photography/iStock/Getty Images
The State Department’s Diplomacy Lab program says it enables students to work on real policy issues, benefitting both their careers and American foreign policy through their research and perspectives. It’s meant to “broaden the State Department’s research base in response to a proliferation of complex global challenges,” according to the program website.
But now the Trump administration’s domestic policy fight against diversity, equity and inclusion could upend this partnership between the State Department and universities. The Guardianreported last week that the department is planning to suspend 38 institutions from the program, effective Jan. 1, because they had what the department dubbed a “clear DEI hiring policy.” It’s unclear how the department defines that phrase or how it determined these institutions have such policies.
On Tuesday, The Guardian—citing what it called an unfinalized “internal memo and spreadsheet”—published the list of institutions that State Department plans to kick out, keep in or add to the program. A State Department spokesperson didn’t confirm or deny the list to Inside Higher Ed or provide an interview, but sent an email reiterating the administration’s anti-DEI stance.
“The Trump Administration is very clear about its stance on DEI,” the unnamed spokesperson wrote. “The State Department is reviewing all programs to ensure that they are in line with the President’s agenda.”
The institutions to be ousted, per The Guardian’s list, range from selective institutions such as Northeastern, Stanford and Yale universities to relatively small institutions including Colorado College, Gettysburg College and Monmouth University. The 10 universities to be added include Gallaudet University, which specializes in educating deaf and hard-of-hearing students, Liberty University, a conservative Christian institution, and the St. Louis and Kansas City campuses of the University of Missouri system. In all, the list shows plans for 76 institutions.
The shakeup appears to be yet another consequence of the Trump administration’s now nearly year-long campaign to pressure universities to end alleged affirmative action programs or policies. The day after his inauguration, Trump signed an executive order mandating an end to “illegal DEI” and calling for restoring “merit-based opportunity.” But Trump’s order didn’t define DEI.
Through cutting off federal research funding and other blunt means, the administration has tried to push universities to end alleged DEI practices. A few have settled with the administration to restore funding; Columbia University agreed this summer to pay a $221 million fine and to not, among other things, “promote unlawful DEI goals” or “promote unlawful efforts to achieve race-based outcomes, quotas, diversity targets, or similar efforts.” Columbia is among the institutions that the State Department intends to keep in the program, according to The Guardian’s list.
Inside Higher Ed reached out Tuesday to the institutions listed to be ousted. Those who responded suggested the program didn’t provide much, or any, funding, and said they didn’t engage in any illegal hiring practices.
The University of Southern California said in a statement that it “appreciated travel funding provided by the Diplomacy Lab program to two USC students in 2017 and looks forward to future opportunities to collaborate.” The university said that was the last time it received funding, and said it “complies with all applicable federal nondiscrimination laws and does not engage in any unlawful DEI hiring practices.”
Oakland University political science department chair and Diplomacy Lab campus coordinator Peter F. Trumbore said through a spokesperson that he hasn’t received notice of a change in status as a partner institution. He also said his university received no funding from the State Department for the program, though “our students have had invaluable experiences conducting research on behalf of State, and working with State Department stakeholders in producing and presenting their work.”
Georgia Institute of Technology spokesperson Blair Meeks said his university also never received funding from the State Department for the program. He also said “Georgia Tech does not discriminate in any of its functions including admissions, educational, and employment programs. We have taken extensive actions over time to eliminate any programs, positions, or activities that could be perceived as DEI in nature.”
Meeks further wrote that the State Department “communicated that cuts or halts to the program were associated with the federal government shutdown” that ended earlier this month. Sarah Voigt, a spokesperson for St. Catherine University, said in an email that the State Department told her university back on Jan. 31 that it was pausing Diplomacy Lab activities, so the institution didn’t apply for research opportunities this semester. Then, last week, the State Department told the university that “‘due to the delays caused by the shutdown,’ they were again pausing Diplomacy Lab activities.”
“Our understanding is that the program was shut down due to a lack of government funding,” she wrote.
“The University had been participating as a Diplomacy Lab Partner Institution since early 2020, and we appreciated the opportunities to offer our students and faculty members very timely research topics through this program,” she added. “If the Department of State were to resume Diplomacy Lab activities, we would review what opportunities were available.”
Two years after the Supreme Court banned the use of race in college admissions decisions and in the wake of the Trump administration’s attacks on diversity, equity and inclusion, colleges’ use of diversity- and identity-related supplemental essay prompts is patchy.
After a boom in prompts about applicant’s identities, several universities have scrapped the essays entirely for the 2025–2026 admission cycle. Still others, especially selective universities, have kept the prompts, saying they are the best way to get to know their applicants.
Kelsea Conlin, who oversees the college essay counseling team for College Transitions, an admissions consulting firm, identified 19 colleges with optional or required diversity essays last admission cycle that either had dropped or reworded those prompts this year.
“I’ve seen very few colleges that still require students to write about diversity; the prompt may still be on their application and students have the opportunity to write about it, but it’s an optional essay,” she said.
Diversity-related essays often ask students to describe how they’ve been shaped by their community, culture or background, sometimes prompting them to describe how those identities will bring something new to a campus. Others ask students to discuss or reflect on issues like diversity, social justice or antiracism more broadly.
In the majority opinion in Students for Fair Admissions v. Harvard, Chief Justice John Roberts said it was acceptable for students to continue discussing race in their essays: “Nothing in this opinion should be construed as prohibiting universities from considering an applicant’s discussion of how race affected his or her life, be it through discrimination, inspiration, or otherwise.”
The following application cycle, several colleges introduced diversity-related essay prompts to their applications, according to research by Sonja Starr, a law professor at the University of Chicago; Conlin also said she observed a surge in these essays in the 2023–2024 application cycle.
But this year, the Department of Justice issued guidance warning institutions against using “proxies” for race in admissions and hiring, and described requirements for applicants to “describe ‘obstacles they have overcome’ or submit a ‘diversity statement’ in a manner that advantages those who discuss experiences intrinsically tied to protected characteristics” as examples.
“The administration basically says, … ‘if you are letting the desire for a diverse campus influence your policies in any way, that is just as unconstitutional as taking the individual applicant’s race into account,’” Starr said. “I think that’s a wrong reading of the law.”
Still, she said she’s not surprised institutions may be wary of maintaining essay questions overtly related to identity, considering the harsh actions the administration has taken against colleges it disagrees with.
“There’s all kinds of ways the federal government can really make it difficult for universities,” she said, pointing out the slew of funding the administration has cut or frozen over the past ten months. “[Some institutions], I think, are just trying to at least stay out of the administration’s way.”
Simplifying the Process
Several institutions told Inside Higher Ed that they cycle out their essay prompts regularly, so the change from last year’s diversity question was par for the course. Others said they eliminated their supplemental essay requirements altogether, in an effort to make the application process less strenuous.
The University of Washington, which removed a supplemental essay asking prospective students to describe how their background and the communities they are involved in would contribute to the campus’s diversity, told Inside Higher Ed in an email that they hope the removal of the essay will make the admissions process less strenuous for applicants.
“During the annual review of our application process, we determined that an additional essay did not provide sufficient value when reviewing students for admission. We discovered that some applicants, like those interested in our honors program, were previously seeing up to four essay prompts. This change simplifies the process for all our applicants,” wrote David Rey, associate director of strategic communications.
A University of Virginia spokesperson gave a similar statement to the campus student newspaper, The Cavalier Daily, about its decision to remove a diversity essay prompt introduced in the 2025–2026 application cycle, saying that its removal aimed to “lighten the load and reduce stress and anxiety around the college application process.” UVA did not respond to Inside Higher Ed‘s request for comment.
Does that mean supplemental essays are falling out of vogue? Not necessarily, Conlin said; a significant number of selective universities still require them, and the students she works with are generally writing just as many supplemental essays as they have in previous years.
Despite some institutions opting to change or remove their diversity prompts, though, Ethan Sawyer, the founder of the admissions consulting firm College Essay Guy, said that a review of 300 institutions’ prompts for the 2025–2026 admission season showed that questions about what a student’s identity will bring to the institution are the most popular for the second year in a row.
He said in an email to Inside Higher Ed that these prompts have proven to be particularly effective at providing colleges with the key information they’re looking for out of an admissions essay. The identity prompt acts as the new “Why Us” essay, but avoids the pitfall of students focusing exclusively on the college’s attributes rather than their own.
“It lets colleges learn what they’ve always wanted to know—how will this student engage with our community? What qualities will they bring?—but through a framing that encourages students to reflect on who they are (as opposed to how awesome the college is). In other words, colleges are still trying to understand fit; they’re just using a lens that better centers the student,” he said.
Students Still Write About Race
While some colleges may be scrapping diversity prompts, many students want to write about their identities, Conlin and Sawyer said.
“They don’t see themselves through just one lens. No student wants to be reduced to a single label or experience. They understand they’re complex people shaped by many different identities, roles, and life moments,” Sawyer wrote. “Part of our job as counselors is to help them express that complexity—to choose which pieces of their story to spotlight in each essay, and to show how those pieces translate into contributions they’ll make on a college campus.”
Many of the new or reworded essay prompts that have replaced diversity-related questions are broad enough that students can still talk about their identities and experience if they choose to, Conlin noted. In her experience, students are often interested in discussing their race or first-generation student status in essays. But students are more reluctant to write about being LGBTQ+ or having mental health struggles.
Diversity essays aside, Conlin also noted two burgeoning categories of essay topics this year: prompts asking students to talk about how they handle conflict and prompts offering students the chance to explain their relationship with AI.
Oakwood University supports the Trump administration’s controversial compact for higher education that would require signatories to make changes to their policies in order to receive a potential edge in federal funding, Religion News Service (RNS) reported.
The historically Black university in Alabama wrote a Nov. 18 letter to the Education Department about its interest in the compact. Oakwood is the second HBCU to show interest in signing on. Like the other HBCU, Saint Augustine’s University, Oakwood officials say the compact needs to change for them to actually sign it.
Of concern for the HBCUS are provisions that would cap undergraduate international student enrollment at 15 percent, require a five-year tuition freeze and limit the use of race in admissions and other decisions.
Most Popular
“While we strongly support the Compact’s overarching goals, several provisions of the draft framework raise important concerns that, if left unaddressed, could unintendedly hinder HBCUs’ ability to participate fully or effectively,” Oakwood President Gina Brown wrote in the letter, according to RNS. “Absent a mission-based exemption, HBCUs would face an untenable choice between compliance and fulfilling their congressionally mandated purpose.”
Oakwood is affiliated with the Seventh Day Adventist Church, and RNS noted that faith-based institutions would still be able to consider religion in admissions and hiring.
The Trump administration invited nine universities to give feedback on the proposed compact. Most of that group declined outright to sign it, saying that federal funding should be based on merit, not adherence to a president’s priorities. Since then, New College of Florida, Valley Forge Military College and Saint Augustine’s have indicated interest in joining the compact.
Four people were arrested at a Nov. 10 protest at UC Berkeley over an event for Turning Point USA.
Justin Sullivan/Getty Images
The Department of Education is reviewing potential violations of the Clery Act at the University of California, Berkeley following violence at a campus protest.
Fights broke out and four people were arrested at a Nov. 10 protest against an event for Turning Point USA, the conservative student group founded by Charlie Kirk, Cal Matters reported. The organization has received newfound attention after Kirk was shot and killed while speaking at Utah Valley University in September, exactly two months before the event at UC Berkeley.
The Department of Education announced the launch of the investigation Tuesday.
“Just two months after Turning Point USA founder Charlie Kirk was brutally assassinated on a college campus, UC Berkeley allowed a protest of a Turning Point USA event on its grounds to turn unruly and violent, jeopardizing the safety of its students and staff. Accordingly, the Department is conducting a review of UC Berkeley to ensure that it has the procedures in place to uphold its legal obligation to maintain campus safety and security,” U.S. Secretary of Education Linda McMahon said in a statement.
ED also accused the university of having “a history of violating the Clery Act” in a news release announcing the investigation, citing a $2.4 million fine and settlement agreement in 2020 for UC Berkeley’s failure to properly classify 1,125 crimes on campus and insufficient record keeping.
The Department of Justice previously announced a probe into the university earlier this month, claiming that “Antifa,” a decentralized, left-wing movement was involved in the Nov. 10 protests.
UC Berkeley spokesperson Dan Mogulof told Inside Higher Ed by email that the university “has an unwavering commitment to abide by the laws, rules and policies that are applicable to the university” and “will continue to cooperate with governmental inquires and investigations.”
Mogulof added that the university provided public reports about two violent crimes that occurred Nov. 10: a fistfight over an attempted robbery and someone being hit by a thrown object. He also highlighted efforts by administrators “to support the First Amendment rights of all by deploying a large number of police officers from multiple jurisdictions, and a large number of contracted private security personnel” and closing off parts of campus on the day of the protest.
The investigation comes as the Trump administration has clashed with the University of California system in recent months as it sought to cut off federal research funding over alleged antisemitism and how administrators handled pro-Palestinian campus protests in spring 2024. The federal government has also demanded the University of California, Los Angeles, agree to a $1.2 billion fine and make a number of changes in response to the administration’s concerns.
A federal judge recently ruled against the federal government and its “blanket policy of denying any future grants” to UCLA and determined that the Trump administration can’t demand payouts from University of California member institutions as it conducts civil rights investigations.
According to an August email obtained by The Chronicle, Jenny Edmonds, associate dean of communications and marketing at Duke’s Sanford School of Public Policy, encouraged faculty to “continue to engage with the media to disseminate [their] research as [they] have always done,” while also cautioning that “media attention to institutions of higher education and discussions about institutional responses to policy changes have become more prominent than ever.”
“In this moment in particular, questions about Duke and current events are being answered by Frank Tramble and his team,” Edmonds wrote. “If you are contacted by the media about overarching issues confronting the University, please forward the requests to [Sanford’s Senior Public Relations Manager Matt LoJacono] and me.”
Although it wasn’t a universitywide directive, The Chronicle obtained emails that show some other departments also gave their faculty similar instructions to route media requests through the university’s central communications channels.
At an Academic Council meeting in October, Duke’s president, Vincent Price, and council chair, Mark Anthony Neal, commended faculty members for not speaking to a New York Times reporter; the reporter had visited the campus while working on a story about the Trump administration targeting Duke’s diversity, equity and inclusion program.
“It was pretty amazing that [the reporter] actually got no commentary from Duke officials and Duke faculty,” Neal continued. “Even if it wasn’t overtly communicated to the community, the community understood the stakes of that mode of inquiry.”
At that meeting Price also called Trump’s higher education compact—which would allegedly give universities preferential funding in exchange for making sweeping institutional policy changes— “highly problematic,” according to The Chronicle. Despite public pressure, Duke hasn’t officially rejected the terms of the compact.
Abedini’s detention makes real the fears of many foreign and American academics who are rethinking or boycotting travel to academic conferences in the U.S.
Peter Zay/AFP/Getty Images
Immigration and Customs Enforcement agents detained a University of Oklahoma professor Saturday while he was on his way to a conference.
Vahid Abedini, a professor of Iranian Studies, was stopped and detained while he was boarding his flight to attend the Middle East Studies Association conference in Washington, D.C. He was released Monday night, according to a LinkedIn post.
“I’m relieved to share that I was released from custody tonight. It was a deeply distressing experience, especially seeing those without the support I had,” Abedini wrote on LinkedIn early Tuesday morning. “My sincere thanks to my friends and colleagues at the University of Oklahoma, the Middle East Studies Association, and the wider Iran studies and political science community for helping resolve this.”
Abedini did not respond to Inside Higher Ed’s request for comment. According to Joshua Landis, Abedini’s colleague and co-director of the Center for Middle East Studies at the University of Oklahoma, Abedini has an H-1B visa.
“ICE arrested our beloved professor Vahid Abedini,” Landis wrote on X Monday. “He has been wrongfully detained because he has a valid H-1B visa—a non-immigrant work visa granted to individuals in ‘specialty occupations,’ including higher education faculty. We are praying for his swift release.”
Reached for comment, a Department of Homeland Security spokesperson told Inside Higher Ed: “This Iranian national was detained for standard questioning. He’s been released.”
Abedini’s detention makes real the fears of many foreign and American academics who are rethinking or boycotting travel to academic conferences in the U.S. due to concerns about wrongful arrests by immigration enforcement.
In a statement, the MESA Board of Directors said they were “disturbed” to learn of Abedini’s detention and “deeply concerned” about the circumstances. The University of Oklahoma declined to comment on the situation.
Colleges and universities sit on a large wealth of data, ranging from student attendance and interactions with learning management systems to employment and earnings data for graduates. But uniting legacy systems and having responsive data remains a wicked problem for many institutions.
This year, Central New Mexico Community College is deploying a new AI-powered predictive analytics tool, CampusLens, part of CampusWorks, to improve data visibility in student retention, early alerts and career outcomes.
In the latest episode of Voices of Student Success, host Ashley Mowreader speaks with Tracy Hartzler, president of Central New Mexico Community College, to discuss the risks with taking on new tools, the college’s approach to change management and the need for more responsive data.
An edited version of the podcast appears below.
Q: Can you introduce yourself, your role and your institution?
Tracy Hartzler, President of Central New Mexico Community College
A: My name is Tracy Hartzler. I’m president of Central New Mexico Community College. We’re located in Albuquerque, New Mexico. We serve three counties around us, and our population is about 900,000 residents in our area, so it’s about half the state of New Mexico who lives in our service area, but it’s an incredibly diverse area.
We have a significant population of Hispanic, Latino students. We have a large population of Indigenous students, as well. We are the largest undergraduate institution in the state, and that’s distinct because we only issue or grant certificates and associate degrees. We are not a bachelor’s degree–granting institution, so our focus really is on those students who are seeking entry into college—whether that’s our dual credit students who are still in high school—but also those who are returning for upskilling. They’ve already earned their bachelor’s degree or degrees, and they’re coming back for some hands-on or applied skills, or those who are getting back into education and training because they’re looking for greater financial stability.
Like so many other colleges, we know we want to learn from others, and so we’re really proud that we work with many of our other colleges across the state of New Mexico, but we certainly engage in conversations with leaders and schools who participate in American Association of Community Colleges who are part of the global community college leader network.
But we’re really pleased and we’ve been really pushed by our peers who are members of the Alliance for Innovation and Transformation—group of higher education institutions, there’s about 60 of us—with some other thought partners to really help us think how we can best leverage technology and change our processes and deliver better education and training for our students and better serve our employers.
We also are relying on lessons learned from those outside of higher education, so whether it’s in hospitality, healthcare, manufacturing and others. So while we know we have great work to do in New Mexico, and we are incredibly fortunate to work with strong partners who tell us what they want and how we can best serve them, we certainly look to other schools and other organizations to help us make those transitions faster so we can better serve our community.
Q: From my vantage point, it seems community colleges are often some of the most nimble when it comes to learning from other institutions.
A: Absolutely. You know, it’s great to be scrappy. I think we and here at CNM, we certainly punch above our weight. We are excited to take on new challenges. We are, frankly, fortunate to be able to move faster. So, if something doesn’t work, we can pivot away from it entirely or continue to revise it. And frankly, the urgency to do so is really placed on us by our employer partners, our community partners and our students. They really are pressing us to be responsive to them because they don’t have time to waste, and they certainly don’t have resources to waste. So, we really step in. And again, I don’t think CNM is unique in that we all respond to the need as quickly and as best we can.
Q: We’re talking today about a new AI-powered predictive analytics tool that you all implemented. Historically, what has been the college’s retention and persistence strategy? What are some of the challenges you have seen when supporting students?
A: Before the pandemic, we were able to and we were participating in a lot of futures work: What do students need? What do they need now? What do employers need in the future? Which, at that time, seemed so far off, and so we were already on a journey again, whether it was working with AFIT or others to help us better identify what we needed to do and how we needed to change to better meet our student and our employer needs.
We knew that that would include certainly technology and leveraging technology, but we also knew it would mean changing how we do things, how we schedule, how we use the data in our systems. And we also knew we have a tremendous amount of information. We have a lot of data, but like so many other places, it’s in seven legacy systems. And we have over 100 applications that help our data systems talk to each other, to generate reports that our staff use, and it’s incredibly challenging to wrangle this data in a way that is useful, that helps us drive and drive change again.
Most of the data is legacy data. It’s what happened last year, and how do we think that’s going to improve? What are we doing now to then improve performance a year ahead, and then we hope that what we do over the next year meets the need, but it takes us too long to really react.
So, we were looking for ways to take the assets we have—which includes our incredible faculty and a number of our leaders and our office of data strategy and some of our contractors, like CampusWorks and consultants—to help us wrangle this data in a way that helps us be data informed in a time-sensitive way.
We had a lot of processes in place that were helping us to do that. A lot of our steps were manual and creating reports, and it really slowed down what our frontline advisers and navigators and employment advisers really were able to do, because we were requiring them to do so much manipulation with the data then to be able to identify what they should do once they got this great report.
So, we were looking for ways to leverage technology. And again, the pandemic happens. We’re increasingly dependent on our systems, using them to greater degrees than we had before, including our learning management system. We are also undertaking a transition conversion from our old student information system to a new student information systemin Workday. We’re making all these changes and upgrading technology, and frankly, AI is coming along that’s really dramatically changing how we work, or could change how we work. We’re trying to figure out a better way to wrangle all these opportunities.
We were so excited to learn about CampusWorks and their product, CampusLens, because we think that tool will help us leapfrog, not only the tool but the experts that that CampusWorks brings to the table to help us to analyze data and develop tools that will help our frontline staff much more quickly and easily identify how they can help students. To register from class A to B, to help them identify all the predictors that say, if the student’s missing one assignment or they haven’t attended class, here are the automatic prompts for you as an adviser or navigator—or if you choose to automate that process you can. But really, how can we help individuals—our employees—still help and better connect with students to keep them on the track of success?
It certainly can also help us schedule, help our faculty and all of our associate deans who do incredible jobs trying to figure out ways to schedule our incredible programming to be most effective for students. Some of this information that we’ve had in different places, when it comes together in a product like CampusLens, will help us generate these tools so that we can we can more quickly assess our situation and better adapt, test, try and iterate ways to better, like I said, schedule classes, schedule our work-based experiences, help our employers predict the number of graduates who are going to come out of our programs at any given time.
When we have employers come to us with dramatic needs, you know, they need 1,000 technicians over three years, well, what do we need to do to scale and ramp up our programming to meet that need?
I’m excited that we have a tool that will help us do that, instead of the army of staff and technical staff that I would have to try to find to help us do that in an efficient way. That’s why a product and a team, a quality, curious and an innovative team at CampusWorks to help us work through some of these projects.
Q: How does the tool work logistically? What are you excited about when it comes to the capabilities of CampusWorks and CampusLens?
A: It helps us better, frankly, use staff time to keep students and others on the right track—on the track that they’ve chosen, by the way.
What is most exciting, at least for us with CampusLens, is their Career Lens. So all institutions, all community colleges, are focused on many phrases, but all go to the federal emphasis, or your statewide emphasis on return on investment. What is the value that a learner gets from your program that can be defined a lot of ways. It could be defined by wages, wages a year out, it could be defined for many years out from completing a certificate or degree. It looks at what’s your job in a particular program. We know the federal government, whether when they’re leading the rules around rule-making for Workforce Pell, we know that those regulations are going to help us require that we analyze our programs for results. Will these programs allow students to be eligible for federal assistance?
We know that we can use all of this data and CampusLens is going to be able to help us identify which programs are eligible for Workforce Pell, what are the wages? It’ll help us report out the successes of our programs, or, frankly, identify those programs where wages are not at the median level. What do we need to do, then, to repackage or reschedule or build up some of our programs to meet the wage requirements that we want individuals to accomplish, to achieve and earn, but also that will meet some of our federal standards?
So, I’m really excited about the workforce component of this, which is really what we’re all looking for. All of us [higher education leaders] want, I’d say, a silver bullet when it comes to unifying this data and being able to tell the story and being able to design programming is responsive and frankly to be able to tell our stakeholders, whether they’re legislators or federal government agencies giving us funding for workforce training, what are the outcomes? That’s so important that we’re able to show and tell the story with really valuable data? And I’m excited that CampusLens allows us to achieve that.
Q: How have you all been thinking about AI as a tool on campus, what are those conversation looking like with your staff and employees?
A: I want to start with our governing board. CNM is governed by a seven-member elected governing board, and our governing board represents geographic areas in our community. They are focused on how we are best serving our students, our employers and our community members and what does that mean for technology? Is the college investing in programming and the right tools? Are we getting the greatest benefit from the tools we’ve purchased? That also includes the question of, how are our faculty and staff using the tools to better help our learners?
We talked about retention and persistence and how we use data, but it takes training and professional development to be able to use the tools to the greatest advantage. And of course, this is all in service to our learners and our employers. So it starts with our leadership, and then it flows through.
I don’t think we’re any different than a lot of other colleges. We’ve looked at our policies, and we’ve built on our existing academic integrity policies around AI use, and we have faculty policies on how they describe and expect use, or have authorized use, if you will, approved use, embedded use of AI in their coursework, in their programs. We have policies in place.
We also have done some pilot work. We’ve created a fund for individuals to come to a group around data, frankly, out of our data governance team and our IT team to be able to pitch ideas for three-month sprints or pilots, and they report back. What was the result? What did we learn? Is it something we should scale
I will say many of those pilots are both on the business side or the operation side of the college, but also on the student and teaching and learning side. So that’s really interesting. We look forward to some of those first official pilots coming forward in the next month or so.
What I’m most excited about, though, is the systemic use of AI across the institution. I appreciate the pilots get us excited and interested. It gets people familiar with tools as they evolve and change. But how do we embed AI into our systems work? That’s why I’m excited about CampusLens.
You can only have so many pilots and scale up pilots, and you can read how many articles that tell you and advice pieces from Gartner to McKinsey to whatever source you may choose that help you try to identify how to scale up pilots. But I wanted something that was going to help us leapfrog that, and frankly, CampusLens allows us to do that with a multi-year co-development opportunity to help us focus on the student journey, but really in a systemic way, look at all of our data sources and our use and all of even our new systems like Workday that help us to leverage a tool that sits above our data sources. We’ll learn the operational side of this as we go on. But I’m really focused on students, and this was the easiest way to take a risk at a systemic change with a trusted partner who has incredible expertise, as we’ve known for years, and our relationship with them to help us take that leap, to help us implement a system-wide approach to using AI and how that can change and enhance all the human work that we do with our students.
It’s not necessarily about eliminating the human touch to what we do. It’s about helping our advisers and our navigators and our faculty members and our intern placement officers, helping them do their work more successfully, always evaluated by student satisfaction, student placement, employer satisfaction and the like.
I appreciate pilots, there’s a great role for them. And I really appreciate that we are able to take a systemic swing at this work.
Q: You used the work risk earlier applied to taking on this system, what do you consider the risks or challenges of this process?
A: There’s always a risk in the investment you’re making initially and the ongoing risk. The risk is not only the contract for the service, hiring the expertise and hiring a partner who’s been affiliated and connected to higher education for decades, who understands students, understands institutional requirements and for compliance and integrity and data governance and permissible uses. Working with a partner that has that basic understanding is critical. That mitigates your risk immediately.
The financial risk is always: Are we chasing AI attachments to every system we have, or are we helping to right size those to be able to leverage a holistic or a system-wide, comprehensive AI-aided business analytics or business intelligence tool? That’s a very different approach then again, enhancing all of the six legacy systems I have, plus using one system or one tool to be able to do that intelligence work. That’s a risk, and that’s something different that we’ve had to navigate.
I don’t underestimate the time and challenge and excitement of staff in using technology, that can be seen as a risk. There’s a real temptation, and I see it almost daily, to just lift what we do currently in our old systems and shift it to a new system and just be satisfied with going faster or generating a nicer looking report. It’s not what we wanted and that is so not what we can do in higher ed. We are called to be more innovative and to really use our information differently. And this tool will allow us to do that in terms of really getting to the intelligence side of predictive analytics.
That’s always seen as almost a holy grail for us, and to see that it’s within reach now, that’s worth the risk for us. We’ll be able to see the analytics and the predictive analytics that we were at one time working on a project, and we thought we might get there in two or three years. The fact that I can probably do this by the end of this academic year is really important for us. And by the way, not just see some results at the end of this academic year, but know that it’s going to be iterative and evolve, so that we’re going to continue to see growth and change and adaptation and be a part of that shaping is really important to us.
I think I mean the risk is time, resources, and security, and we face those all the time. But I will dare say the risk is also not doing anything. If we aren’t moving in this direction, you are risking putting resources, and particularly too much money and technologies that you still have to reconcile in some way. You risk, frankly, burning out your staff by adding another dashboard they’ve got to learn instead of one that’s much more comprehensive. You’re still going to have them look at 10 different screens to come up with all the information they need to advise one student. So, you don’t want to burn out your staff. You actually need them to be more efficient and effective and spend time with the student in a different way.
The risks of not taking a step like this are substantial, because the world will continue. Students will still demand more, and they always demand more to make their work easier, which means our work can be a little bit harder, and employers are expecting us to be responsive. So if you don’t act and take certain risks, you’re either irrelevant or your students are unprepared for the world that they’re going to be entering, and we just don’t have time for that. That’s just not an option.
Q: I appreciated your comment on the risks of using a new tool to do the exact same thing. We know that faculty and staff are often crunched for time and ensuring that we’re creating new systems that are evolutionary and actually creating efficiencies for everyone involved is important.
A: Yeah, and that’s scary. It does mean that we will be changing how we work. It means we will be removing some of the guesswork of whether our efforts will work. We can see whether, if I move certain levers in a student journey, does this really make the difference? Does it really move the needle, not only for that student, but maybe very similarly situated students?
It’s really important. This will change how we work. We’ll be asking our employees, my colleagues, to think differently and do their work differently, because they’ll have more information available to them with suggestions on how to act, so they don’t have to always consider and frankly, reinvent the wheel. That’s really important, but I don’t underestimate what that change looks like, because when you have expertise in old systems or even evolving systems, and that expertise can be threatened or seem to be threatened, then we have to navigate that, and again, always make sure we’re serving our students and doing it the best way we can.
This technology, the tools, the guidance and the continued evolution will, I think, go a long way toward mitigating that fear. When I brought this option for CampusLens to my team, I made sure my team kicked the tires. This wasn’t a president’s folly. It was sincere, deliberate vetting by many individuals across the college to say, is this the right approach? What are our questions? What are our fears? What’s my role? Will it really better serve our students, and what does that look like with professional development? How do I use this team of experts that I’m not used to working with? How are they going to integrate and challenge us and help us do our work? So there were a number of challenges in the five or six months that it took us to ascertain whether this is the right approach for us, and I appreciate that it’s a collaborative effort, and that that is continuing as we talk about change management and the work that we have to test the tool and move it out in the college.
Q: Where are you all at in this change process? What are you looking forward to as the next step?
A: We’re still early in our stages of implementing CampusLens. Much of what we hope for centers arounds adoption and effectiveness and we really hope for a long-term operational integration. Again, my interest is not only in pilots, but in helping us make systemic change and better leveraging all the legacy data sources that we have.
What we are hoping to see in the next 12 months would be how we move from tracking legacy data and focusing on what has happened to helping us think about what is likely to happen based on the data we see. So again, shift in mindset from always reporting out past data, old data, lagging data to what do we think will happen? And then how do we change behavior to improve what we think will happen or change the trajectory, if that’s what we want to do? I think it’s really important for our community, for us to continue to test the model, the tool and the logic, so it’s going to continue to be refined. I know that as we go through over time, we will continue to improve, refine, revise the model so that it better reflects what our community here in Central New Mexico needs and what our students need.
We’re early in the stages. What I’ve seen so far is exciting, and it’s what we wanted to accomplish, and this tool is going to help us accomplish it, I think, sooner, and to be able to test our work sooner.
Trustees at Nassau Community College are poised to file a lawsuit after the State University of New York’s Board of Trustees denied their presidential pick.
At a special meeting on Sunday, the Nassau Community College Board of Trustees unanimously voted to allow the board chair to file a lawsuit challenging the SUNY board’s decision, with one board member absent, Newsdayreported. Earlier this month, SUNY trustees voted unanimously, with three members absent, to reject Maria Conzatti, who has run the college as interim or acting president for almost four years. A SUNY official told Newsday it was the first time the system’s board disapproved a presidential nominee.
The resolution voted on asked that Conzatti’s appointment by Nassau Community College’s board be “disapproved” with no further explanation.
“SUNY is committed to excellent leadership for all of our campuses and the success of our students, and we will vigorously defend ourselves against any frivolous lawsuit,” a spokesperson for the system said in a statement to Inside Higher Ed.
The college’s Student Government Association also passed a measure on Monday expressing “gratitude and appreciation” for Conzatti while also acknowledging the SUNY board vote and encouraging the college to “conduct an equitable, transparent and expeditious search for a new permanent president.”
The conflict comes amid broader tensions between the college’s faculty union and the administration over the consolidation of academic departments and a union contract that expired in August, among other issues. The union sued the college last year arguing the elimination of 15 department chairs violated state regulations, but a judge dismissed the case. The union has since appealed.
Nassau has also reported less-than-optimal student outcomes in recent years. It has the lowest two-year graduation rate and second lowest three-year graduation rate among community colleges in the SUNY system, 9.4 percent and 23.6 percent respectively.
Join HEPI for a webinar on Thursday 11 December 2025 from 10am to 11am to discuss how universities can strengthen the student voice in governance to mark the launch of our upcoming report, Rethinking the Student Voice. Sign up now tohear our speakersexplore the key questions.
This blog was kindly authored by Professor Ronald Barnett, Emeritus Professor at the Institute of Education, University of London and President Emeritus of the Philosophy and Theory of Higher Education Society, and Secretary of the Global Forum for Re-Humanizing Education.
We are faced today, especially in the UK, with a policy discourse in higher education that speaks entirely of ‘skills’ and an academic discourse, especially in the humanities and social theory, which speaks of ‘education’. In the skills discourse, there is typically no mention of education per se; and in the education discourse, there is no mention of skills per se.
It will be said that this is an exaggeration, to which I invite such commentators to look at the evidence. In the policy discourse, rafts of blogs, public pronouncements by politicians, and reports from think tanks speak of skills without the idea of education being even mentioned as such, let alone raised up for consideration. On the other side, whole papers in the academic literature and even books can be found that speak of education, student development, criticality, self-formation and so on, while paying only perfunctory attention to the matter of skills, if that.
On the skills side of the debate, we may observe a HEPI blog entitled ‘Bridging the Gap: How Smart Technology Can Align University Programmes with Real-World Skills’ (Pete Moss, 22 July, 2025). The term ‘skills’ appears twenty times, with an additional mention of ‘reskilling’ and the phrases ‘skills gap’ and ‘skills taxonomy’. The term ‘education’ appears just three times, with two of those instances being in the form of phrases – ‘higher education’ and ‘university education’.
Only once does the term ‘education’ appear unadorned, and that in the last line: ‘After all, education is a journey. It’s time the map caught up’. Nowhere are we treated to any indication as to the nature of the journey beyond it being the acquisition of skills. What education as such is, we are left to ponder.
This debate in higher education is not really a debate at all, but rather a situation in which ships pass in the night and without even acknowledging each other. There is an occasional – if rather perfunctory – doffing of the hat towards skills on the educational side; but pretty well a near-complete silence about education on the skills side.
Does this matter? After all, it might be suggested that what we have here is nothing more than a continuation of the polarisation of the liberal-vocational perspectives that have been with us in the United Kingdom for two hundred years or more. Nothing new here, it may be said. I disagree.
First, the intensity of this polarisation is now extreme. As remarked, characteristically, as I see it, positions are taken up of a kind that exhibits a total blankness towards the other side. As a result, there is no mutual engagement of positions.
Second, this blankness is particularly marked on the skills side, so to speak; and that is where the power lies. As a result, the framing of higher education in terms of skills becomes the dominant discourse.
Third, the skills side is not only utilitarian, but it is also instrumental. Every aspect of higher education comes to be valued insofar as it demonstrably has an outcome, and this logic is extended to students themselves. They become ends towards external purposes, now of economic, societal and national advancement. The development of students, understood as human beings, is rendered invisible.
Fourth, the world is facing great difficulties: egregious inequalities (of a like not seen for hundreds of years), crises of the natural environment, non-comprehension across peoples, violence (both material and discursive), and a degrading instrumentality in the way states treat their citizens are just indicative. What, against these horizons, might ‘higher education’ mean? Simply to speak of skills misses the point.
Lastly, the world is in difficulty partly due to its institutions of higher education losing sight of their educational responsibilities. At best, those institutions have become institutions of higher skills. In the process, universities have played a part in forging the instrumentality that is now dominant in the world. That the world is in grave difficulty can be laid, in part, at the door of universities.
What, then, is to be done? The answer is obvious. We need a rebalancing in our debates, our language, our practices, our evaluation mechanisms, and the ways in which we identify what is of value in higher education. It is right for skills – and knowledge too, for that matter – to have a place, but that place has to be against the horizon of what is good for the education of students as human beings on and in this troubled Earth.
But this rebalancing calls for a reordering, where concerns with education have to precede concerns with skills. Wisdom, critical reflection, dialogue for understanding, care, consideration, carefulness, self-understanding, the world, Nature, dispute, antagonism, and mutuality have to become part of the vocabulary of student formation in constructing a proper policy debate. Unless and until this happens, the policy framework will be blind and surrender itself into the interests and technologies of the powerful.
Artificial Intelligence (AI) is driving one of the most significant transformations in academic publishing since the advent of peer review. There has been a steady increase in the use of AI in research manuscripts since 2022, when OpenAI launched ChatGPT. Recent data demonstrate that up to 22% of computer science papers indicate the usage of large language models (LLMs). Another study on LLM use in scientific publishing revealed that 16.9% of peer review texts contained AI content.
The integration of AI tools in every stage of publication has been met with appreciation and criticism in equal measure. Proponents advocate AI as a catalyst that boosts efficiency by automating mundane tasks such as grammar editing, reference formatting, and initial screening. However, critics warn against possible pitfalls in quality and ethical considerations. The question remains: is AI disrupting traditional publishing models, or is it fostering a natural evolution within the academic enterprise?
Rise of AI in Research and Publication
AI has become embedded throughout the research lifecycle, from idea generation to final manuscript submission. This pervasive adoption is reshaping researchers’ approach to their work; they use applications for grammar checks, plagiarism detection, format compliance, and even assessment of their research significance.
Recent studies show that LLMs have an unmatched ability to generate prose and summarize content, enabling researchers to write literature reviews and experimental descriptions more efficiently. In fact, many researchers utilize features of LLMs to assist them with brainstorming, rephrasing, and clarifying their arguments.
Besides writing assistance, AI-driven platforms can streamline other laborious tasks. For instance, searching massive databases to find relevant citations and assessing the context and reliability of references. Likewise, available text translation and summarization tools support 30+ languages— a feature that breaks down linguistic barriers and allows for international collaboration in research.
One perspective piece published in PLOS Biology notes: “A primary reason that science has not yet become fully multilingual is that translation can be slow and costly. Artificial intelligence (AI), however, may finally allow us to overcome this problem, as it can provide useful, often free or affordable, support in language editing and translation”.
The integration of AI tools in research and publication has fundamentally shifted research workflows. Automation technologies have quickened the pace of writing and made academic writing and publishing more accessible.
Ethical Implications and Risks
The increased efficiency and renewed research approach also come with significant ethical concerns. AI systems operate by learning patterns in existing data and can amplify hidden biases. If an AI tool is trained on past editorial outcomes, it may score submissions from well-known institutions or English-speaking authors more highly.
Stanford researchers found that even when non-native English researchers use LLMs to refine their submissions, peer reviewers still show bias against them. Another study revealed that AI text-detection tools often misidentify non-native English writing as machine-generated. In other words, if there are two authors with equal merits, they may face unequal scrutiny if one uses slightly different phrasing.
Beyond systemic biases, there is concern that AI integration into the peer review process can lead to over-reliance on its functionalities. An analysis of peer review highlighted that overdependence of editors and reviewers on AI-generated suggestions can result in deterioration in review quality, and not to mention, factual mistakes that slip through evaluation.
Even as AI is used for initial screening, the recommendations made by such tools about a manuscript’s quality or reviewer selection may be opaque or even include hallucinated reasoning. In the absence of transparency, it is difficult to identify and correct misjudgments. This accountability challenge can seriously undermine trust in the editorial process.
However, the most concerning risk is the potential of AI to create a feedback loop of maintaining the status quo. AIs are trained on existing published literature and are designed to evaluate new submissions based on the present data. Given this data, it is likely that the AI system may inadvertently suppress new and innovative ideas that do not coincide with established patterns.
The Irreplaceable Role of Human Editorial Judgment
Despite these technological advances, the fundamental responsibility for maintaining scientific integrity ultimately rests with human editors and reviewers. Academic publishers serve as gatekeepers of knowledge, shaping what research reaches the broader scientific community and, by extension, informing public understanding and policy decisions. This role carries immense responsibility. Editorial decisions can accelerate breakthrough discoveries or inadvertently stifle groundbreaking research that challenges conventional thinking.
Rather than delegating these critical judgements to algorithms, the academic publishing community must recognize this moment as a call to elevate editorial standards and practices. Editors must recommit to rigorous, nuanced evaluation that prioritizes scientific merit over efficiency metrics. The stakes are too high– the advancement of human knowledge and credibility of scientific enterprise itself– to entrust these decisions to systems that, however sophisticated, lack the contextual understanding, ethical reasoning, and innovative thinking that human expertise provides.
Dr. Su Yeong Kim is a Professor of Human Development and Family Sciences at the University of Texas at Austin. A leading figure in research on immigrant families and adolescent development, she earned the honor of being a Fellow on multiple national psychology associations and is the Editor of the Journal of Research on Adolescence. Dr. Kim has authored and published more than 160 works and is recognized with Division 45 of American Psychological Association’s Distinguished Career Contributions to Research Award. Her research, funded by the National Institutes of Health, National Science Foundation, and other bodies, covers bilingualism, language brokering, and cultural stressors in immigrant-origin youth. Dr. Kim is an enthusiastic mentor and community advocate, as well as a member of UT’s Provost’s Distinguished Leadership Service Academy.
Kim, H., J. C. Little, J. Li, B. Patel, and D. Kalderon. “Hedgehog-Stimulated Phosphorylation at Multiple Sites Activates Ci by Altering Ci–Ci Interfaces without Full Suppressor of Fused Dissociation.” PLOS Biology 23, no. 4 (2025): e3003105. https://doi.org/10.1371/journal.pbio.3003105