Tag: Disruptions

  • Data, Decisions, and Disruptions: Inside the World of University Rankings

    Data, Decisions, and Disruptions: Inside the World of University Rankings

    University rankings are pretty much everywhere. Though the earliest university rankings in the U. S. date back to the early 1900s and the modern ones from the 1983 debut of the U. S. News and World Report rankings. The kind of rankings we tend to talk about now, international or global rankings, really only date back to 2003 with the creation of the Shanghai Academic Rankings of World Universities.

    Over the decade that followed that first publication, a triumvirate emerged at the top of the rankings pyramid. The Shanghai Rankings, run by a group of academics at the Shanghai Jiao Tong University, the Quacquarelli Symonds, or QS Rankings, and the Times Higher Education’s World University Rankings. Between them, these three rankings producers, particularly QS and Times Higher, created a bewildering array of new rankings, dividing the world up by geography and field of study, mainly based on metrics relating to research.

    Joining me today is the former Chief Data Officer of the Times Higher Education Rankings, Duncan Ross. He took over those rankings at a time when it seemed like the higher education world might be running out of things to rank. Under his tutelage, though, the Times Impact Rankings, which are based around the 17 UN Sustainable Development Goals, were developed. And that’s created a genuinely new hierarchy in world higher education, at least among those institutions who choose to submit to the rankings.  

    My discussion with Duncan today covers a wide range of topics related to his time at THE. But the most enjoyable bit by far, for me anything, was the bit about the genesis of the impact rankings. Listen a bit, especially when Duncan talks about how the Impact Rankings came about because the THE realized that its industry rankings weren’t very reliable. Fun fact, around that time I got into a very public debate with Phil Beatty, the editor of the Times Higher, on exactly that subject. Which means maybe, just maybe, I’m kind of a godparent to the impact rankings. But that’s just me. You may well find other points of interest in this very compelling interview. Let’s hand things over to Duncan.


    The World of Higher Education Podcast
    Episode 3.20 | Data, Decisions, and Disruptions: Inside the World of University Rankings 

    Transcript

    Alex Usher: So, Duncan, let’s start at the beginning. I’m curious—what got you into university rankings in the first place? How did you end up at Times Higher Education in 2015?

    Duncan Ross: I think it was almost by chance. I had been working in the tech sector for a large data warehousing company, which meant I was working across many industries—almost every industry except higher education. I was looking for a new challenge, something completely different. Then a friend approached me and mentioned a role that might interest me. So I started talking to Times Higher Education, and it turned out it really was a great fit.

    Alex Usher: So when you arrived at Times Higher in 2015, the company already had a pretty full set of rankings products, right? They had the global rankings, the regional rankings, which I think started around 2010, and then the subject or field of study rankings came a couple of years later. When you looked at all of that, what did you think? What did you feel needed to be improved?

    Duncan Ross: Well, the first thing I had to do was actually bring all of that production in-house. At the time, even though Times Higher had rankings, they were produced by Clarivate—well, Thomson Reuters, as it was then. They were doing a perfectly good job, but if you’re not in control of the data yourself, there’s a limit to what you can do with it.

    Another key issue was that, while it looked like Times Higher had many rankings, in reality, they had just one: the World University Rankings. The other rankings were simply different cuts of that same data. And even within the World University Rankings, only 400 universities were included, with a strong bias toward Europe and North America. About 26 or 27 percent of those institutions were from the U.S., which didn’t truly reflect the global landscape of higher education.

    So the challenge was: how could we broaden our scope and truly capture the world of higher education beyond the usual suspects? And beyond that, were there other aspects of universities that we could measure, rather than just relying on research-centered metrics? There are good reasons why international rankings tend to focus on research—it’s the most consistent data available—but as you know, it’s certainly not the only way to define excellence in higher education.

    Alex Usher: Oh, yeah. So how did you address the issue of geographic diversity? Was it as simple as saying, “We’re not going to limit it to 400 universities—we’re going to expand it”? I think the ranking now includes over a thousand institutions, right? I’ve forgotten the exact number.

    Duncan Ross: It’s actually around 2,100 or so, and in practice, the number is even larger because, about two years ago, we introduced the concept of reporter institutions. These are institutions that haven’t yet met the criteria to be fully ranked but are already providing data.

    The World University Rankings have an artificial limit because there’s a threshold for participation based on the number of research articles published. That threshold is set at 1,000 papers over a five-year period. If we look at how many universities could potentially meet that criterion, it’s probably around 3,000, and that number keeps growing. But even that is just a fraction of the higher education institutions worldwide. There are likely 30,000—maybe even 40,000—higher education institutions globally, and that’s before we even consider community colleges.

    So, expanding the rankings was about removing artificial boundaries. We needed to reach out to institutions in parts of the world that weren’t well represented and think about higher education in a way that wasn’t so Anglo-centric.

    One of the biggest challenges I’ve encountered—and it’s something people inevitably fall into—is that we tend to view higher education through the lens of our own experiences. But higher education doesn’t function the same way everywhere. It’s easy to assume that all universities should look like those in Canada, the U.S., or the UK—but that’s simply not the case.

    To improve the rankings, we had to be open-minded, engage with institutions globally, and carefully navigate the challenges of collecting data on such a large scale. As a result, Times Higher Education now has data on around 5,000 to 6,000 universities—a huge step up from the original 400. Still, it’s just a fraction of the institutions that exist worldwide.

    Alex Usher: Well, that’s exactly the mission of this podcast—to get people to think beyond an Anglo-centric view of the world. So I take your point that, in your first couple of years at Times Higher Education, most of what you were doing was working with a single set of data and slicing it in different ways.

    But even with that, collecting data for rankings isn’t simple, right? It’s tricky, you have to make a lot of decisions, especially about inclusion—what to include and how to weight different factors. And I think you’ve had to deal with a couple of major issues over the years—one in your first few years and another more recently.

    One was about fractional counting of articles, which I remember went on for quite a while. There was that big surge of CERN-related articles, mostly coming out of Switzerland but with thousands of authors from around the world, which affected the weighting. That led to a move toward fractional weighting, which in theory equalized things a bit—but not everyone agreed.

    More recently, you’ve had an issue with voting, right? What I think was called a cartel of voters in the Middle East, related to the reputation rankings. Can you talk a bit about how you handle these kinds of challenges?

    Duncan Ross: Well, I think the starting point is that we’re always trying to evaluate things in a fair and consistent way. But inevitably, we’re dealing with a very noisy and messy world.

    The two cases you mentioned are actually quite different. One is about adjusting to the norms of the higher education sector, particularly in publishing. A lot of academics, especially those working within a single discipline, assume that publishing works the same way across all fields—that you can create a universal set of rules that apply to everyone. But that’s simply not the case.

    For example, the concept of a first author doesn’t exist in every discipline. Likewise, in some fields, the principal investigator (PI) is always listed at the end of the author list, while in others, that’s not the norm.

    One of the biggest challenges we faced was in fields dealing with big science—large-scale research projects involving hundreds or even thousands of contributors. In high-energy physics, for example, a decision was made back in the 1920s: everyone who participates in an experiment above a certain threshold is listed as an author in alphabetical order. They even have a committee to determine who meets that threshold—because, of course, it’s academia, so there has to be a committee.

    But when you have 5,000 authors on a single paper, that distorts the rankings. So we had to develop a mechanism to handle that. Ideally, we’d have a single metric that works in all cases—just like in physics, where we don’t use one model of gravity in some situations and a different one in others. But sometimes, you have to make exceptions. Now, Times Higher Education is moving toward more sophisticated bibliometric measures to address these challenges in a better way.

    The second issue you mentioned—the voting behavior in reputation rankings—is completely different because it involves inappropriate behavior. And this kind of issue isn’t just institutional; sometimes, it’s at the individual academic level.

    We’re seeing this in publishing as well, where some academics are somehow producing over 200 articles a year. Impressive productivity, sure—but is it actually viable? In cases like this, the approach has to be different. It’s about identifying and penalizing misbehavior.

    At the same time, we don’t want to be judge and jury. It’s difficult because, often, we can see statistical patterns that strongly suggest something is happening, but we don’t always have a smoking gun. So our goal is always to be as fair and equitable as possible while putting safeguards in place to maintain the integrity of the rankings.

    Alex Usher: Duncan, you hinted at this earlier, but I want to turn now to the Impact Rankings. This was the big initiative you introduced at Times Higher Education. Tell us about the genesis of those rankings—where did the idea come from? Why focus on impact? And why the SDGs?

    Duncan Ross: It actually didn’t start out as a sustainability-focused project. The idea came from my colleague, Phil Baty, who had always been concerned that the World University Rankings didn’t include enough measurement around technology transfer.

    So, we set out to collect data from universities on that—looking at things like income from consultancy and university spin-offs. But when the data came back, it was a complete mess—totally inconsistent and fundamentally unusable. So, I had to go back to the drawing board.

    That’s when I came across SDG 9—Industry, Innovation, and Infrastructure. I looked at it and thought, This is interesting. It was compelling because it provided an external framework.

    One of the challenges with ranking models is that people always question them—Is this really a good model for excellence? But with an external framework like the SDGs, if someone challenges it, I can just point to the United Nations and say, Take it up with them.

    At that point, I had done some data science work and was familiar with the tank problem, so I jokingly assumed there were probably 13 to 18 SDGs out there. (That’s a data science joke—those don’t land well 99% of the time.) But as it turned out, there were more SDGs, and exploring them was a real light bulb moment.

    The SDGs provided a powerful framework for understanding the most positive role universities can play in the world today. We all know—well, at least those of us outside the U.S. know—that we’re facing a climate catastrophe. Higher education has a crucial role to play in addressing it.

    So, the question became: How can we support that? How can we measure it? How can we encourage better behavior in this incredibly important sector?

    Alex Usher: The Impact Rankings are very different in that roughly half of the indicators—about 240 to 250 across all 17 SDGs—aren’t naturally quantifiable. Instead, they’re based on stories.

    For example, an institution might submit, This is how we combat organized crime or This is how we ensure our food sourcing is organic. These responses are scored based on institutional submissions.

    Now, I don’t know exactly how Times Higher Education evaluates them, but there has to be a system in place. How do you ensure that these institutional answers—maybe 120 to 130 per institution at most—are scored fairly and consistently when you’re dealing with hundreds of institutions?

    Duncan Ross: Well, I can tell you that this year, over 2,500 institutions submitted approved data—so it’s grown significantly. One thing to clarify, though, is that these aren’t written-up reports like the UK’s Teaching Excellence Framework, where universities can submit an essay justifying why they didn’t score as well as expected—what I like to call the dog ate my student statistics paper excuse. Instead, we ask for evidence of the work institutions have done. That evidence can take different forms—sometimes policies, sometimes procedures, sometimes concrete examples of their initiatives. The scoring process itself is relatively straightforward. First, we give some credit if an institution says they’re doing something. Then, we assess the evidence they provide to determine whether it actually supports their claim. But the third and most important part is that institutions receive extra credit if the evidence is publicly available. If you publish your policies or reports, you open yourself up to scrutiny, which adds accountability.

    A great example is SDG 5—Gender Equality—specifically around gender pay equity. If an institution claims to have a policy on gender pay equity, we check: Do you publish it? If so, and you’re not actually living up to it, I’d hope—and expect—that women within the institution will challenge you on it. That’s part of the balancing mechanism in this process.

    Now, how do we evaluate all this? Until this year, we relied on a team of assessors. We brought in people, trained them, supported them with our regular staff, and implemented a layer of checks—such as cross-referencing responses against previous years. Ultimately, human assessors were making the decisions.

    This year, as you might expect, we’re introducing AI to assist with the process. AI helps us filter out straightforward cases, leaving the more complex ones for human assessors. It also ensures that we don’t run into assessor fatigue. When someone has reviewed 15 different answers to the same question from various universities, the process can get a bit tedious—AI helps mitigate that.

    Alex Usher: Yeah, it’s like that experiment with Israeli judges, right? You don’t want to be the last case before lunch—you get a much harsher sentence if the judge is making decisions on an empty stomach. I imagine you must have similar issues to deal with in rankings.

    I’ve been really impressed by how enthusiastically institutions have embraced the Impact Rankings. Canadian universities, in particular, have really taken to them. I think we had four of the top ten last year and three of the top ten this year, which is rare for us. But the uptake hasn’t been as strong—at least not yet—in China or the United States, which are arguably the two biggest national players in research-based university rankings. Maybe that’s changing this year, but why do you think the reception has been so different in different parts of the world? And what does that say about how different regions view the purpose of universities?

    Duncan Ross: I think there’s definitely a case that different countries and regions have different approaches to the SDGs. In China, as you might expect, interest in the rankings depends on how well they align with current Communist Party priorities. You could argue that something similar happens in the U.S. The incoming administration has made it fairly clear that SDG 10 (Reduced Inequalities) and SDG 5 (Gender Equality) are not going to be top priorities—probably not SDG 1 (No Poverty), either. So in some cases, a country’s level of engagement reflects its political landscape.

    But sometimes, it also reflects the economic structure of the higher education system itself. In the U.S., where universities rely heavily on high tuition fees, rankings are all about attracting students. And the dominant ranking in that market is U.S. News & World Report—the 600-pound gorilla. If I were in their position, I’d focus on that, too, because it’s the ranking that brings in applications.

    In other parts of the world, though, rankings serve a different purpose. This ties back to our earlier discussion about different priorities in different regions. Take Indonesia, for example. There are over 4,000 universities in the country. If you’re an institution like ITS (Institut Teknologi Sepuluh Nopember), how do you stand out? How do you show that you’re different from other universities?

    For them, the Impact Rankings provided an opportunity to showcase the important work they’re doing—work that might not have been recognized in traditional rankings. And that’s something I’m particularly proud of with the Impact Rankings. Unlike the World University Rankings or the Teaching Rankings, it’s not just the usual suspects at the top.

    One of my favorite examples is Western Sydney University. It’s a fantastic institution. If you’re ever in Sydney, take the train out there. Stay on the train—it’s a long way from the city center—but go visit them. Look at the incredible work they’re doing, not just in sustainability but also in their engagement with Aboriginal and Torres Strait Islander communities. They’re making a real impact, and I’m so pleased that we’ve been able to raise the profile of institutions like Western Sydney—universities that might not otherwise get the recognition they truly deserve.

    Alex Usher: But you’re still left with the problem that many institutions that do really well in research rankings have, in effect, boycotted the Impact Rankings—simply because they’re not guaranteed to come first.

    A lot of them seem to take the attitude of, Why would I participate in a ranking if I don’t know I’ll be at the top?

    I know you initially faced that issue with LERU (the League of European Research Universities), and I guess the U.S. is still a challenge, with lower participation numbers.

    Do you think Times Higher Education will eventually crack that? It’s a tough nut to crack. I mean, even the OECD ran into the same resistance—it was the same people saying, Rankings are terrible, and we don’t want better ones.

    What’s your take on that?

    Duncan Ross: Well, I’ve got a brief anecdote about this whole rankings boycott approach. There’s one university—I’m not going to name them—that made a very public statement about withdrawing from the Times Higher Education World University Rankings. And just to be clear, that’s something you can do, because participation is voluntary—not all rankings are. So, they made this big announcement about pulling out. Then, about a month later, we got an email from their graduate studies department asking, Can we get a copy of your rankings? We use them to evaluate applicants for interviews. So, there’s definitely some odd thinking at play here. But when it comes to the Impact Rankings, I’m pretty relaxed about it. Sure, it would be nice to have Oxford or Harvard participate—but MIT does, and they’re a reasonably good school, I hear. Spiderman applied there, so it’s got to be decent. The way I see it, the so-called top universities already have plenty of rankings they can focus on. If we say there are 300 top universities in the world, what about the other 36,000 institutions?

    Alex Usher: I just want to end on a slightly different note. While doing some background research for this interview, I came across your involvement in DataKind—a data charity that, if I understand correctly, you founded. I’ve never heard of a data charity before, and I find the idea fascinating—intriguing enough that I’m even thinking about starting one here. Tell us about DataKind—what does it do?

    Duncan Ross: Thank you! So, DataKind was actually founded in the U.S. by Jake Porway. I first came across it at one of the early big data conferences—O’Reilly’s Strata Conference in New York. Jake was talking about how data could be used for good, and at the time, I had been involved in leadership roles at several UK charities. It was a light bulb moment. I went up to Jake and said, Let me start a UK equivalent! At first, he was noncommittal—he said, Yeah, sure… someday. But I just kept nagging him until eventually, he gave in and said yes. Together with an amazing group of people in the UK—Fran Bennett, Caitlin Thaney, and Stuart Townsend—we set up DataKind UK.

    The concept is simple: we often talk about how businesses—whether in telecom, retail, or finance—use data to operate more effectively. The same is true in the nonprofit sector. The difference is that banks can afford to hire data scientists—charities often can’t. So, DataKind was created to connect data scientists with nonprofit organizations, allowing them to volunteer their skills.

    Of course, for this to work, a charity needs a few things:

    1. Leadership willing to embrace data-driven decision-making.
    2. A well-defined problem that can be analyzed.
    3. Access to data—because without data, we can’t do much.

    Over the years, DataKind—both in the U.S. and worldwide—has done incredible work. We’ve helped nonprofits understand what their data is telling them, improve their use of resources, and ultimately, do more for the communities they serve. I stepped down from DataKind UK in 2020 because I believe that the true test of something successful is whether it can continue to thrive without you. And I’m happy to say it’s still going strong. I kind of hope the Impact Rankings continue to thrive at Times Higher Education now that I’ve moved on as well.

    Alex Usher: Yeah. Well, thank you for joining us today, Duncan.

    Duncan Ross: It’s been a pleasure.

    And it just remains for me to thank our excellent producers, Sam Pufek and Tiffany MacLennan. And you, our viewers, listeners, and readers for joining us today. If you have any questions or comments about today’s episode, please don’t hesitate to get in touch with us at podcast@higheredstrategy.com. Worried about missing an episode of the World of Higher Education? There’s a solution for that. Go to our YouTube page and subscribe. Next week, our guest will be Jim Dickinson. He’s an associate editor at Wonkhe in the UK, and he’s also maybe the world expert on comparative student politics. And he joins us to talk about the events in Serbia where the student movement is challenging the populist government of the day. Bye for now.

    *This podcast transcript was generated using an AI transcription service with limited editing. Please forgive any errors made through this service.

    Source link

  • 2025 Higher Ed Disruptions | Collegis Education

    2025 Higher Ed Disruptions | Collegis Education

    The coming year promises to be transformative for higher education as institutions find new ways to manage enrollment targets, operating costs, and shifting student expectations. Several existing and emerging trends have the potential to alter the higher ed landscape as we know it in 2025.

    Disruption isn’t necessarily a bad thing. It often leads to innovation and more efficient ways to meet the needs of students, faculty, and administrators. The good news is we have the tools and the know-how to address these challenges head-on. Institutions focused on building foundational capabilities in the coming months are best positioned to leverage technology effectively and position themselves for continued success.

    Here are the trends I predict will significantly impact higher ed this year and what we can do to take advantage of them.

    1. More Urgent Digital Transformation Plans
    Institutions need access to valid, reliable, and meaningful data to operate effectively. Thousands of schools still rely on proprietary, on-premise student information systems (SIS) with fragmented data sources, which limits their ability to make data-enabled decisions. Given that migrating to cloud-based solutions can take 18+ months, it’s important for schools to start the process now. Although the process is difficult given the significant change management associated with large cloud migrations, it will enable them to operate more efficiently and compete more effectively. I anticipate we’ll see SIS cloud migrations at the top of the priority lists at many institutions.

    2. Heightened Focus on Cybersecurity
    Cyber attackers have targeted higher education for years because they know they are a rich source of student and institutional data and the digital infrastructures at most schools are outdated. Hackers continue to find new ways to access networks and data, especially as the number of connected devices and applications swells. Institutions need to stay vigilant to cyber threats while also complying with various data privacy laws. I count 18 states with privacy regulations in addition to U.S. federal and European Union requirements. It’s an extremely complex situation only made more difficult by the shortage of cybersecurity professionals, especially at smaller schools. In 2025, institutions will focus on automating network security protocols and finding outside resources to augment their security capabilities.

    3. Expanded Use of Shared Services
    Smaller schools need access to the same technology and technical expertise as larger schools, just on a reduced scale with a more limited budget. I anticipate that smaller schools will seek strategic partners to manage critical IT and other specialized services to support data access, reliability, and usability. It’s a smart way to reduce costs while maintaining essential day-to-day services, enhancing security protocols, and being prepared for technology advancements.

    4. Continuation of Mergers and Acquisitions
    Financial pressures and the threat of closures will continue to drive acquisitions of some smaller schools. Mergers create significant challenges to combine the data, applications, and systems of the two institutions. However, the benefits to both sides are worth it in streamlining operations, retaining existing students, and growing enrollment. Given the breadth and depth of our functional and technical expertise, we can help schools to navigate the challenges and drive positive results.

    5. Adoption of Data-Enabled Technologies
    As institutions embrace digital transformation, they build out the foundation needed to take advantage of data-enabled technologies such as artificial intelligence (AI). With the initial hype of AI behind us, I anticipate in 2025 that schools will start to deploy AI-enabled solutions that feed off rich data sources to personalize recruitment efforts, improve retention, invigorate online classes, and predict demand for course offerings, as a few examples.

    6. Refinement of Enrollment and Retention Strategies
    The declining number of high school graduates and international enrollments continues to intensify the effort to recruit and retain students. I expect to see schools expand how they leverage data to personalize recruitment campaigns, target secondary audiences (such as adult learners and transfer students), and nurture existing students. The success of this strategy relies heavily on having the right infrastructure in place to support centralized data access, emerging technologies, and analytic tools.

    7. Shift to Career-Focused and Flexible Learning Opportunities
    Like changing recruitment and retention methods, institutions will continue to adapt their academic programs to meet the evolving needs of the workforce. They will highlight outcomes based on data about job placement rates, alumni success stories, and collaborations with area businesses to illustrate tangible benefits. They will evaluate current course offerings and pivot when it makes sense to expand hybrid learning models, professional development programs, and skill-based credentials. This effort may involve launching new programs in high-demand fields, such as data science and cybersecurity, or retooling existing programs to incorporate emerging technologies.

    8. Consolidation of Ed-tech Solution Providers
    Many higher education software and service providers have discussed how they could combine forces to serve colleges and universities better. I think we’ll see M&A activity pick up over the next two years. These consolidations will further the need for institutions to standardize business processes and accelerate cloud migrations as legacy systems will eventually become unsupported.

    I’m excited about the positive impact these eight trends might make in 2025. Higher education institutions that prioritize foundational improvements by aligning their data, technology, and talent are best suited to successfully address mounting challenges like demographic shifts and affordability concerns.

    I can’t wait to see how schools that adopt AI and predictive analytics are able to improve decision-making and enhance student experiences. However, the real breakthrough will come from integrating systems and breaking down data silos. Institutions that invest in building these foundational capabilities will be better positioned to leverage emerging technologies, drive measurable outcomes, and fulfill their mission to support lifelong learners.

    — Kim Fahey, CEO Collegis Education

    Source link

  • 2025 Higher Ed Disruptions | Collegis Education

    2025 Higher Ed Disruptions | Collegis Education

    As 2024 draws to a close, the holiday season inspires gratitude and reflection. Personally, I’m very grateful for the incredible partners and colleagues I’ve had the privilege to work with this year. Together, we accomplished so much.

    • We collaborated with our partners throughout the year to deliver great experiences for their students, alumni, and staff.
    • We collectively navigated some of higher education’s biggest challenges, driving partner growth and enabling impact.
    • We pushed the boundaries of innovation, embracing the power of data-enabled technologies.

    I’m so proud of the positive impact the Collegis team generated with our partners across the entire student lifecycle, from the moment prospective students first inquire about a program to the day they graduate.

    Let’s look back at some of 2024’s meaningful results

    Recruitment and Enrollment Growth

    We supported double-digit year-over-year (YoY) enrollment growth –– as high as 57% –– for many partner institutions in first-year, program-specific, transfer, and graduate populations. Engagement from our enrollment teams was instrumental in connecting students with the right programs and guiding them through the admissions process.

    IT Managed Services and Student Support

    Our IT team ensured seamless operations, providing reliable technology solutions that empower students and faculty. Some of my favorite examples from 2024 include:

    • Integrating systems to drive process improvement across enrollment, financial aid, academics, and career services.
    • Modernized campus infrastructures and networks to drive student engagement at a college’s main hall.
    • Significantly improved student experience by implementing a user-friendly, single sign-on (SSO) solution across student-facing systems.
    • Led an institution through a critical component of its digital transformation journey by migrating its on-premise, legacy ERP to a cloud-based, next-generation solution.

    Innovative Learning Experiences

    Our instructional design team enabled partners to grow their online course offerings on platforms such as Brightspace, Canvas, Coursera, and Blackboard Ultra, including course and online library development, course migrations, maintenance, faculty support, and term start/end deployment activities.

    We collaborated with the nursing program at one partner to revamp the entire library of online courses to meet new accreditation standards. Another partner was able to add 200 online courses to fill the needs of 13 online programs at three schools.

    Marketing Impact

    Our web team conducted a user survey and other research to refine a partner’s website, which increased clicks to inquire by 82%, the request for information (RFI) click rate to 71%, and clicks to apply by 7.5%.

    Another shining example was uncovering a way to target a healthcare provider’s employees who are eligible for a tuition discount. Because of healthcare regulations, the partner could not provide an audience list, so Collegis addressed this niche audience using in-platform targeting tools available on social media platforms. The return on ad spend (ROAS) is 2.2:1 overall in 2024 with plans to expand the program next year.

    Student Success

    Our student support team provided essential services to help students thrive and continue to pursue their academic goals.

    • At a public, four-year institution in Ohio, Collegis Student Success Coaches helped new students with the registration process, driving admit-to-enroll numbers and YOY growth of +66% in Fall 2024.
    • At a private, four-year institution in Texas, Fall retention was 97%, with a 90% retention rate since the partnership launched.
    • At a private, four-year institution in New York, term-over-term retention from Summer to Fall is 96%, with a 91% average retention rate since the partnership launched.
    • At a public, two-year institution in the Pacific Northwest, Collegis helped drive the sixth consecutive term of enrollment growth, with Fall enrollment trending toward +8%.

    Research and Portfolio Planning

    Because we are ingrained in every step of the student lifecycle, partners often ask us to assist with forward-looking strategies. For example, our team helped a partner understand the pros and cons of expanding their full-time Accelerated Bachelor of Science in Nursing (ABSN) with a part-time program. With our marketplace analysis, recommendations for how to offer courses, and a marketing launch plan, the institution is currently accepting applications for Summer 2025.

    Another institution asked for Collegis’s assistance to develop a multi-year strategic approach to graduate enrollments. The partner’s team lead noted that, “[Collegis] led productive brainstorming and strategic planning sessions with the team. Their deep knowledge of graduate enrollment trends, market analysis for graduate programs and expertise in leading our team from conceptualization to the delivery of specific recommendations on our next steps were invaluable.”

    Strategic Innovation and Workshop Design

    Our strategy and solutions team helped colleges and universities unpack complex problems and find innovative, human-centered solutions. We architected and facilitated numerous design thinking workshops, guiding leadership teams through critical strategic discussions about the future of their institutions. I’ll let some of the participants of the workshops explain the value they got out of the sessions:

    • “Collegis didn’t just help us evaluate our processes — they led us on a journey to uncover areas of improvement we hadn’t even considered. Their expert guidance illuminated the path forward, empowering us to create a more positive, streamlined, and truly enjoyable student experience.”
    • “My team and I were thoroughly impressed with your ability to take what essentially was a speck of an idea and collaboratively ideate possibilities for [the university] to offer new academic programs and training to underserved high school populations.”
    • “Collegis took the time to meet with leadership prior to the sessions and came prepared to tackle the challenges at the college. The activities were well thought-out and allowed individuals time to really think about the core issues. Thanks to Collegis, I am hopeful that our college can make key changes that will benefit our student experience and lighten our faculty/staff workload.”

    Looking Forward

    As you can see, 2024 has been a year of growth, innovation, and collaboration. We are grateful for the opportunity to work with our partners and look forward to even greater achievements in the years to come.

    I’d like to extend my sincere gratitude to my Collegis colleagues, who amaze me with their creativity, expertise, and dedication to delivering exceptional results. I can’t wait to see what you do in 2025 to continue inspiring each other and driving growth for our partners.

    Happy Holidays and best wishes for a prosperous New Year!

    — Kim Fahey, CEO Collegis Education

    Source link