administrators

B Lab Releases Standards for Colleges

B Lab is a nonprofit group that issues a seal of approval to companies across 120 industries that adhere to voluntary standards based on social and environmental performance, accountability and transparency. After a two years of work, the group on Friday released a new benchmarking tool for colleges. The voluntary standards are designed to enable comparisons of both nonprofit and for-profit institutions.

"B Lab recognizes that the cost and outcomes of higher education, particularly regarding for-profit institutions, have become increasingly controversial, but regardless of structure institutions should put their students’ needs first," Dan Osusky, standards development manager at B Lab, said in a written statement. "We see our role as the promoter of robust standards of industry-specific performance that can be used by for-profits and nonprofits alike to create the greatest possible positive impact and serve the public interest, ultimately by improving the lives of their students."

A committee of experts, working with HCM Strategists and with funding from the Lumina Foundation, devised the standards. Laureate Education, a global for-profit chain, already uses the assessment tool.

Man With Machete Killed by Police at Boulder

A man with a machete was shot and killed by police officers when he was wielding the weapon in the sports medicine facility of the University of Colorado at Boulder, The Denver Post reported. “Given the weapon the suspect was armed with, given the statement already made to our initial victim and given the nature of how he was maneuvering through the Champions Center, we believe it was in the best interest of the university that it was a deadly force situation,” said the university's campus police chief, Melissa Za, at a news conference.

Ad keywords: 

An evaluation of whether performance funding in higher education works (essay)

More than 30 states now provide performance funding for higher education, with several more states seriously considering it. Under PF, state funding for higher education is not based on enrollments and prior-year funding levels. Rather, it is tied directly to institutional performance on such metrics as student retention, credit accrual, degree completion and job placement. The amount of state funding tied to performance indicators ranges from less than 1 percent in Illinois to as much as 80 to 90 percent in Ohio and Tennessee.

Performance funding has received strong endorsements from federal and state elected officials and influential public policy groups and educational foundations. The U.S. Department of Education has urged states to “embrace performance-based funding of higher education based on progress toward completion and other quality goals.” And a report by the National Governors Association declared, “Currently, the prevailing approach for funding public colleges and universities … gives colleges and universities little incentive to focus on retaining and graduating students or meeting state needs …. Performance funding instead provides financial incentives for graduating students and meeting state needs.”

But with all this state activity and national support, does performance funding actually work? As we report in a book being published this week, Performance Funding for Higher Education (Johns Hopkins University Press), the answer is both yes and no.

Based on extensive research we conducted in three states with much-discussed performance funding programs -- Indiana, Ohio, and Tennessee -- we find evidence for the claims of both those who champion performance funding and those who reject it. In keeping with the arguments of PF champions, we find that performance funding has resulted in institutions making changes to their policies and programs to improve student outcomes -- whether by revamping developmental education or altering advising and counseling services.

Underpinning those changes have been increased institutional efforts to gather data on their performance and to change their institutional practices in response.

But we often cannot clearly determine to what degree performance funding is driving those changes. Many of the colleges we studied stated they were already committed to improving student outcomes before the advent of performance funding. Moreover, in addition to PF, the states often are simultaneously pursuing other policies -- such as initiatives to improve developmental education or establish better student pathways into and through higher education -- that push institutions in the same direction as their PF programs. As a result, it is nearly impossible to determine the distinct contribution of PF to many of those institutional changes.

Meanwhile, supporting the arguments of the PF detractors, we have not found conclusive evidence that performance funding results in significant improvements in student outcomes -- and, in fact, we’ve discovered that it produces substantial negative side effects. In reviewing the research literature on PF impacts, we find that careful multivariate studies -- which compare states with and without performance funding and control for a host of factors besides PF that influence student outcomes -- largely fail to find a significant positive impact of performance funding on student retention and degree attainment. Those studies do find some evidence of effects on four-year college graduation and community college certificates and associate degrees in some states and some years. However, those results are too scattered to allow anyone to conclude that performance funding is having a substantial impact on student outcomes.

Various organizational obstacles may help explain that lack of effect. Many institutions enroll numerous students who are not well prepared for college. In addition, state performance metrics often do not align well with the missions of broad-access institutions such as community colleges, and states do not adequately support institutional efforts to better understand where they are failing and how best to respond.

Even if performance funding ultimately proves to significantly improve student outcomes, the fact remains that it has serious unintended impacts that need to be addressed. Faced both by state financial pressures to improve student outcomes and substantial obstacles to doing so easily, institutions are tempted to game the system. By reducing academic demands and restricting the enrollment of less-prepared students, broad-access colleges can retain and graduate more students, but only at the expense of an essential part of their social mission of helping disadvantaged students attain high-quality college degrees. Policy makers should address such negative side effects, or they could well vitiate any apparent success that performance funding achieves in improving student outcomes.

In the end, performance funding, like so many policies, is complicated and even contradictory. To the question of whether it works, our answer has to be both yes and no. It does prod institutions to better attend to student outcomes and to substantially change their academic and student-service policies and programs. However, performance funding has not yet conclusively produced the student outcomes desired, and it has engendered serious negative side effects. The question is whether, with further research and careful policy making, it is possible for performance funding to emerge as a policy that significantly improves student retention, graduation and job placement without paying a stiff price in reduced academic quality and restricted admission of disadvantaged students. Time will tell.

Kevin Dougherty is a senior research associate at the Community College Research Center, Teachers College, Columbia University and an associate professor at Teachers College. Sosanya M. Jones is an assistant professor at Southern Illinois University. Hana Lahr is a research associate, Rebecca S. Natow is a senior research associate, Lara Pheatt is a former research associate and Vikash Reddy is a postdoctoral research associate, all with CCRC.

Editorial Tags: 

Report Proposes Alternate Form of Accreditation

The Center for American Progress today released a report that proposes a "complementary competitor" to the current system of accreditation.

The report describes three primary components for an outcomes-focused, alternative system, which, like current accreditors, would serve as a gatekeeper to federal financial aid.

  • High standards for student outcomes and financial health;
  • Standards set by private third parties;
  • Data definition, collection and verification, as well as enforcement of standards by the federal government.

"If implemented, this new system would provide a pathway to address America’s completion and quality challenges through desperately needed innovation," the report said. "And it would do so while establishing strong requirements to ensure that students and taxpayers get their money’s worth."

Differing Views on Free College, State Disinvestment

Public Agenda, a nonpartisan group, on Thursday released results of two recent national surveys of American adults on higher education. Respondents generally favor using taxpayer money to make public colleges free for students from low- and middle-income families, with roughly two-thirds calling it a good idea. However, the survey found that Democrats are much more likely to like free college proposals (86 percent) than Republicans (34 percent). Respondents were also divided by age, with those under 49 liking the free-college idea (73 percent) more than those who are at least 50 (58 percent).

The group also found a partisan divide on a question about cuts in state government funding of public colleges. Democrats were more likely to call disinvestment a problem (79 percent) than were Republicans (57 percent).

What current college rankings do and don't tell us (essay)

The Wall Street Journal and Times Higher Education have partnered to produce yet another college ranking. Should we applaud, groan, ignore or something else? I choose applause -- with suggestions.

This new project represents a positive step. For starters, any ranking that further challenges the hegemony of what I have termed the “wealth, reputation and rejection” rankings from U.S. News & World Report is welcome. Frank Bruni said much the same thing in his recent New York Times column, “Why College Rankings Are a Joke.”

I traveled the country for two years for the U.S. Department of Education -- I called myself the “listener in chief” -- to hear what students and colleges wanted or worried about in the federal College Scorecard. I explained that one of the most important reasons to develop the College Scorecard was to help shift the focus in evaluating higher education institutions to better questions and to the results and differences that should really matter to students choosing colleges and taxpayers who underwrite student aid. Just this week, the White House put it this way:

By shining light on the value that institutions provide to their students, the College Scorecard aligns incentives for institutions with the goals of their students and community. Although college rankings have traditionally rewarded schools for rejecting students and amassing wealth instead of giving every student a fair chance to succeed in college, more are incorporating information on whether students graduate, find good-paying jobs and repay their loans.

Some ratings have already blazed new trails by giving weight to compelling dimensions. Washington Monthly, for example, improved the discourse when it added public service to its criteria. The New York Times ranking high-performing institutions by their enrollment rates for Pell-eligible students was an enormous contribution to rethinking what matters most. The Scorecard in turn contributed by adding some (but admittedly not all) of the central reasons we as a nation underwrite postsecondary education opportunity: completion, affordability, meaningful employment and loan repayment.

The WSJ/THE entrant also offers positive approaches to appreciate. The project shares some sensible choices with the Scorecard, starting with not considering selectivity. That’s good: counting how many students a college or university rejects often tells us more about name recognition and gamesmanship than learning.

The WSJ/THE rankings also incorporate repayment, which is a useful window that reflects such contributing factors as an institution’s sensitivity to affordability, which in turn includes net price and time to degree. It also grows out of whether students are well counseled and realistic about debt and projected income -- and use their flexible repayment options, like income-contingent choices, so they can handle living costs and also their loan obligations, even if they choose work that pays only moderately.

And it was a very smart choice to use value-added outcome measures, drawing on work by the Brookings Institution melding Scorecard information and student characteristics. That approach is designed “to isolate the contribution to the college to student outcomes.” It is also important because value-added metrics help respect the accomplishments of colleges that are taking, or want to increase enrollment of, populations of students that might not graduate or achieve other targets as easily as others.

That said, the WSJ/THE transition to outcomes-based measures is incomplete. A fresh new ranking is a chance to recognize colleges and universities that do a remarkable job in achieving strong results for students at affordable prices. Including a metric for per-student expenditure is an unfortunate relic from old-fashioned input-based rankings. It’s a problem not just because it mixes inputs and outcomes. More significantly, it clouds the focus on results, giving an advantage at the starting gate to incumbents that simply have a lot of money, even if other institutions are achieving better results more economically. That’s counterproductive when the goal should be to identify and reward institutional efficiency and affordability measures that generate good results as quickly as possible.

Even Better Questions

But it’s not too late. The tool already allows sorting by the component pillars, and Times Higher Education plans to work with the data to explore relationships and additional questions. WSJ/THE could rerun their analysis without the wealth measure and its conservative influence to see whether and how that alters the rankings. It will be interesting to see whether publics rise if resources are factored out. I’d like to think that a true results-based version would surface colleges and universities, perhaps a bit lesser known, that outperform their spending and bank accounts. Whether institutions achieve that through innovation, culture, dedication or some other advantages or efficiencies, it’s well worth a look.

These rankings include a new dimension, using a new 100,000-student survey that asks students about the opportunities their colleges and universities provide to engage in high-impact education practices. The survey also asks about students’ satisfaction and how enthusiastically they would recommend their institutions. We definitely need fresh ways to understand such education outcomes as well-being, lifelong learning, competence and satisfaction. How effectively does this survey, or the Gallup-Purdue survey, answer that need?

My first question is whether this really fits the sponsors’ desire to focus on outcome measures. Is it just an input or process measure, albeit an interesting one? I’d also like to know more about the relationship of opportunities to actual engagement in those practices, and I wonder how seriously or consistently students answered the questions. What does it mean to different students to have opportunities to “meet people from other countries,” for example, and does simply meeting them make for a more successful education?

This is also my chance to ask about the strength of the causal links between the opportunity to participate in a particular educational practice (whether an internship or speaking in class) and the outcomes, from learning to job getting and performance, with which they may be associated. This question applies to questions with the WSJ/THE “Did you have an opportunity to …” format, and is most vivid for the Gallup-Purdue study. Maybe the people whose personalities and preferences incline them to choose projects that give them close, long-term connections with faculty members -- or whom faculty members chose for those projects based on characteristics including charm, social capital or prior experiences -- are engaging, optimistic and, yes, privileged in ways that would also make them engaged and healthy later in life. Was the involvement with those practices in college really causal?

To evaluate the WSJ/THE question about recommending the college or university to others, it would be valuable to know more about the basis for the students’ responses: Were they thinking about academic or quality-of-life considerations? Past experience or projections for how their educations would serve them later? I wonder if their replies were colored by an intuition that their answers could affect their institution’s standing, thus kicking in both their loyalty and self-serving desire to promote it. In short, it’s hard to know how much these new measures tell us. But it’s a worthy effort to build new tools beyond the few crude metrics we have now for understanding differences among institutions.

Serious Work to Do

On a broader level, none of the ratings and rankings have made much progress in expressing learning outcomes, although we urgently need more powerful and sensitive ways to articulate them. Whether or not students have developed the knowledge and skills, the capacities and problem-solving abilities, appropriate to their program should be at the heart of any assessment of educational results.

It’s also time to move beyond Washington Monthly’s good but simple additions to capture intangible and societal outcomes more successfully. By that failure of imagination, not only we ratings designers, but also all of us in higher education, have allowed income to take center stage among outcomes -- and played into the damaging transformation in perception of higher education from a public to a private good.

I said many times in the Scorecard conversation that it’s sensible and understandable for families to want to know if an educational experience would typically generate what they would consider a decent living, including the ability to handle the loans assumed to pay for it, and how employment or income results compare across programs or institutions. But, I went on, that does not mean that income can stand alone as though that’s all that matters. If an affordable housing advocate or journalist is satisfied in her preparation to do that work, or a dancer or teacher got exactly the education he needs to pursue his goals and be a good citizen, how can we measure and reflect that? This is not a news flash but an echo: we need better ways to communicate with families and students about how higher education makes a difference to both the student and society far beyond just economic returns.

There’s other work to be done in the college information enterprise. This month, as the Scorecard celebrated its first anniversary, the Education Department marked the occasion with a data refresh and also wove together new partnerships to expand the Scorecard’s value. One of the big challenges for any useful college information source is to make sure it reaches the students who need it the most: the ones who aren’t sure if, or may already doubt that, they can afford college, who know least about the options available, who are uncertain about the outcomes they can expect from college or at different schools. They’re the ones who suffer from the severe “guidance gap” at far too many high schools and among many adult populations. This intensified collaboration among government, counselor organizations and higher education institutions at every level is a wise strategy.

Ultimately, however, the most transformative role of the well-conceived rankings and scorecards will turn out to lie not in whether every student reads them but in their value in supporting institutional improvement. They are a gold mine for benchmarking and can help institutions choose which outcomes really matter and then work across functions to improve them.

The fact is that for decades colleges have invested far too much energy striving -- or replaying games -- to get better on measures that don’t really matter. Asking smarter questions about genuinely significant priorities can help us graduate more low-income students into rewarding work and find affordable paths to solid learning outcomes for citizens and workers. Better ratings, continuously improved and building on each other’s contributions, give us a chance to put higher education’s intense competitive energies into worthy races.

Jamienne S. Studley, former deputy under secretary of the U.S. Department of Education and president of Skidmore College, is national policy adviser with Beyond 12 and consultant to the Aspen Institute, colleges and nonprofits.

Editorial Tags: 
Image Source: 
iStock/filo

How effectively interviewing for a job is often like speed dating (essay)

Elaine P. Maimon explains why effectively interviewing for a job is often like speed dating.

Ad keywords: 
Editorial Tags: 
Show on Jobs site: 
Image Source: 
iStock/sturti

Appeals court revives student rights suit in case over training for medical procedure

Smart Title: 

Appeals court revives suit over student complaints about being pressured to participate in medical procedure.

Eastern Kentucky Senate Rejects Health Plan

Eastern Kentucky University’s Faculty Senate voted overwhelmingly this week to reject a new health insurance plan, The Richmond Register reported. Some 83 percent of the senate symbolically opposed the plan, which they said would raise premiums for faculty and staff members between 200 to 400 percent. A single employee on the standard plan, for example, would see a monthly increase from $41 to $170. A family plan would jump from $472 to $963, according to the resolution.

The faculty statement says those who earn least would be most adversely affected by the changes, which represent “a grievous moral and ethical lapse, particularly for a public university that has defined itself as a champion of first-generation college students who come from lower-income backgrounds.” Sarah Pitt, the university’s chief human resources officer, addressed the senate, saying that meetings to explain the changes to employees are scheduled, according to the Register. The university is facing state funding cuts of 4.5 percent for each year of the current two-year budget and expects the new plan will result in health care cost savings.

Kristi Middleton, university spokesperson, said via email that the Faculty Senate "correctly referenced the increase in insurance premiums if employees choose to keep the same deductibles, copays etc. of the former standard plan. With three tiers of plans offered by [the university], the change in premium rates varies based on each employee’s current and future benefit choices. [The university] is working with individual policyholders to find the best option to meet their medical needs and family budget." (Note: This story has been updated from an earlier version, in which Middleton said that the Faculty Senate misrepresented the scope of changes to health care premiums.)

Ad keywords: 

Colleges respond to social media panic over alleged clown sightings

Smart Title: 

Colleges are victims of a social media hoax, prompting concern that people dressed as clowns were stalking students and challenging campus security leaders.

Pages

Subscribe to RSS - administrators
Back to Top