Office of Institutional Effectiveness
The mission of the Office of Institutional Effectiveness (IE) is two fold:
- To promote educational effectiveness throughout »Æ¹ÏÊÓƵ Community College (»Æ¹ÏÊÓƵ)
by collecting, analyzing, and disseminating data concerning students, faculty and
staff, educational programs, administrative and support services, and institutional
planning, policy development and decision making;
- To promote continuous improvement of student learning through assessment across all
academic programs, facilitating the assessment process, and providing support to all
College departments with their assessment efforts.
Specific Project in Mind?
Please fill out our Data Request Form and we will follow up with you.
Data Request Form
|
Contact Us and Request Data
Do you need any help with:
- Accreditation and Program Review?
- Research and Survey Design?
- Data Management and Data Analysis?
ie@berkshirecc.edu
|
Assessment Guidance
First Things First
You are not alone. The Office of Institutional Effectiveness serves as a support staff
for all institutional assessment activities. Whether you are brand new to assessment
or you are a grizzled veteran, whether you need a lot of help or just a little, IE
is a resource that is available to you.
The Importance of Assessment
How do you know that the things you are doing are working? This question is the essence
of assessment. It is best to think of assessment as a continuous cycle. The cycle
begins with analysis of evidence (Pre), then moves to action that is designed to affect
change (Intervention), and then back to analysis of evidence to determine efficacy
of that action (Post) and identify the need for any further action, which would start
another cycle. By measuring at the beginning and end of the process, we can understand
the effect that our actions have. When we assess our activities, we can move beyond
feeling like we accomplished our work to knowing that we did.
-
A Key Performance Indicator (KPI) is a measurable value used to evaluate an outcome.
We often use KPI's such as Graduation Rates, Retention Rates, and Course Pass Rates
to evaluate the overall work of the College, but KPI's can be as grand or as granular
as the situation demands. Here are a few tips for selecting meaningful KPI's:
- Select KPI's that match the scale of the work. For instance, the Library may decide
that number of students served is a KPI they want to focus on instead of graduation
rate. We sometimes refer to this as "taking the pulse close to the heart". KPI's like
graduation are the culmination of hundreds (if not more!) of factors and should only
be used to assess the College at a holistic level. Select KPI's that are appropriate
to the scale of the work.
- Select KPI's that you cannot directly manipulate. Continuing with the Library example,
they could decide to measure their success based on the number of books in their collection.
The problem with this approach is that it does not account for how useful the books
are or how they enhance the collection. The Library could buy one thousand copies
of the same book and this metric would go up, but the increase in service to students
would be minimal. The Library could choose to buy books that are cheaper (so they
can buy more), but less useful than more expensive books. Changing behavior solely
to increase the KPI's without regard for the real-world impact is known as "dial turning"
and is a pitfall of selecting KPI's that you can artificially control.
- Select KPI's that are reliable. Every metric has an inherent amount of instability,
known as variance. Metrics with a low variance are stable across time and/or different
groups and are good candidates for KPI's. When there is little natural fluctuation,
it is easier to see if changes you make are having an effect. Other metrics have high
variance with significant changes from year to year or group to group with no simple
underlying cause. This happens most when groups are small and a shift of just a few
members can tip the scales or when outcomes rely on complex and numerous inputs that
are difficult to isolate. Selecting stable metrics gives confidence that any changes
you are seeing are a product of your work and not natural fluctuation in the data.
- Understand the difference between Lagging and Leading indicators. Those who are familiar
with Formative vs Summative Assessment will recognize the dynamic at play here. The
aptly named Lagging indicators occur at the end of an event and measure what happened
in the past. They are usually easy to measure, but don't provide the opportunity to
change past outcomes. Leading indicators measure early in an event, sometimes even
before an event occurs. They can be more difficult to measure, but they provide an
opportunity to influence the outcome. Both types of indicators are useful for different
purposes; be sure you understand which one you are using and why.
-
Benchmarking measures »Æ¹ÏÊÓƵ against its peers. Many of the large, institutional reports
that are available have built-in benchmark cohorts for comparison. The IPEDS and VFA
systems both include benchmark groups with national peers using Carnegie Classification
(more on Carnegie Classification below). Most reports will have a section listing
peer institutions and the criteria used to select them.
However, some do not which may prompt you to create your own benchmark group from
data that are available to you. When deciding which institutions to use for benchmarking,
it is important to select those schools that are like us in the ways that are important
to the data being studied. While criteria can vary depending on the data set, you
will generally want to select schools with matching Carnegie Classifications. For
»Æ¹ÏÊÓƵ that would be Small (<1000 FTE), Public, two-year, and Rural. Within the MA state
system, school matching these criteria are Greenfield CC and Cape Cod CC.
-
You may be asked to make projections or set goals for the future. Remember that projections
and goals are targets to aim for, but that you probably won't hit all of your marks
for a myriad of reasons. After all, if we could accurately predict the future we would
all be lottery millionaires. That said, there are a few things you can do to improve
your projections. If we think of projections as a continuum from "Guessing" to "Certainty",
we hope to move away from Guessing and toward Certainty using a few tips:
- Use strong rationale for your projection/goal. If you project that a student outcome
will increase by 5% as a result of a project, you should know why. Picking 5% because
it is a satisfyingly round number is a weak rationale. Picking 5% because three other
schools who have tried the same intervention have each seen at least 5% increase is
a much stronger rationale.
- Set appropriate goals. We would all love to see 100% completion, retention, and graduation,
but that simply is not realistic. Setting unrealistic goals risks demoralizing and
alienating those you need achieve that goal. Good goals can be a stretch, but need
to be attainable to keep people engaged. For internal work, there is usually nothing
wrong with setting a goal that will be difficult (but not impossible) to achieve.
Even if you fall a little short, that is still progress and should be celebrated!
There may also be some times when you want to set "safe" goals that you know you can
achieve like for an outside accreditor or grantor where goals need to be met to satisfy
conditions. Use your best judgement.
- Choose the right metric. Limit your projections/goals to things that could reasonably
be affected by the intervention. For instance, making projections about an increase
in graduation rate based on a change in new student orientation is probably reaching
a bit too far. There are hundreds, if not thousands of events between orientation
and graduation that impact that outcome. It would be better to look for changes in
something like the SENSE or CCSSE surveys which measure student connections in the
first few weeks of a semester.
- Beware statistical pitfalls. »Æ¹ÏÊÓƵ is a small school and we normally deal with small
populations. As a result, we often see natural fluctuations in student behavior that
can appear to be something on the surface, but that don't hold up under statistical
scrutiny. Looking at full-time freshmen retention rates from year to year, the rate
swings from the high 40's to the mid 50's and back again with regularity. The high
variance of this metric makes using it for projection problematic. This is an area
where the Office of Institutional Effectiveness can be a great help.
- Use concrete language. The more concrete the language you use, the easier it is assess.
Words like "increase" "decrease" "add" and "eliminate" are strong words that aid in
assessment. Soft language, like "foster" "encourage" "promote" "strive" and "help",
is very difficult to measure and makes assessment much more difficult.
-
There are a number of data sources available to measure student outcomes, each with
their strengths and weaknesses. The table below lists several data sets that are available
to the College community. While these data sets certainly cannot answer every question,
they are a great place to start looking at the College.
The PMRS has been singled out as of particular interest by the MA Commissioner of
Higher Education in the Strategic Planning process.
Name |
Target Population |
Frequency |
Governing Body |
Provides Benchmarks |
Type of Data |
IPEDS
|
Students
|
Annual
|
Federal
|
Yes
|
Outcomes
|
PMRS (HEIRS)
|
Students
|
Annual
|
State (MA)
|
No
|
Outcomes
|
VFA
|
Students
|
Annual
|
Independent
|
Yes
|
Outcomes
|
NECHE Data First Forms
|
Students
|
Annual
|
NECHE
|
No
|
Outcomes
|
SENSE/CCSSE
|
Students
|
Biennial
|
Independent
|
Yes
|
Early Connections
|
Great College to Work For
|
Staff
|
Biennial
|
Independent
|
Yes
|
Satisfaction
|
Name |
Strengths |
Weaknesses |
IPEDS
|
Highly standardized
|
Focus on Freshmen
|
PMRS (HEIRS)
|
Highly standardized, Reflect Commissioner's Priorities
|
Scope Limited to Commissioner's Priorities
|
VFA
|
n/a
|
Focus mainly on most successful students
|
NECHE Data First Forms
|
n/a
|
n/a
|
SENSE/CCSSE
|
Best measure of early connection
|
n/a
|
Great College to Work For
|
Disaggregated by work group
|
n/a
|