NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE... This user guide is for users of AAUDE shared data...

advertisement
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
1
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm
This user guide is for users of AAUDE shared data from NSSE 2.0 (version 2, known here as v2,
introduced 2013) and data from NSSE 1.0 (2000-2012, aka v1) converted to v2 formats where
items and response scales were sufficiently comparable (known throughout as v1_v2).
See the website for codebook with extensive mark-ups; variable list in Excel; data and
background on response rates, populations, and samples; information on AAUDE questions; and
procedures for AAUDE data sharing and using AAUDE questions.
Super quick guide





START by running Enhancements.sas – set the choices and locations in the code as
desired. Described below.
THEN run Longitudinal.sas – set choices as desired, to select schools and years.
YOU WILL THEN HAVE

Temp dataset long1, row = respondent (in year, school, class level), v2 and v1_v2

Permanent dataset BaseStats, row = year*school*class level*item, with
marginals, v2 and v1_v2
NOTE, in both datasets, the “key variables,” described below.
USE the marked-up codebook and key variables to plan and execute your analyses.
Contents of this doc listed below. This is a chicken-and-egg doc, so many sections use terms etc.
defined in later sections. Sorry! See section near the end for changes 2013 to 2014. These are
minimal, affecting only two demographic vars. We did not revisit the v1-to-v2 decisions we
made with the 2013 data; we simply took that model and added 2014 and 2015 data.
NOTE that the contents points to a section on What’s new in (data from administration year
XXXX)
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
2
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm
Files to have at your fingertips ........................................................................................................ 2
Files – SAS datasets, SAS formats/code........................................................................................... 2
Key variables – for ID/selecting years, schools, respondents ......................................................... 5
Variables UCB added or modified ................................................................................................... 6
Notes on other variables – HIPsum.., engagement indicators, duration, sexorient ....................... 6
Picking years and schools and respondents for analyses ................................................................ 8
Missing data – Handling in SAS, patterns of response, PctValid vars, comparison to v1 ............... 8
Naming conventions – Files, variables, labels ................................................................................. 9
Majors – Use the CIPs .................................................................................................................... 10
Institutional metadata – Populations, samples, response rates ................................................... 11
AAUDE Questions – Changes, comparisons, CU calcs – runaround, obstacles ............................. 12
VSA................................................................................................................................................. 13
Things CU-Boulder did in creating the v2.0 response-level dataset ............................................. 13
Creating the v1_v2 dataset (in 2013; we didn’t recheck all this) .................................................. 15
Adding a new year ......................................................................................................................... 16
2015 changes/updates .................................................................................................................. 17
2014 changes/updates .................................................................................................................. 17
NOTES, CAUTIONS specific to ONE YEAR....................................................................................... 18
PBA file refs and summary of Boulder to do’s............................................................................... 18
Files to have at your fingertips

This user guide

The marked up codebook

The SAS datasets and code from the passworded page.

The ResponseRates Excel for the N_compare_Pivot tab to see patterns of administration and
use of AAUDE questions over years
Files – SAS datasets, SAS formats/code

SAS datasets – all version 9. PBA ref: L:\IR\survey\NSSE\AAUDE\All\datasets, aka nssedb

METADATA -- Institutional metadata

Row = school x year x IRclass (1, 4) value

Includes 2005 – most current

Population and sample counts, response rates, derivations, weights.

Same info is in ResponseRates Excel with plots, narrative
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
3
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm


BASESTATS

means and SD's (plus N and PctMiss) by institution, class level, and item. Row =
school x year x IRclass (1, 4) value, plus marginals, x item. 2001-most current, with
v1 data converted to v2 formats.

Stats are based on compare=1 only

Sorted by item x year x inst x class level with most recent year on top. Items are
ordered by var_order

CAUTION: here it’s variable CLASS, with values Freshman, Senior, All, rather than
IRClass (values 1, 4)

Contains handy key variables including public, Partic2013, Partic2014, Partic2015,
Partic2016, v2, LastV1, compare (always 1 in this dataset)

And item-specific variables: variable (name), LabelShort, q_num, Var_Order (as in
MatchVars_v2), and InV1_v2 so you can limit rows to items with both v1 and v2
data if desired.

Boulder had thoughts of, but decided against, taking on getting the stats into
Tableau. If anyone else does, let us know, for sharing on the AAUDE-NSSE website.

Boulder considered but decided not to add distributions (percentages) from layout
(e.g., pct1, pct2, pct3, with pct who selected response values 1, 2, 3)
Response-level data in v2 formats

NSSEALLYRS_V2: 2013 (and later) response-level data, all institutions

Vars include base questionnaire, NSSE recodes from base (e.g., estimated hrs
from x to y scale), NSSE engagement indicator scales (e.g., HO, higher-order
learning), NSSE-coded majors and CIPs, NSSE weights, items from all NSSE
modules, AAUDE questions, and CU-added utility variables (e.g., COMPARE)

See Enhancements.sas below to change to shorter variable labels and attach
formats.

CONVERT_V1_V2: Converted version 1 data to version 2 data.

2001-2012 response-level data, all institutions.

Contains the 120 items from v1 that could be converted to a v2 item. Items have
been renamed to the v2 name and values converted to v2 values.

233,000 rows – this N will not increase. This dataset will not be updated for
future administrations aside from adding e.g. Partic2015 flags to all rows of
schools participating that year

Use the provided SAS code to put these two datasets together and select from them
for longitudinal analyses.

NEW Nov. 2015: NSSEALLYRS_V2_TEXT: 2014 (and later) response-level data for
the open-ended response questions, all institutions. These items were not on what
AAUDE got from NSSE in 2013.

Row = 1 respondent, with year, IPEDS, INST, IRclass, and surveyID.

Can be merged with NSSEALLYRS_V2 on YEAR, IPEDS, INST, IRclass and
SURVEYID. (surveyID may not be unique across the other keys)

Not all respondents are in this dataset. The dataset contains only respondents
who made at least one non-blank response to one of the below items.

Vars include

GENDERID_TXT (gender identity),

SEXORIENT_TXT (sexual orientation),
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
4
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm




ADV03_TXT (academic advising module, primary source of advice),
CIV03_TXT_PLUS (civic engagement module, most meaningful experiences),
FYSFY06B_6_TXT (FY experiences & SR transitions module, considered
leaving)

FYSSR01A_TXT (plans after graduation),

FYSSR08_TXT_PLUS (anything institution could have done better to prepare
for career or further education)

Note: CIV03_TXT_PLUS and FYSSR08_TXT_PLUS are a combination of
variables due to very long responses. CIV03_TXT_PLUS has a length of 1,062
and FYSSR08_TXT has a length of 3,843.

CAUTIONS from Lou on using this dataset

Some of the items follow a closed item with text additions to alternatives,
so you should not analyze the text items w/o reference to the closed item.

The dataset is structured to be efficient for merging and analysis, not
storage. Most of the space in it is blank.

Many of the items come from modules, not base NSSE, so were presented
to a minority of respondents.

We masked nothing. Some of the comments name individuals, many berate
offices or individuals.

We did not include the open-ended response items on major/discipline in either
dataset. Use the CIP code variables for this instead.

Going forward, all future open-ended response items will be in this dataset.MATCH_V2 'Match' dataset used to assign formats and labels. Row=variable. For v2, with a map of
the conversion of v1 to v2 variables. Comes from MatchVars_v2 Excel, which is the
master source. The Match_v2 dataset is used in generating code that handles the
conversion of v1 to v2.
SAS formats and code for use with v2 and v1_v2 data.

All have file extension .sas but are ASCII files that can be opened in any text editor or
even in Word.

Functions

Put v2 and v1_v2 data together

Switch from the extremely long variable labels provided by NSSE to shorter labels

Drop variables from the NSSE modules

Create SAS formats to label values

Attach formats to variables

Select populations of schools, years, and respondents to suit the analyses you’re
after

Formats_v2.sas – creates formats to label values.

see codebook markup or MatchVars Excel or match dataset for what formats go
with what vars.

Enhancements.sas attaches formats to variables in a temporary dataset.

PBA ref: L:\IR\survey\NSSE\AAUDE\All\gencode\Formats_v2.sas

Enhancements.sas

Run FIRST.

Allows user to switch to shorter variable labels; also creates and attaches formats to
variables in the dataset. Calls formats_v2.sas
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
5
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm

PBA ref: L:\IR\survey\NSSE\AAUDE\All\Enhancements.sas
Longitudinal.sas

Run AFTER Enhancements.sas.

Code to concatenate nsseallyrs_v2 with convert_v1_v2 to allow longitudinal analysis
across v1 to v2, and to select comparable populations over time.

Creates temp dataset long1, with all v2 vars (except those from modules, if
dropped) plus two vars from v1: LastV1 (0/1 marking latest year a school did v1)
and aau17v1 (the AAUDE “obstacles” question from v1, for reference – not similar
enough to v2 AAUDE obstacles question to recode). New vars in v2 will be missing
in v1 years.

PBA ref: IR\survey\NSSE\AAUDE\All\Longitudinal.sas 
convert_v1_v2.sas

For reference, what we did in converting v1 to v2. Users do not need to run or
understand this!!

PBA ref: L:\IR\survey\NSSE\AAUDE\All\Convert_V1_V2.sas

NSSEinit.sas

Initialization file for CU-Boulder users. Suggest you establish similar with your
locations and directives

PBA ref: l:\ir\survey\nsse\aaude\NSSEinit.sas
In a separate ZIP - V1 only datasets, with v1 variable names – You should not need these at
all.

Datasets: Basestats_v1, Match_v1, nsseallyrs_v1 (same as nsseallyrs through 2012
data), nsse_oneyear_v1 (with 170 variables used in only one year)

Use ONLY if you have code for v1 that you want to use on v1 only data

PBA ref: L:\IR\survey\NSSE\AAUDE\All\datasets\archive, aka nsardb


Key variables – for ID/selecting years, schools, respondents

KEY variables are noted with the word KEY in the label -- meaning, critically important – not
all are literally keys, nor are all literal keys so labelled. Generally near top of dataset. In
both convert_v1_v2 and nsseallyears_v2 datasets except as noted

YEAR -- Survey Year (KEY) – Numeric, e.g. 2013 – coverage depends on dataset. IS A
KEY. No missing. YYYY. 2001+.

INST -- Institution Name (KEY) – Short (16 char) version; e.g. ‘Colorado’ – these are
identical over years and 1-1 with IPEDS so you can use them as a key. No missing.

IPEDS -- IPEDS code (KEY) – Character, e.g. ‘126614’. IS A KEY. The 6-digit IPEDS
code. Text variable. XXXXXX for McGill, YYYYY for Toronto. (NSSE uses 24002000
McGill, 35015001 Toronto )

PUBLIC -- Public Institution (0=no, 1=yes)

TRGTINST -- Target institution (set to 1 for own institution) – 0/1 numeric; initially
all zero.

Partic2013, Partic2014, Partic2015, Partic2016 are 0/1 vars set for a school on all
years administered – These make selecting comparable populations easy. E.g.
“where Partic2013” gives all years for the schools that administered in 2013.

Partic2016 is already set (based on 10-26-15 info; could change). We’ll add
Partic2017 etc. in annual updates.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
6
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm




AAUDEQ - 0/1 (zero, one) variable where 1 = respondent answered at least one
AAUDE question. 0 = did not. 0 if school didn’t ask AAUDE q’s in a year, OR if
student didn’t answer any of them.
Property of an individual response/row

COMPARE – Respondent-level. COMPARABLE over schools: Random sample of
eligible first years/seniors (KEY) – 0/1 numeric. Critically important = use “where
compare” for ALL comparisons across schools.

IRclass -- Institution-reported: Class level (KEY) – IS A KEY to stats.

Values 1, 4 – Numeric, for first-year, seniors

Other values occur but not within COMPARE=1.

Really not “class level” – e.g., 1=first year can include soph status.

SurveyID – NSSE-assigned ID number – important only for linking to your own
student records
V2 -- NSSE version 2 (2013 & later) - 0=no, 1=yes.
LastV1 – 0/1, marks cases from the school’s last v1 administration. Not in the
nsseallyears_v2 dataset. Useful for comparing v2 responses with responses from only
the last v1 year for a school.
Variables UCB added or modified

YEAR, PUBLIC, TRGTINST, COMPARE, V2, Partic2013, Partic2014, Partic2015, Partic2016,
LastV1

CIP codes and descriptors of major codes (see section on Majors, below)

SATT (V+M)

MULTRE – calc, were multiple racial/ethnic categories checked

PctValidQ1_19 and PctValidQ20-36 – pct of questions student made a valid response to
– see section on Missing Data, below

We changed NSSE variable “logdate” from a date-time to a date.

A new var, logdatetime, records the date-time so you can see how many did it at
3am

AAU13Avg, NValid, N_1, N_34 – derivations from items AAU13a-g, the new-format
obstacles question

BACH – 0/1 expecting bachelors at this school – derived from AAUDE q
Notes on other variables – HIPsum.., engagement indicators, duration, sexorient

Variables HIPsumFY and HIPsumSR are calc’d by NSSE from items judged “high impact
practices.”

The variable labels tell the varnames of the items that contribute – 3 for FY, 6 for SR

HIPsumFY: Number of high-impact practices for first-year students (learncom,
servcourse, and research) marked 'Done or in progress'

HIPsumSR: Number of high-impact practices for seniors (learncom, servcourse,
research, intern, abroad, and capstone) marked 'Done or in progress'

The HIPsumXX vars range 0 to 3 or 6, with one point for each item marked “done or in
progress.”

HIPsumFY is missing for all IRclass=4 (seniors), and vice versa

If a student answered none of the 3 or 6 items, the HIPsumxx variable is missing. It is
NOT missing if the student answered even one of the 3 or 6.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
7
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm







HIPsumFY is missing for about 15% of IRclass=1
HIPsumSR is missing for about 13% of IRclass=4
Note that the 3 items in the HIPsumFY scale are also in the HIPsumSR scale, with
internship, study abroad, and capstone added.
The Engagement Indicators (EI) – variables HO RI LS QR CL DD SF ET QI SE – see p 17 of the
codebook markup. The var labels also show what items are included. All range 0 to 60,
oddly.

The v1 benchmarks and CU scales have not been carried into v1_v2 data.

The EI indicators have not been calc’d for v1_v2 because most rely on new or
extensively-modified variables; the exception is the HO indicator (which we didn’t calc
for v1_v2 either).
DURATION (in minutes) – it’s elapsed time not time on task!!! And <1 min was handled
differently in v1 and v2. Use MEDIANS.

Duration MEDIAN = 13 v2, 14 v1.

But duration MEAN = 34 v2, 24 v1. More in v2 were interrupting and taking up later.
Use mean duration with GREAT CAUTION

Most v2 respondents are in the 8-22 minute range, centered around 12 min.
SEXORIENT - Sexual orientation – New 2013, at school option.

2013: Colorado and Ohio State posed this Q to their respondents; Iowa St, Kansas,
Michigan State, Nebraska, Washington did not.

Most students at CU and OSU did respond. (who got to the demog section)

Distributions are similar – 6% “prefer not to respond’ (which we have NOT set to
missing), 1-2% each gay, lesbian, bisexual, and questioning/unsure, 87-92%
heterosexual.

NSSE has modified this q for 2014, adding “Another sexual orientation, please specify”
to the options. See
http://websurv.indiana.edu/nsse/interface/img/Changes%20to%20Demographic%20Ite
ms%20(NSSE%202014).pdf

NSSE chose to create a new variable SEXORIENT14 with the additional response but we
have decided to rename the new variable back to SEXORIENT since fewer than 0.5% of
respondents chose “Another sexual orientation, please specify.”

“Another sexual orientation, please specify” has a value of 5 in 2014. The value for the
response “Questioning or unsure” changed from 5 in 2013 to 6 in 2014. This change has
been accounted for in the 2013 data.

The write-in responses for the respondents that chose “Another sexual orientation,
please specify” are in the variable SEXORIENT_TXT.
GENDERID – gender identity – new in 2014, has replaced the variable Gender

The new question GENDERID asks “What is your gender identity?” and the response
values are “Man,” “Woman,” “Another gender identity,” and “I prefer not to respond.”

The old question GENDER asked “What is your gender?” and the response values were
“Male,” and “Female.”

We chose to equate the old question to the new GENDERID question since fewer than
0.2% of participants chose “Another gender identity.” This also meant changing the
Convert_V1_V2 dataset since the variable name was GENDER there as well.

The write-in responses for the respondents that chose “Another gender identity” are in
the variable GENDERID_TXT.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
8
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm
Picking years and schools and respondents for analyses

See the sample where clauses and if statements in the code Longitudinal.sas

Look at ResponseRates.xlsx, especially the tab N_Compare_Pivot. This shows counts of
random-sample respondents by school over time with various filters available.

Variables Public, Partic2013, Partic2014, Partic2015, Partic2016, v2, and LastV1 are your
friends in selecting schools and years

Variable COMPARE is essential for picking respondents
Missing data – Handling in SAS, patterns of response, PctValid vars, comparison to v1

Coding missing data

We used SAS missing-data values, mostly . (default), for missing data = student did not
answer the question at all, do not include in stats

NSSE reports often show 9’s separately, not as missing, so you may need to take this
into account when matching NSSE reports to your own analyses.

On a few questions we used SAS special missing values

Variable disability – No=0, Yes=1, I prefer not to respond = .P, missing but
differentiatable from ., did not provide any response. This is noted in the codebook
and acknowledged in formats.

Variable sexorient, sexual orientation, was optional per institutional request. If the
school did NOT request, we set this variable to .N, to differentiate from those who
were at schools that asked (Colorado and Ohio State) but provided no response.

These are shown in the codebook markup

V2 2013, what the respondents actually did

The codebook markup shows where NSSE questionnaire screens broke, and where
respondents dropped off.

We used variable PctValidq1_19 (19 is the last q before demographics) to explore
respondent patterns, and catergorized them with format pvalid.

Of those respondents with 90%+ valid responses on the ratings questions (1-19),
none had durations zero or one minute, 7% were 8 or fewer minutes. 12 still the
center.

THEREFORE, to get a measure of “seriousness” of the respondent, use the
PctValidq1_19 measure, not duration.

As noted in the codebook,

3% of R’s stopped at or before question 1e – midway through the very first block of
questions. PctValidq1_19 =< 7%.

Another 3% stopped after/around q 2a, the start of the second block, still on upper
half of screen 1; PctValidq1_19 =< 15%.

An additional 8% of R's stopped with QRevaluate -- bottom of screen 1. Total ~14%
never made it to screen 2. PctValidq1_19 =< 40%.

Another ~6% of R's stopped at/about q 13c, QIfaculty, bottom of screen 2. Total
20% never made it to screen 3; PctValidq1_19 =< 70%.

Those who DID go to screen 3 generally completed entire rest of the instrument
including demographics and AAUDE q's.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
9
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm

Relationship of PctValid to duration – yes – Duration is tricky because it’s in minutes but
extends to 17 days, reflecting students who started, stopped, and came back before
submitting. It’s elapsed time not time on task. But it clearly relates to PctValidq1_19,
and is unnervingly low for those who checked out early:
--Percent of questions 1 - 19 not missing, of 86 total
Lower
Upper
N Obs
Min
Quartile Median Quartile
Max
------------------------------------------------------------------------------Stopped in q01
415
0
0
0
0
117
Stopped in q02
506
0
0
0
0
5412
Stopped in q06, end screen 1
1067
0
0
2
4
7529
Stopped in q13, end screen 2
820
0
4
6
9
3643
Completed all screens/parts
11093
1
11
15
22 26222

Code to produce:
proc means data = long1 min q1 median q3 max maxdec=0 ;
var duration ;
where v2 ;
class pctvalidq1_19 ;
format pctvalidq1_19 pvalid. ;
run ;


V2 has more missing data than in v1, and very clear relationship to position in questionnaire.
True across all schools.

E.g. genderid self-report (a late question) – 19% missing v2, 11% v1. This is such an
easy, common question that missing indicates the respondent had stopped/finished by
the time it was posed.

The steady increase in pct missing w question number in 2013 is visible in BaseStats
dataset. 20-21% on q14 and all remaining. This was Probably true in v1 too but didn’t
get to as high a level.
Are there diffs by school? Yes. And the lower the overall response rate at a school, the
lower the average PctValidq1_19 at the school. Ouch.
Naming conventions – Files, variables, labels

Files

V2 in a name means, from version 2, 2013 and later, only

V1 in a name means, from version 1, through 2012, only, with formats, variable names,
response alternatives, wording, etc. all as through 2012

V1_V2 in a name means, data from v1 through 2012, that are comparable to v2 data
and with formats, variable names as in v2. The v1_to_v2 files are designed to combine
with v2 data for longitudinal analyses.

Variable names – max length 16 -- In V2 (and V1_to_V2)

We have not changed the var names NSSE sent, so names should match those in data
you received directly from NSSE.

In some cases the NSSE names are confusing, not systematic – especially those
involving majors. See section below.

ALL CAPS = added by CU Boulder: YEAR, PUBLIC, TRGTINST, COMPARE, V2

Some others added by CU Boulder are not in all caps: Partic2013, Partic2014,
Partic2015, Partic2016, LastV1

A few all caps are NSSE calcs, such as WEIGHT1/2

Engagement indicator scales are the 2-letter acronym, such as HO for higher-order
learning
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
10
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm


NSSE numeric vars, student responses to individual items – all lowercase

EXCEPT items that contribute to the EI Engagement Indicators have the 2-character
initial of the indicator in all caps as prefix. E.g.,

empstudy = q 14a, (How much does your institution emphasize the following) Spending significant amounts of time studying and on academic work, not in any
EI

SEacademic = q 14b, Providing support to help students succeed academically,
in EI SE, Supportive Environment

NSSE modules – items have 3-character acronym prefixes in all caps, followed by item
number (letter). E.g., ADV02d for the advising module

AAUDE questions have prefix AAU
Variable labels – in V2 and v1_to_v2, same labels

Are VERY long. These are from NSSE. One, in the diversity module, is 207 characters
long. Only 10% are < 40, the old limit. 90% are < 120.

UCB added the acronym for engagement indicators to the front of labels for items with
the EI prefix. E.g.

RIconnect, label = RI: Connected ideas from your courses to your prior experiences
and knowledge

Start with “—“ (two hyphens) when CU-Boulder added the variable or calc or modified
the var

ALTERNATIVE SET of labels is

Max 83 long – using abbreviations, discarding some words . . .

Attaches the question number and source info to the end of the label. E.g.

LS: Reviewed your notes after class (09b)

Est'd hrs/wk: tmrelax recoded by NSSE using the midpoints (15f calc)

Institution reported: First-time first-year (INST)

Diversity: Discussed: Different political viewpoints (DIV03d)

Academic quality of your major program (AAU08)

The codebook is organized by question number, so having this in the label makes
tracking exact questions easier. However, the short labels no longer have the full
question wording of the NSSE labels.

TO USE the alt set run Enhancements.sas
Majors – Use the CIPs

Bottom line – use the CIPS

NSSE vars, v2, carried into nsseallyrs_v2

MAJnum
How many majors do you plan to complete? (Do not count minors.) –
Numeric. In the 2013 data takes on values 1 and 2 and missing only; 19% of both frosh
and seniors said two. Not converted from v1 – there was something but it was too iffy
and too little utility to be worth pushing.

MAJfirstcode
First major or expected major - NSSE code – numeric value, really a
code, assigned by NSSE based on what the student wrote in. More differentiated than
the codes used in v1. Not converted from v1.

MAJsecondcode Second major - NSSE code – ditto.

MAJfirstcol
First major category NSSE code (discipline area) from MAJFIRSTCODE –
This is also a numeric var where a value means a discipline area, called variously
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
11
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm


“discipline” and “category” and “college” (as in the var name) and “broad field” - not
converted from v1. The varname SHOULD have the word “code” in it but doesn’t.

MAJsecondcol Secondary major category NSSE code (discipline area) from
MAJSECONDCODE – ditto
NSSE vars, v2, NOT carried into nsseallyrs_v2 – 6 vars each 256 characters, holding the
verbatim write-ins for first and second major. These appeared useless to us – NSSE already
coded into the vars noted above. Not worth 1500 characters!

MAJfirst, MAJFI0, MAJFI1, MAJsecond, MAJSE0, MAJSE1
Vars added by CU Boulder

No v1 conversion for these:

MAJfirstDesc --First major or expected major - NSSE description and code

MAJsecondDesc
--Second major - NSSE description and code

Both about 60 characters long, with the code in parens at the end – from the
NSSE format. DESC here stands for description.

DiscFirst
--First major category NSSE description (discipline area) from
MAJFIRSTCOL

DiscSecond --Secondary major category NSSE description (discipline area) from
MAJSECONDCOL

We realize these names do not match NSSE’s – we used DISC for discipline (as in
broad discipline, college). Same 60-char format as above.

CIP codes – all 6-character character vars. All HAVE BEEN COVERTED from v1. So for
longitudinal analyses of majors, USE THESE. CIPS are from the NSSE crosswalk, which
NSSE did by itself based on the one AAUDE did for it years ago (Mary Black and Linda
Lorenz)

CIP1First
--CIP1 from MAJfirstcode (first/expected major)

CIP2First
--CIP2 from MAJfirstcode (first/expected major)

CIP3First
--CIP3 from MAJfirstcode (first/expected major)

CIP1Second --CIP1 from MAJsecondcode (second major)

CIP2Second --CIP2 from MAJsecondcode (second major)

CIP3Second --CIP3 from MAJsecondcode (second major)

If you want a broad-field grouping for longitudinal (or even just v2) data, use your
own usual grouping of CIP codes.

You’re on your own figuring out how to use 3 CIPS per major
Institutional metadata – Populations, samples, response rates
The institutional metadata dataset, provided by NSSE, contains the number of freshmen and
seniors, by year of administration and participating institution, in each institution's population
and NSSE-selected random sample from the eligible population (COMPARE=1). It also contains
the number of responding students, the response rate (computed as the number of responses
divided by the number of students in the sample), and NSSE-computed sampling error (stored in
the dataset as a percentage).

The ResponseRates Excel is based on the same metadata but has additional info – plots,
narrative, reference to NSSE studies, more.

NSSE attempts an explanation of its sampling error formula at
http://nsse.iub.edu/html/error_calculator.cfm. An example of sampling error usage that
NSSE provides is: “assume that 60% of your students reply ‘very often’ to a particular item.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
12
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm


If the sampling error is +/-5%, then the true population value is most likely between 55%
and 65%.” This of course ignores non-response bias.
A row in the metadata dataset is a combination of institution, IPEDS no., year of
administration, and IRClass (1 = first year, 4 = senior). This dataset can therefore be merged
easily with the all-years dataset, which has a similar structure.
Both the metadata dataset and the all-years response-level dataset are sorted by year,
IPEDS, and IRclass.
AAUDE Questions – Changes, comparisons, CU calcs – runaround, obstacles

See the PDF codebook markup pages 18-19 for v2 questions, var names, formats.

See the AAUDE optional q Excel with map from v1 for questions dropped

Q 1-12 in v2 came directly from v1 so have v1_v2 data.

CAUTION: Q 10 in v2, on runaround, changed position – in v1 it was q14 and followed items
on academic advising, in v2 those items were dropped. But the response scales were
identical, with “strongly agree” listed first in both v1 and v2.

Every 2013 school shows more agreement with “At this university, students have to run
around from one place to another to get the information or approvals they need.”

This can’t be due to chance – I looked and looked to see if/how we’d messed up the
conversion. But we didn’t.

SO – the increases in agreement with “have to run around” are due to some
combination of

Real – there’s really more run-around required

Real – students now have less patience with run-around now (that is, higher
expectations) than students surveyed with v1

Effect of dropping the prior advising items

Effect of other changes in entire instrument

Something else

Obstacles to academic progress – most of this is also noted in the codebook and Excel

Old q17 asked students to pick the ‘biggest’, pick one, and “I have no real obstacles” was
a listed option.

New q 13 is actually 7 separate q’s, “please rate the following as obstacles to your
academic progress during the current AY” with 1=not an obstacle at all, 2=minor, 3=
moderate, 4=major. Furthermore, 13g is an alternative not listed at all in v1, Personal
health issues, physical or mental.

We obviously couldn’t recode old 17 into new 13a-g. But we DID carry the v1 data intact
into v1_v2 dataset, named AAU17v1, with format obstac2_. to label the responses. This
should allow you to compare old to new however is appropriate for you.

AAU13a – AAU13g are packed with info but difficult to summarize. We therefore
created and stored additional vars AAU13Avg, AAu13NValid, AAU13N_1 (number rated
1=not an obstacle), AAU13N_34 (number rated moderate/major obstacle).

A quick look at results, 2013, 2014, and 2015 as shown in the first couple.

AAU13a – money, work obligations, finances – by far highest rating each year

Same as in v1 (which was “pick one”), where money, work, finances = 38%, with lack
of personal motivation 15%, rest all 4-6%.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
13
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm




All others alts 13b-g similar to one another each year. 13f (poor academic performance)
usually the lowest rating.Avg of 1.6 items (of 7 max) were listed as 3 or 4
(moderate/major); avg of 3 were listed as not an obstacle.
In v1, about 27% checked the option “no real obstacles.” In v2, only 5% rated all 7 items
as 1=not an obstacle, but 26% rated none of the 7 as 3-4/mod/major. So v2 is definitely
yielding more discrimination.
V2 items 13a-g are positively correlated. So it’s NOT “if the kid has obstacle X he
doesn’t have obstacle Y” – it’s just the opposite. Strongest correlations

a-b – money with family

c-d – advising with getting courses
In my opinion the obstacle data were useful before and are terrific now with great
potential utility.
VSA

http://nsse.iub.edu/html/vsa.cfm explains the role of NSSE in VSA (Voluntary System of
Accountability) “assessment.”

If a school uses NSSE to satisfy VSA, it reports “results for seniors within six specified areas
that academic research has shown to be correlated with greater student learning and
development: group learning, active learning, experiences with diverse groups of people
and ideas, student satisfaction, institutional commitment to student learning and success,
and student interactions with faculty and staff.”

The webpage links to lists of items in v1, and in v2, grouped into the six areas, showing what
responses are to be grouped (e.g. for SEacademic, “Very much; Quite a bit; Some (Exclude
Very Little)”).

The figures required for entry to VSA can be produced from the nsseallyrs_v2 or long1
datasets using a tabulate or freq with formats to do the groupings of response alternatives.

Sample, what VSA looks like with NSSE 2.0 data/items:
http://www.collegeportraits.org/CO/CU-Boulder/student_experiences
Things CU-Boulder did in creating the v2.0 response-level dataset

Started with SAV (SPSS) file from NSSE

Made no modifications to ROWS – all rows sent, are still in even though some are (by NSSE’s
categorization) “ineligible to take survey.” [Probably individuals removed from the
population by a school who nevertheless got a survey invitation and responded.]

Calculated COMPARE, the very important variable carried from AAUDE v1. COMPARE = 1
marks rows that were first-year or senior, random sample, and eligible. Almost but not
quite all rows are COMPARE = 1.

Added the name of the institution (variable INST).

Moved various important variables, including INST and COMPARE, to the top/front of the
dataset

Located two NSSE-provided vars which are unique to respondent and match to the samenamed vars in the datasets NSSE provided directly to institutions: PopID and SurveyID.
These two have been moved near top of the dataset.

Developed the formats working backwards from SPSS value labels, but created just one
format for each unique set of response alternatives, rather than one per item
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
14
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm





Created a new set of labels using the 2-character variable prefixes that NSSE used to
indicate which “Engagement Indicator” (aka scale, no longer known as benchmarks) an item
contributes to. Both old and new labels are in the Excel.
Added, to NSSE numeric codes for majors 1 and 2

The NSSE description of the code

NSSE’s up-to-three CIP codes. I didn’t even look at these, just used them. These are at
http://nsse.iub.edu/html/commonDataRecodes.cfm because I asked NSSE to do them.
Let me or NSSE know if changes are in order.
DROPPED some vars NSSE sent us:

Institutional student ID number!!! If you want to match to your own student records
you can using SurveyID and other datasets from NSSE. NSSE will not provide AAUDE
with institutional student IDs in future.

GROUP variables which contained institution-provided data that varied from school to
school. Most schools didn’t use these, and the info was useless and together they were
over 2000 characters so we dropped!

The write-in major(s). Again, lots of characters. NSSE coded the write-ins to major
codes, which ARE in the dataset.
Looked for but did NOT FIND two sets of NSSE variables

The experimental items on course evaluation. No AAUDE received these items. No
school with two modules or consortium and module got them.

The student comments
We LEFT IN items in the NSSE modules (advising, writing, etc.). Some of them had no
AAUDE users so every item is 100% missing. PARTIC Excel shows who used what module,
and below. We put the module abbreviation on the labels of all these items.
Longitudinal.sas also drops the 118 variables from the NSSE modules. Or use this drop:
drop adv0: div0: trn0: civ0: tec0: wri0: gpi0: inl0: ;
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
15
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm



List of all modules can be found at http://nsse.iub.edu/html/modules.cfm
2013 modules and what data we have (publics)

ADV – Academic Advising : Michigan State, Ohio State

CIV – Civic Engagement - none

DIV – Experience with Diverse Perspectives – none

TEC – Learning with Technology - none

TRN – Development of Transferable Skills – Iowa State, Nebraska

WRI - Writing – Experiences with Writing - Colorado

None – Kansas, Washington
2014 modules and what data we have (publics)

ADV – Academic Advising : Boston U, Georgia Tech, McGill, SUNY-Stony Brook

CIV – Civic Engagement - none

DIV – Experience with Diverse Perspectives – Wisconsin

GPI – Global Perspectives (new) – Illinois

INL – Information Literacy (new) – none

TEC – Learning with Technology - none

TRN – Development of Transferable Skills –Maryland and SUNY-Buffalo

WRI - Writing – Experiences with Writing – none


2015 modules and what data we have (publics).

ADV – Academic Advising : Illinois, Indiana, Oregon, Syracuse

CIV – Civic Engagement - Tulane

DIV – Experience with Diverse Perspectives – Missouri

FYS – First Year Experiences & Senior Transitions (new) – Indiana, Rutgers

GPI – Global Perspectives – Case Western

INL – Information Literacy - Arizona

TEC – Learning with Technology - none

TRN – Development of Transferable Skills – Case Western, Illinois

WRI - Writing – Experiences with Writing – none
Creating the v1_v2 dataset (in 2013; we didn’t recheck all this)
– Started with NSSE item by item documentation on what changed. Coupled that with our over
time record of variable names etc. Built matchVars_v2.xlsx. Used that to write code to convert
varnames and drop v1 vars not carried to v2. Checked distributions of responses before and
after. Iterated time and again. Etc.

How checked: take all publics from 2013, get their single most recent admin 2009 and later.
Check. Drop Ohio State, diff pop in 2013. Drop Michigan State, response rate 2013 was half
of prior. (Colorado *N* is < prior but that was due to sampling in 2013.) Keep COMPARE=1
obs, 12,000 v1, 10,300 v2. From Colo, Iowa State, Kansas, Nebraska, UW. About equal
except Kansas smaller. Resulting N’s used in comparison checks:
INST(--Institution Name (KEY))
YEAR(--Survey Year (KEY))
Frequency |
2009|
2010|
2011|
2013|
-----------+--------+--------+--------+--------+
Colorado
|
2896 |
0 |
0 |
1915 |
-----------+--------+--------+--------+--------+
Iowa St
|
0 |
0 |
2282 |
2615 |
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
Total
4811
4897
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
16
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm
-----------+--------+--------+--------+--------+
Kansas
|
0 |
1605 |
0 |
1231 |
-----------+--------+--------+--------+--------+
Nebraska
|
0 |
2557 |
0 |
2072 |
-----------+--------+--------+--------+--------+
Washington |
0 |
0 |
2782 |
2495 |
-----------+--------+--------+--------+--------+
Total
2896
4162
5064
10328


2836
4629
5277
22450
Details:

NSSE created an item by item comparison of v1 and v2, and categorized change – none,
major, minor, dropped, new, and whether change involved stem, response alternatives,
or both. We used this to create ItemComparisonLongitudinal Excel, and added some
info based on our looks, about AAUDE questions, etc.

For all years of v1, Boulder had MatchVars http://www.colorado.edu/pba/surveys/nsseaaude/MatchVars.xls , tracking each variable over time.

We put the item-comparisons together with MatchVars info on v1 to create
MatchVars_v2 Excel, with Match_v2 tab.

The Match_v2 tab is the master source for translating v1 data to v2 formats.

Ultimately Match_v2 produced Match dataset, which is the source of SAS-written keep,
drop, format, and rewrite statements. These are used, and more complex recodes
done, in convert_v1_v2.sas
Troublesome vars in the conversion

LIVING – where living – response alts changed order, recode required, done

DDRACE –

V2: q 8a, During the current school year, about how often have you had discussions
with people from the following groups? - People of a race or ethnicity other than
your own.

V1: How often… Had serious conversations with students of a different race or
ethnicity than your own?

Level up significantly at every school. E.g. UW 2.79 to 3.3. Others similar.

Decided – DROP old. Do not equate. Change in wording and change in context
(from 1 of 2 items, to first of 4-5) are probably source of level change. Would be
improper to equate and infer that real activity increased significantly at every
school.

DURATION

V1 has missing and some value=0. V2, 2013, has no missing and more value=0. Also
value 1, 2 etc. = minutes. Most with short durations answered few questions. Bet
NSSE handled < 1 min differently in v1 and v2, but rest of data are OK, don't try to
fix.

2014 has missing duration, wildly long durations well over 500 minutes. No zero
values but several under 0.05 minutes.

2015 has NO missing duration, over 1,200 cases of wildly long duration over 500
minutes, no zero values but 100 cases 0.20 minutes and less.
Adding a new year
We account for any changes in variable names with both the V2 dataset and the Convert_V1_V2
dataset. Other processing is similar to 2013.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
17
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm
2015 changes/updates

Variables added by NSSE but not added to the all years dataset

BROWSER_FIRSTLOGIN, BROWSER_LASTLOGIN, OS_FIRSTLOGIN, OS_LASTLOGIN. The
browser/operating system used by the respondent when beginning/finishing the survey.

Question 31a on international student status. NSSE added an additional question on
country of citizenship if the student answered ‘yes’ to being an international student. Two
variables were added to the NSSE dataset that we also added to the all years dataset.

COUNTRY and COUNTRYCOL (region, eight categories). Both are coded numerically so
use the formats for text values.

Recoded variables added by NSSE but not added to the all years dataset. These variables
recode items that have a ‘Not Applicable’ response from missing to a value of 9. We
decided these were not needed.

This occurred exclusively on question 13: indicate the quality of your interactions with
the following people at your institution. QIstudentr, QIadvisorr, QIfacultyr, QIstaffr, and
QIadminr.

Also in the Academic advising module. ADV02AR – ADV02IR.

The variable POPID (unique population number) was removed by NSSE. This variable was
previously in 2013 and 2014 but will no longer be provided. We’ve removed this variable
from the all years dataset. The var SURVEYID is unique within year/institution (at least) and
can be used to merge to your own datasets.

Modules

New module on first-year experience and senior transitions. All of these variables begin
with FYSfy or FYSsr.

Academic advising module has a new question ADV04_15. How often have your
academic advisors reached out to you about your academic progress or performance?
We’ve added this to the dataset.

2014 changes/updates

Gender and sexual orientation q’s changed 2013 to 2014. See
http://websurv.indiana.edu/nsse/interface/img/Changes%20to%20Demographic%20Items
%20(NSSE%202014).pdf

Gender was dropped and replaced by GenderID by NSSE. Response values of
‘Another gender identity’ and ‘Prefer not to respond’ were added.

Decided – in AAUDE/NSSE data, equate old to new since only 0.2% chose
‘Another gender identity’ in 2014. The new value has been added to the format.

V1 datasets updated to reflect change. Gender replaced by GenderID

NSSE dropped the variable sexorient (Sexual orientation) and replaced by
sexorient14. Response value of ‘Another sexual orientation’ was added.

Decided – in AAUDE/NSSE data, equate old and new but kept variable name as
sexorient since less than 0.5% chose ‘Another sexual orientation’ in 2014. The
new value has been added to the format.

IRGender (institutional reported gender) name changed by NSSE to IRSex. Response values
were not changed.

We adopted the new variable name. V1 datasets updated to reflect the name change.
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
NSSE-AAUDE Data User Guide – For NSSE v2 and NSSE v1 converted to v2
18
University of Colorado Boulder - Last update 11/2015
Posted from http://www.colorado.edu/pba/surveys/nsse-aaude/index.htm




Question 16 on ‘time spent preparing for class in a typical week about how many hours are
on assigned reading’ changed response values from an hours estimate to a descriptive scale
(very little to almost all).

Decided – could not equate so added variable READING (NSSE’s name; for 2014 and
later only) and kept the old var TMREAD for 2013 data only.

Caution – looking at estimated time per week on reading assignments using 2013 vs
2014 and later data will be tricky at best and probably impossible.
Question on duration (in minutes) it took to answer the AAU questions changed from
AAUDUR to DURATIONAAU.
Three modules were added.

Global Perspectives – Cognitive and Social. Illinois answered these in 2014

Experiences with Information Literacy. No AAU publics answered.
First-year Experiences and Senior Transitions were listed by NSSE but the variables were not
part of the NSSE dataset (we’re assuming no institutions answered these questions)
NOTES, CAUTIONS specific to ONE YEAR
2013

Ohio State senior population.

N in the population dropped by 72% while N in frosh population dropped by only 14%.
This was due to OSU change in definition of senior pop, restricting to those who’d
declared intent to graduate by about October prior to spring.

Ratio of Sr:Fr now different for OSU 2013 than for all others.

Therefore CAUTION using OSU seniors

I’ve corresponded with Julie and others and recommended change in definition next
time
2014

No known issues. Just watch out for change in name of gender variables.
2015 No known issues
PBA file refs and summary of Boulder to do’s
L:\IR\survey\NSSE\AAUDE\All\UserGuide_v2.docx
Scavenged from
L:\IR\survey\NSSE\AAUDE\All\NSSE_AAUDE_v2_release2.docx
L:\IR\survey\NSSE\AAUDE\All\archive\UserGuide_v1toscavenge.docx
And copied here/memo from L:\IR\survey\NSSE\AAUDE\NSSE_AAUDE_201311.docx
CU-Boulder PBA:Lou McClelland –Document1 – 7/12/2016
Download