2020/25 Beginning Postsecondary Students (BPS:20/25) Field Test
Supporting Statement Part B
OMB # 1850-0631 v.20
Submitted by
National Center for Education Statistics
U.S. Department of Education
Contents
B. Collection of Information Employing Statistical Methods 3
b. BPS:20/22 Field Test Sample 3
c. BPS:20/25 Field Test Sample 5
d. BPS:20/22 Full-scale Sample 6
e. BPS:20/25 Full-scale Sample 6
3. Methods for Maximizing Response Rates 7
a. Tracing of Sample Members 7
b. Training for Data Collection Staff 8
e. Refusal Aversion and Conversion 10
4. Tests of Procedures or Methods 10
a. Experiment #1, Data Collection: Address Confirmation Incentive 11
b. Experiment #2, Data Collection: CATI Reminders 11
c. BPS:20/25 Field Test Data Collection Procedures 12
5. Reviewing Statisticians and Individuals Responsible for Designing and Conducting the Study 15
Tables and Figures
This submission requests clearance for the 2020/25 Beginning Postsecondary Students Longitudinal Study (BPS:20/25) field test data collection materials and procedures. BPS:20/25 is the second follow-up of sample members drawn from the 2019-20 National Postsecondary Student Aid Study (NPSAS:20). Specifically, those who began their postsecondary education during the 2018-19 (field test sample) or 2019-20 (full-scale sample) academic year. For details on the NPSAS:20 sampling design see NPSAS:20 Supporting Statement Part B (OMB# 1850-0666 v.25). Specific plans for the BPS:20/25 data collection are provided below.
The respondent universe for BPS:20/25 consists of all students who
began their postsecondary education for the first time during the
2018-19 (field test sample) or 2019-20 (full-scale sample) academic
year at any Title IV-eligible postsecondary institution in the
United States.
In prior BPS data collections, a NPSAS
field test, conducted a year before the full-scale, has been used to
identify a field test sample for the BPS longitudinal follow up. For
example, the NPSAS:12 field test was the source of the sample for
the BPS:12/14 field test. However, because NPSAS:20 did not conduct
a field test, the BPS:20/22 field test sample was drawn from the
NPSAS:20 full-scale sample. The BPS:20/25 sample is created from the
BPS:20/22 sample.
While the BPS:20/25 full-scale cohort
will be comprised of students who first enroll in postsecondary
education after high school during the 2019-20 academic year, the
BPS:20/25 field test cohort will follow students whose first
postsecondary enrollment after high school occurred during the
2018-19 academic year. This is necessary to test the temporal
components of the survey, which ask about enrollment and experiences
during academic years 4-6 after their first enrollment. The
BPS:20/25 field test sample will include students from the NPSAS:20
full-scale sample who were identified as confirmed or potential
2018-19 academic year first-time beginner students based on survey,
institution, or other administrative data.
The BPS:20/25 field test will be implemented to test procedures, methods, and systems planned for the BPS:20/25 full-scale study in a realistic operational environment prior to implementation in the full-scale study. The field test is designed to test and validate data collection and monitoring procedures that will obtain the most accurate data in the least amount of time. Specific field test evaluation goals for the student survey include the following:
Identifying problematic data elements in the BPS:20/25 student interview
Testing two data collection experiments aimed at reducing nonresponse error and the potential for nonresponse bias:
an “address confirmation” incentive experiment where respondents receive an additional $5 if they confirm their phone number, mailing and e-mail address before the start of the survey, and
a reminder mode experiment where we explore the implications of dropping telephone reminders for sample members who have responded to previous survey requests.
Additionally, we will evaluate the time required to complete the survey and identify possible modifications to reduce respondent burden in the full-scale survey, if needed. We will also conduct reliability re-interviews to evaluate the temporal consistency of selected interview items.
The sample for the BPS:20/25 field test is built upon the sample from the prior field test data collection with this cohort, BPS:20/22. Individuals who self-identified as first-time beginners (FTBs) in the 2018-19 academic year (i.e., prior-year FTBs) in the NPSAS:20 survey, and those identified as potential prior-year FTBs using administrative data, were eligible for selection for the BPS:20/22 field test. A field test sample of approximately 3,700 students was identified through enrollment lists, administrative data and survey responses collected during the NPSAS:20 full-scale study. The BPS:20/22 field test sample consisted of four distinct groups based on the NPSAS:20 full-scale data collection outcomes:
1) NPSAS:20-eligible survey respondents who self-identified as FTBs between July 1, 2018 and April 30, 2019 (approximate n = 2,440).
2) NPSAS:20-ineligible survey respondents who self-identified as FTBs between July 1, 2018 and April 30, 2019 (approximate n = 180). This constituted an unusual group because they were not enrolled during the NPSAS:20 academic year but were enrolled during the prior year and have either completed a short-term credential, stopped or dropped out. Lack of enrollment during the NPSAS:20 academic year made them ineligible for NPSAS:20. However, because a BPS first follow-up data collection included students who were not enrolled one year after they first begin postsecondary education, this group was useful in field testing BPS:20/22 methods and procedures across different eligible populations. While these students were not eligible for NPSAS:20, they were asked to complete a subset of survey questions to confirm FTB status in 2018-19. If found eligible, we notified them about the potential for follow up and administered the locating section of the NPSAS:20 survey so that they could be contacted and asked to participate in the BPS:20/22 field test.
3) NPSAS:20 survey nonrespondents who were potential academic year 2018-19 FTBs based on administrative data (approximate n = 540).
4) NPSAS:20 sample members who were not invited to participate in the survey (administrative-only individuals) who were also potential academic year 2018-19 FTBs based on administrative data (approximate n = 540).
For the first group, NPSAS:20-eligible survey respondents, rather than relying on random sampling from 2018-19 academic year FTB students in the NPSAS:20 sample, the BPS:20/22 field test used a quota sampling approach to reduce burden on NPSAS:20 respondents. Potentially eligible BPS:20/22 field test sample members received an additional question module targeted at 2018-19 FTB status in the NPSAS:20 instrument. This set of survey items took approximately 10 minutes to administer. Once quotas were met, the administration of the additional items was stopped. In contrast, a random sampling approach would have resulted in additional NPSAS:20 respondents being administered additional items unnecessarily. Final BPS:20/22 field test sample counts included here differ slightly from those described in the BPS:20/22 OMB package, since that package was prepared prior to implementation of the quota sampling approach. Table 1 summarizes the BPS:20/22 sampled, eligible, and responding individuals by their NPSAS:20 full-scale data collection outcomes. The count of eligible students was updated after BPS:20/22 data collection concluded. A derived variable calculating month and year of first postsecondary enrollment identified those with enrollment before the targeted academic year, thus reducing the total eligible sample. This derived variable will also be applied to the full-scale sample. This process allows for a detailed reporting of eligibility removing the need for an eligibility screener in the BPS:20/25 field test and full-scale data collections.
Table 1. Counts of BPS:20/22 field test sampled and eligible students and response rates, by NPSAS:20 full-scale data collection outcomes: 2020–22
NPSAS:20 full-scale data collection outcomes |
Sampled students |
Eligible students1 |
Respondents |
Response Rate |
Total |
3,700 |
3,280 |
2,120 |
0.65 |
NPSAS:20-eligible survey respondent |
2,440 |
2,310 |
1,800 |
0.78 |
NPSAS:20-ineligible survey respondent |
180 |
160 |
110 |
0.70 |
NPSAS:20-eligible survey nonrespondent |
540 |
430 |
50 |
0.12 |
NPSAS:20-eligible administrative only |
540 |
370 |
150 |
0.42 |
1
Sample member eligibility was determined during the student screener
combined with derived variables based on institutional records. The
use of derived month and year of first postsecondary enrollment was
included to align with the approach used for full-scale data
collection and to remove the need for an eligibility screener during
BPS:20/25 field test data collection.
NOTE: Sample sizes
rounded to the nearest 10. Percentages are based on the unrounded
count of eligible students within the row under consideration.
Detail may not sum to totals because of rounding.
SOURCE: U.S.
Department of Education, National Center for Education Statistics,
2020/22 Beginning Postsecondary Students Longitudinal Study
(BPS:20/22).
The BPS:20/25 field test is the second follow-up data collection, to be conducted 3 years after BPS:20/22 field test. The BPS:20/25 field test sample is constructed from the BPS:20/22 field test sample. The main source of change from the BPS:20/25 field test sample to the BPS:20/22 field test sample is the exclusion of sample members determined to not be FTB students and therefore ineligible for BPS. Deceased individuals are also excluded from the BPS:20/25 field test sample. Table 2 shows the distribution of the approximately 3,280 sample members determined to be eligible, by study member status and survey response status, and identifies groups that will be fielded.
Table 2. BPS:20/25 sample member status, by NPSAS:20 and BPS:20/22 field test response status and known eligibility status: 2020–25
Group |
NPSAS:20 Survey Respondent |
BPS:20/22 Field Test Respondent |
Known Eligibility1 |
Count |
Fielded in BPS:20/25 Field Test |
Total |
|
|
|
3,280 |
|
1 |
Yes |
Yes |
Yes |
1,910 |
Yes |
2 |
Yes |
No |
Yes |
550 |
Yes |
32 |
Yes |
No |
No |
20 |
No |
4 |
No |
Yes |
Yes |
200 |
Yes |
53 |
No |
No |
Yes |
370 |
No |
63 |
No |
No |
No |
230 |
No |
1
Eligibility remains unknown if the individual did not have enough
administrative data and did not complete the BPS:20/22 field test
eligibility screener.
2 Group 3 not fielded as these
individuals were NPSAS:20 ineligible and do not have enough
administrative data to determine if they are FTBs.
3 Groups 5
and 6 are not included as they did not respond to either the
NPSAS:20 survey or the BPS:20/22 field test survey.
NOTE:
Sample sizes rounded to the nearest 10.
SOURCE: U.S. Department
of Education, National Center for Education Statistics, 2020/22
Beginning Postsecondary Students Longitudinal Study (BPS:20/22).
While BPS:20/25 field test sample members who did not respond to either the base-year NPSAS:20 survey or the BPS:20/22 field test survey are eligible for the BPS:20/25 field test, these sample members will not be fielded due to low expected participation rates. This results in approximately 2,660 BPS:20/25 field test sample members to be fielded.
Based
on past second follow-up BPS field test data collections, we expect
a response rate of 70 percent. Because we are only including
individuals known to be eligible for the BPS:20/25 field test
sample, we have an overall eligibility rate of 100 percent. Table 3
below displays the BPS:20/25 field test sample size by NPSAS:20 and
BPS:20/22 field test data collection outcome groups and the expected
yield of completed interviews.
Table 3. BPS:20/25 Field Test Expected Completes by NPSAS:20 and BPS:20/22 Field Test Data Collection Outcome Groups
NPSAS:20 Survey Respondent |
BPS:20/22 Field Test Respondent |
Sample Size |
Expected Response Rate |
Expected Completes |
Total |
|
2,660 |
0.70 |
1,880 |
Yes |
Yes |
1,910 |
0.85 |
1,620 |
Yes |
No |
550 |
0.28 |
150 |
No |
Yes |
200 |
0.50 |
100 |
NOTE:
Sample sizes rounded to the nearest 10.
SOURCE: U.S.
Department of Education, National Center for Education Statistics,
2020/22 Beginning Postsecondary Students Longitudinal Study
(BPS:20/22).
The BPS:20/25 full-scale sample will be constructed from the BPS:20/22 full-scale sample. The BPS:20/22 sample was composed of students who began their postsecondary education for the first time during the 2019-20 academic year at any Title IV-eligible postsecondary institution in the United States. Approximately 26,470 respondents to the NPSAS:20 full-scale student surveys self-identified as FTBs. In addition, to have full population coverage of the BPS sample, a subsample of approximately 10,860 students who are indicated by the NPSAS sample institution or administrative data to be potential FTBs were selected. This subsample consisted of a combination of NPSAS:20 survey nonrespondents and administrative data-only sample members who were NPSAS:20 study respondents.
As shown in Table 4 below, the starting full-scale sample size was approximately 37,330, with an eligible sample of about 34,240. Eligibility was determined using screener responses and administrative data and deceased individuals are also removed from eligible individual counts. We observed an unweighted response rate of 65 percent among the eligible sample members, which yielded approximately 22,350 responding FTBs.
Table 4. BPS:20/22 full-scale study eligibility and unweighted response rates by base-year outcome status: 2020-22
|
Number |
Eligible1 |
|
Respondents |
||
NPSAS:20 full-scale outcome |
Number |
Rate |
Number |
Rate2 |
||
Total |
37,330 |
34,240 |
92% |
|
22,320 |
65% |
NPSAS Survey respondents |
26,470 |
25,230 |
95% |
|
18,900 |
75% |
NPSAS Survey nonrespondents |
5,510 |
4,470 |
81% |
|
980 |
22% |
NPSAS Administrative-only students |
5,350 |
4,540 |
85% |
|
2,450 |
54% |
1
Sample member eligibility was determined during the student survey
or from institutional records in the absence of a student survey.
Individuals found to be deceased during data collection have also
been removed from the total eligible sample members.
2
Unweighted response rate.
NOTE: Sample sizes rounded to the
nearest 10. Percentages are based on the unrounded count of eligible
students within the row under consideration. Detail may not sum to
totals because of rounding.
SOURCE: U.S. Department of
Education, National Center for Education Statistics, 2020/22
Beginning Postsecondary Students Longitudinal Study (BPS:20/22).
BPS:20/25 will be the second follow-up data collection, conducted three years after BPS:20/22. The major difference from the BPS:20/22 sample to the BPS:20/25 sample is the exclusion of BPS:20/25 sample members who were determined to not be FTB students and are therefore ineligible for BPS. Deceased individuals are also excluded from the BPS:20/22 sample. Table 5 shows the distribution of the 34,240 sample members determined to be eligible for BPS:20/25 data collection, by study member status and survey response status, and identifies groups that will be fielded (included in data collection activities with an objective of obtaining a complete survey) in BPS:20/25. These 34,240 sample members are the same individuals identified as eligible students in Table 4.
Table 5. BPS:20/25 sample member disposition, by NPSAS:20 and BPS:20/22 response status: 2020–22
Group |
NPSAS:20 data collection outcomes |
BPS:20/22 Respondent |
Count |
Field in BPS:20/25 Full-scale |
Total |
|
|
34,240 |
|
1 |
NPSAS Survey respondents |
Yes |
18,900 |
Yes |
2 |
NPSAS Survey nonrespondents |
Yes |
980 |
Yes |
3 |
NPSAS Administrative-only students |
Yes |
2,450 |
Yes |
4 |
NPSAS Survey respondents |
No |
6,330 |
Yes |
51 |
NPSAS Survey nonrespondents |
No |
3,490 |
No |
61 |
NPSAS Administrative-only students |
No |
2,100 |
No |
1
Groups 5 and 6 are not included as they did not respond to either
the NPSAS:20 survey or the BPS:20/22 survey.
NOTE: Sample sizes
rounded to the nearest 10. Detail
may not sum to totals because of rounding.
SOURCE:
U.S. Department of Education, National Center for Education
Statistics, 2020/22 Beginning Postsecondary Students Longitudinal
Study (BPS:20/22).
While BPS:20/25 sample members who did not respond to either the NPSAS:20 student survey nor the BPS:20/22 survey are eligible for BPS:20/25, these sample members (groups 5 and 6 in Table 5) will not be fielded as they have responded at very low rates in previous administrations. This results in approximately 5,590 BPS:20/25 sample members not being fielded. Instead, they will be treated as study nonrespondents for purposes of response rate calculation and accounted for with weight adjustments.
Table 6 presents the BPS:20/25 sampled and expected responding individuals, by NPSAS:20 outcome and BPS:20/22 response status. Based on administrative data, all BPS:20/25 sample members are considered eligible. The response rate estimates are based on the BPS:20/22 data collection and the BPS:12 longitudinal cohort that followed the NPSAS:12 data collection (Bryan et al. 2016).
Table 6. Expected BPS:20/25 full-scale study response rates by base-year and BPS:20/22 outcome status: 2020-22
NPSAS:20 data collection outcomes |
BPS:20/22 Respondent |
Sample Size |
Expected Response Rate |
Expected Completes |
Total |
|
34,240 |
0.59 |
20,020 |
NPSAS Survey respondents |
Yes |
18,900 |
0.81 |
15,310 |
NPSAS Survey nonrespondents |
Yes |
980 |
0.55 |
540 |
NPSAS Administrative-only students |
Yes |
2,450 |
0.75 |
1,830 |
NPSAS Survey respondents |
No |
6,330 |
0.37 |
2,340 |
NPSAS Survey nonrespondents1 |
No |
3,490 |
0.00 |
0 |
NPSAS Administrative-only students1 |
No |
2,100 |
0.00 |
0 |
1
Expected response rate for these groups is zero as they will not be
fielded for data collection.
NOTE: Sample sizes rounded to the
nearest 10.. Detail may not sum to totals because of
rounding.
SOURCE: U.S. Department of Education, National Center
for Education Statistics, 2020/22 Beginning Postsecondary Students
Longitudinal Study (BPS:20/22).
Achieving high response rates in the BPS:20/25 field test study data collection will depend on successfully identifying and locating sample members and being able to contact them and gain their cooperation. As was used successfully in prior NCES longitudinal studies, shortly before data collection begins, we will send an initial contact mailing/e-mail to remind sample members of their inclusion in the study.
To yield the maximum number of located cases with the least expense, we designed an integrated tracing approach with the following elements. Advance tracing activities, which will occur prior to the start of data collection, include initial batch database searches, such as to the National Change of Address (NCOA) databases, for cases with sufficient contact information to be matched. To handle cases for which contact information is invalid or unavailable, additional advance tracing through proprietary interactive databases will expand on leads found.
Hard copy mailings, e-mails, and text messages will be used to maintain ongoing contact with sample members, prior to and throughout data collection. The contacting materials, which will be developed with a design appealing to sample members in 2024, are provided in Appendix C. The data collection mailing to sample members will include a letter announcing the start of data collection, requesting that the sample member complete the web survey, and including a toll-free number, the study website address, a Study ID and password, and a study brochure. We will send a similar e-mail message mirroring the information provided in the mailing.
Sample members will have a variety of means to provide updated contact information and contact preferences. Students can use an Update Contact Information page on the secure BPS:20/25 website to provide their contact information, including cell phone number, as well as provide contacting preferences with respect to phone calls, mailings, e-mails, and text messages. Help Desk calls and e-mails providing information about a sample member’s text message preferences will be monitored and the sample member’s data record updated as soon as the information becomes known.
The BPS:20/25 field test will include an experiment where a randomly selected group of sample members will be asked to confirm contacting information at the start of the data collection, for which they will be paid $5. Once they have provided the contact information, they can continue with the survey.
The telephone locating and surveying stage includes calling all available telephone numbers and following up on leads provided by parents and other contacts.
The pre-intensive batch tracing stage consists of the LexisNexis SSN and Premium Phone batch searches that will be conducted between the telephone locating and surveying stage and the intensive tracing stage.
Once all known telephone numbers are exhausted, a case will move into the intensive tracing stage during which tracers will conduct interactive database searches using all known contact information for a sample member. With interactive tracing, a tracer assesses each case on an individual basis to determine which resources are most appropriate and the order in which each should be used. Sources that may be used, as appropriate, include credit database searches, such as Experian, various public websites, and other integrated database services.
Other locating activities will take place as needed, including a LexisNexis e-mail search conducted for nonrespondents toward the end of data collection.
Telephone data collection will include supervisors and interviewers. Training programs for these staff members are critical to maximizing response rates and collecting accurate and reliable data.
Team supervisors, who are responsible for all supervisory tasks, will attend their own project-specific training, in addition to the interviewer training. They will receive an overview of the study, background and objectives, and the data collection instrument through a question-by-question review. Supervisors will also receive training in the following areas: providing direct supervision during data collection; handling refusals; monitoring interviews and maintaining records of monitoring results; problem resolution; case review; specific project procedures and protocols; reviewing reports generated from the ongoing Computer Assisted Telephone Interviewing (CATI); and monitoring data collection progress.
Training for interviewers is designed to help staff become familiar with and practice using the CATI case management system and the survey instrument, as well as to learn project procedures and requirements. Particular attention will be paid to quality control initiatives, including refusal avoidance and methods to ensure that quality data are collected. Interviewers will receive project-specific training on telephone interviewing and answering questions from web participants regarding the study or related to specific items within the survey. At the conclusion of training, all data collection staff must meet certification requirements by successfully completing a certification interview. This evaluation consists of a full-length interview with project staff observing and evaluating interviewers, as well as an oral evaluation of interviewers’ knowledge of the study’s Frequently Asked Questions.
Surveys will be conducted using a single web-based survey instrument for both web (including mobile devices) and CATI data collection. Control of data collection activities will be accomplished through a CATI case management system, which is equipped with the numerous capabilities, including: on-line access to locating information and histories of locating efforts for each case; a questionnaire administration module with full “front-end cleaning” capabilities (i.e., editing as information is obtained from respondents); sample management module for tracking case progress and status; and automated scheduling module which delivers cases to interviewers. The automated scheduling module incorporates the following features:
Automatic delivery of appointment and call-back cases at specified times. This reduces the need for tracking appointments and helps ensure the interviewer is punctual. The scheduler automatically calculates the delivery time of the case in reference to the appropriate time zone.
Sorting of non-appointment cases according to parameters and priorities set by project staff. For instance, priorities may be set to give first preference to cases within certain sub-samples or geographic areas; cases may be sorted to establish priorities between cases of differing status. Furthermore, the historic pattern of calling outcomes may be used to set priorities (e.g., cases with more than a certain number of unsuccessful attempts during a given time of day may be passed over until the next time period). These parameters ensure that cases are delivered to interviewers in a consistent manner according to specified project priorities.
Restriction on allowable interviewers. Groups of cases (or individual cases) may be designated for delivery to specific interviewers or groups of interviewers. This feature is most commonly used in filtering refusal cases, locating problems, or foreign language cases to specific interviewers with specialized skills.
Complete records of calls and tracking of all previous outcomes. The scheduler tracks all outcomes for each case, labeling each with type, date, and time. These are easily accessed by the interviewer upon entering the individual case, along with interviewer notes.
Flagging of problem cases for supervisor action or supervisor review. For example, refusal cases may be routed to supervisors for decisions about whether and when a refusal letter should be mailed, or whether another interviewer should be assigned.
Complete reporting capabilities. These include default reports on the aggregate status of cases and custom report generation capabilities.
The integration of these capabilities reduces the number of discrete stages required in data collection and data preparation activities and increases capabilities for immediate error reconciliation, which results in better data quality and reduced cost. Overall, the scheduler provides an efficient case assignment and delivery function by reducing supervisory and clerical time, improving execution on the part of interviewers and supervisors by automatically monitoring appointments and call-backs, and reducing variation in implementing survey priorities and objectives.
The
survey will employ a web-based instrument and deployment system,
which has been in use since NPSAS:08. The system provides multimode
functionality that can be used for self-administration, including on
mobile devices, CATI, or data entry. The survey instrument can be
found in Appendix E.
In
addition to the functional capabilities of the case management
system and web survey instruments described above, our efforts to
achieve the desired response rate will include using established
procedures proven effective in other large-scale studies. These
include:
Providing multiple response modes, including mobile-friendly self-administered and interviewer-administered options.
Offering incentives to encourage response.
Assigning experienced CATI interviewers who have proven their ability to contact and obtain cooperation from a high proportion of sample members.
Training the interviewers thoroughly on study objectives, study population characteristics, and approaches that will help gain cooperation from sample members.
Maintaining a high level of monitoring and direct supervision so that interviewers who are experiencing low cooperation rates are identified quickly and corrective action is taken.
Making every reasonable effort to obtain a completed interview at the initial contact, but allowing respondent flexibility in scheduling appointments to be interviewed.
Thoroughly reviewing all refusal cases and making special conversion efforts whenever feasible.
Implementing and assuring participants of confidentiality procedures, including:
Requiring respondents to answer security questions before obtaining and resuming access to the survey;
Restricting the ability for the respondent to view survey responses from prior log in sessions (i.e., no ability to use navigation buttons to go to “Previous” survey questions from another log in session); and
The survey automatically logging out of a session after 10 minutes of inactivity.
Recognizing and avoiding refusals is important to maximize the response rate. We will emphasize this and other topics related to obtaining cooperation during interviewer training. Supervisors will monitor interviewers intensely during the early days of outbound calling and provide retraining as necessary. In addition, the supervisors will review daily interviewer production reports produced by the CATI system to identify and retrain any data collectors who are producing unacceptable numbers of refusals or other problems.
Refusal conversion efforts will be delayed for at least 1 week to give the respondent time after the initial refusal. Attempts at refusal conversion will not be made with individuals who become verbally aggressive or who threaten to take legal or other action. Refusal conversion efforts will not be conducted to a degree that would constitute harassment. We will respect a sample member’s right to decide not to participate and will not impinge this right by carrying conversion efforts beyond the bounds of propriety.
The BPS:20/25 field test will include two data collection experiments focused on survey participation:
1) an “address confirmation” incentive experiment where respondents receive an additional $5 if they confirm their phone number, mailing and e-mail address before the start of the survey, and
2) a reminder mode experiment where we explore the implications of eliminating telephone reminders for sample members who have responded to NPSAS:20 and BPS:20/22.
The
following section describes the two data collection experiments,
followed by a description of the field test data collection
procedures. We will evaluate the generalizability of the results
from these experiments by assessing the representativeness of the
BPS:20/25 field test sample. Even limited generalizability will
provide valuable insight on the potential of implementing the
address confirmation incentive and reducing CATI reminders as
enhancements for the BPS:20/25
full-scale study.
Appendix
C provides materials designed for contacting student sample members
(e.g., letters, brochures, e-mails, text messages).
Input from NPSAS and BPS focus groups (Appendix D), open-ended questions in NPSAS and BPS surveys, as well as posts on social media platforms have indicated that some sample members question the legitimacy of NPSAS/BPS survey invitations and therefore may be hesitant to complete the survey. Previous NPSAS-family studies (e.g., NPSAS:20 calibration, NPSAS:20 full-scale, and BPS:20/22) have used prepaid incentives via PayPal in an attempt to increase sample members’ perceived legitimacy of the survey invitation. However, in these studies, the incentive was automatically deposited into a sample member’s PayPal account without any action from the sample member (i.e., separate from the survey invitation itself). This may have limited a sample member’s ability to make a connection between the incentive delivered to their PayPal account and the survey request.
Experiment 1 will build on this methodology by investigating the effects of offering sample members a monetary incentive in exchange for verifying their address information (e.g., phone number, mailing, and e-mail address) at the beginning of the survey. Immediately upon completion of this address confirmation (i.e., before the survey starts), sample members in the experimental group will receive a $5 incentive payment via their method of choice (PayPal or check), while sample members in the control group will receive no incentive for completing this confirmation. This immediate payment may increase the salience of the connection between the incentive and the survey request, increasing the legitimacy of the study: giving sample members $5 at the beginning of the survey may give them more confidence that they will receive their post-paid incentive at the end of the survey.
All BPS:20/25 field test sample members will receive a request to confirm their phone number, mailing address, and e-mail address. However, the incentive offer in exchange for completing this address confirmation involves splitting the BPS:20/25 field test sample into two randomly assigned groups:
Control Group (n = 1,331): Sample members will receive no incentive for completing the address confirmation.
Experimental Group (n = 1,331): Sample members will receive a $5 incentive if they complete the address confirmation.
This design will allow for a comparison of response rates among equally-sized experimental groups and provide enough power to detect a 4 percentage point difference in response rates assuming 80 percent power, type I error of 5 percent, and a base response rate of 66 percent. This calculation assumes a 2-sided chi-square test of response proportions.
The experiment described above will allow us to test the following hypotheses:
H1a. There is no statistically significant difference in response rates between the Control and the Experimental Group.
H1b. There are no statistically significant differences in representativeness (demographic characteristics) between the Control and the Experimental Group.
H1c. There are no statistically significant differences in timeliness of responses (days to complete) between the Control and the Experimental Group.
Survey interviews conducted via CATI calls represent a small percentage of completes in recent NPSAS-family studies (4 percent of completes in NPSAS:20, 5 percent of completes in the BPS:20/22 full-scale). CATI is also the most expensive mode as it requires a human interviewer. Eliminating CATI prompting for certain groups of sample members (e.g., sample members who have responded to survey requests in previous waves like NPSAS:20 and BPS:20/22) may present an opportunity to reduce data collection costs in BPS:20/25 without adversely affecting response rates or representativeness.
Experiment 2 will explore this idea in BPS:20/25 by investigating the effects of eliminating CATI reminders for sample members who are NPSAS:20 and BPS:20/22 interview respondents, excluding final partials (i.e., the default data collection group; see Field Test Data Collection Procedures below for more information about this data collection group). Eight weeks into data collection, nonresponding sample members in the default group will either receive CATI prompting as usual (control group) or not receive any CATI prompting (experimental group). All other data collection activities (e.g., sending reminder e-mails and texts) will continue throughout the experiment for both groups.
Eight weeks into data collection, nonresponding sample members in the default data collection group will be split into two randomly assigned groups:
Control Group (n ~= 368; assuming a 63 percent response rate at 8 weeks): Members of this group will receive telephone reminders until the end of data collection, in additional to all other reminders.
Experimental Group (n ~= 368; assuming a 63 percent response rate at 8 weeks): Members of this group will not receive telephone reminders but will receive all other reminders.
This design will allow for a comparison of response rates among roughly equally-sized experimental groups and provide enough power to detect a 7 percentage point difference in response rates assuming 80 percent power, type I error of 5 percent, and a base response rate of 34 percent. This calculation assumes a 2-sided chi-square test of the response proportions.
The experiment described above will allow us to test the following hypotheses:
H2a. There is no statistically significant difference in response rates between the Control and the Experimental Group.
H2b. There are no statistically significant differences in representativeness (demographic characteristics) between the Control and the Experimental Group.
BPS:20/25
field test data collection will use two distinct data collection
groups and three main data collection phases. This approach builds
upon the designs implemented in other longitudinal studies where it
has contributed to maximizing response rates and minimizing the
potential for nonresponse bias (e.g., BPS:12/14, BPS:12/17, BPS:
20/22, B&B:16/17, and B&B:08/18). In BPS:20/25, we plan to
implement differential treatments based on prior round response
status, an approach that was successfully implemented in the
B&B:16/17 field test, where NPSAS:16 field test nonrespondents
received either an aggressive or a default protocol. The response
rate among NPSAS:16 field test nonrespondents who received the
aggressive protocol was about 12 percentage points higher than the
group that received the default protocol (37 percent; default 25
percent response rate t(2,097) = 3.52, p < .001).
For
the BPS:20/25 field test design, we will use the following two data
collection groups:
Default Group: Any sample member that responded to all NPSAS:20 and BPS:20/22 survey requests (n = 1,989), including:
NPSAS:20 and BPS:20/22 survey respondents, excluding BPS:20/22 final partials (n = 1,741)
NPSAS:20 administrative only cases who were also BPS:20/22 survey respondents, excluding BPS:20/22 final partials (n = 142)
NPSAS:20 ineligible respondents who, via a screener, self-identified that they began their postsecondary education between July 1, 2018 and April 30, 2019, excluding BPS:20/22 final partials (n = 106).
Aggressive Group: NPSAS:20 survey nonrespondents (n = 50), BPS:20/22 survey nonrespondents (n = 547), or BPS:20/22 survey final partials (n = 76). The goal of this treatment is to convert reluctant sample members (i.e., those who have not responded to a previous survey request) to participate in the study as early in data collection as possible.
Table 7 below presents the type and timing of interventions to be applied in the field test data collection by groups and protocol and is described in more detail in the next section.
Table 7. BPS:20/25 field test data collection protocols by data collection phase and group assignment
|
Data Collection Group Assignments |
|
|
Default Group |
Aggressive Group |
Sample |
|
|
Data Collection Protocols |
|
|
Prior to data collection |
|
|
Early completion phase (Weeks 1-4)
|
|
|
Production Phase (Weeks 5-10)
|
|
|
Nonresponse Follow-Up (Weeks 11 – 16) |
|
|
Total incentives |
|
|
Turning to incentives, the baseline incentive for the prior year respondents in the default group will be $30. Including the possible $5 address confirmation incentive (if a sample member is in the experimental group discussed above), and a $10 boost postpaid incentive (see incentive boosts section below), yields a maximum total incentive of $45 for sample members in the default group.
The
baseline incentive for sample members in the aggressive group will
be $45. An experiment conducted in BPS:12/14 showed that a $45
baseline incentive yielded the highest response rates; however, this
was
underpowered
to detect differences from $30 in the lower propensity response
groups (Wilson et al. 2015). Nonetheless, the $30 baseline incentive
offered to these sample members in prior studies was not sufficient
to encourage response (i.e., n = 50 Field Test sample members did
not respond to NPSAS:20 and n = 623 did not provide complete
responses to BPS:20/22). Therefore, we recommend implementing a
higher baseline incentive given that the $30 baseline incentive was
not enough to encourage these sample members to respond in prior
years. Further, the $40 BPS:20/22 field test incentive yielded a
completion rate of only 23 percent among sample members in the
aggressive group, while a $45 baseline incentive for the aggressive
group in the BPS:20/22 full-scale yielded a completion rate of 38
percent. The baseline incentive will be paid in addition to a
possible $5 address confirmation incentive (if in the experimental
group as discussed above), and a $20 boost postpaid incentive (see
incentive boosts section below). The maximum possible total
incentive is $70 in this aggressive data collection protocol.
Beyond
the baseline incentives, both survey data collection protocols
employ similar interventions, although the timing of these
interventions differs across groups: interventions occur sooner in
the aggressive protocol.
Field
Test Data Collection Protocol Design Elements:
Greeting
card.
The first mailing that individuals in the default and aggressive
data collection protocols will receive is a greeting card expressing
gratitude for being part of the study and announcing the upcoming
data collection. Greeting cards have been shown to significantly
increase response rates in longitudinal studies (Griggs et al. 2019)
and we will use this method as a precursor to the invitation letter
for both groups. The greeting card will be mailed a few weeks in
advance of data collection, upon OMB approval.
Contact
information confirmation incentive.
We will implement a $5 contact information confirmation incentive
experiment as described above.
Reminders.
Text messaging has been shown to significantly increase response
rates in different survey modes (e.g., Callegaro et al. 2011;
Schober et al. 2015). Results from the BPS:20/22 field test showed
that the response rate for sample members who received only text
message reminders was not statistically significantly different from
the response rate of sample members who received only telephone
reminders. Therefore, both data collection groups will receive text
message reminders. Hard copy mailings, e-mails, and text messages
will be used to maintain ongoing contact with sample members in both
data collection groups, prior to and throughout data collection.
CATI
calling.
Sample members in the default group will not receive any CATI
calling if they are in the Experiment 2 experimental group. Sample
members in the default group will receive light outbound CATI
calling at 8 weeks (during the Production phase) if they are in the
Experiment 2 control group. Light CATI involves a minimal number of
phone calls, used mainly to prompt web response (as opposed to
regular CATI efforts that involve more frequent phone efforts, with
the goal to locate sample members and encourage their
participation). All cases in the aggressive group will receive
earlier (during the Early Completion phase) and more intense
telephone prompting than eligible cases in the default group.
Incentive
boosts. Researchers
have used incentive boosts as a nonresponse conversion strategy for
sample members who have implicitly or explicitly refused to complete
the survey (e.g., Groves and Heeringa 2006; Singer and Ye 2013).
These boosts are especially common in large federal surveys during
their nonresponse follow-up phase (e.g., The Center for Disease
Control and Prevention's National Survey of Family Growth) and have
been implemented successfully in other postsecondary education
surveys (e.g., HSLS:09 second follow-up; BPS:12/17; NPSAS:20). In
NPSAS:20, a $10 incentive boost increased the overall response rate
by about 3.2 percentage points above the projected response rate.
Therefore, a $10 incentive boost increase to the BPS:20/25 baseline
incentive is planned during the Nonresponse Follow-Up phase for all
remaining nonrespondents in the default data collection protocol.
Remaining nonrespondents in the aggressive data collection protocol
will be offered a $20 incentive boost increase to the baseline
incentive. This is because the $10 incentive boost in NPSAS:20 did
not show any effect on this group. If necessary, incentive boosts
may be targeted only at certain groups of nonrespondents to achieve
response goals (e.g., targeting nonrespondents from certain states
to ensure representativeness, targeting aggressive group
nonrespondents to reduce the potential for nonresponse bias).
Other interventions. While all BPS studies are conducted by NCES, the data collection contractor, RTI International, has typically used the study-specific e-mail “@rti.org” to contact and support sample members. Changing the e-mail sender to the NCES project officer or the RTI project director may increase the perceived importance of the survey and help personalize the contact materials, thereby potentially increasing relevance. Switching the sender during data collection also increases the chance that the survey invitation is delivered to the sample member rather than to a spam filter.
BPS:20/25 is being conducted by NCES. The following statisticians at NCES are responsible for the statistical aspects of the study: Dr. David Richards, Dr. Tracy Hunt-White, Dr. Sean Simone, Dr. Elise Christopher, and Dr. Gail Mulligan. NCES's prime contractor for BPS:20/25 is RTI International (Contract# 919900-18-C-0039), and subcontractors include Activate Research; EurekaFacts; HR Directions; Leonard Resource Group; Research Support Services; and Strategic Communications, Inc. The following staff members at RTI are working on the statistical aspects of the study design: Dr. Joshua Pretlow, Dr. Jennifer Wine, Dr. Nestor Ramirez, Mr. Darryl Cooney, Mr. Michael Bryan, Dr. T. Austin Lacy, Dr. Emilia Peytcheva, and Mr. Peter Siegel. Principal professional RTI staff not listed above, who are assigned to the study include: Ms. Ashley Wilson, Ms. Kristin Dudley, Mr. Jeff Franklin, Ms. Chris Rasmussen, and Ms. Donna Anderson.
References
Bryan,
M., Cooney, D., Elliot, B., and Richards, D. (2019). 2012/17
Beginning Postsecondary Students Longitudinal Study (BPS:12/17):
Data File Documentation (NCES 2020-522). U.S. Department of
Education. Washington, DC: National
Center for Education Statistics. Retrieved from
https://nces.ed.gov/pubsearch
Callegaro, M. Ayhan, O., Gabler, S., Haeder, S., & Villar, A. (2011). Combining landline and mobile phone samples. A dual frame approach. Working Papers 2011/13. Gesis Leibniz-Institut fuer Sozialwissenschaften.
Griggs,
A., Powell, R., Keeney, J., Waggy, M., Harris, K., Halpern, C. and
Dean, S. (2019) Research Note: A Prenotice Greeting Card's Impact on
Response Rates and Response Time, Longitudinal and Life Course
Studies, 10(4): 421-432.
Groves R.M., and Heeringa, S.G.
(2006). Responsive Design for Household Surveys: Tools for Actively
Controlling Survey Errors and Costs. Journal of the Royal
Statistical Society Series A-Statistics in Society, 169(3): 439-457.
Schober,
M.F,. Conrad, F.G., Antoun, C., Ehlen, P., Fail, S., Hupp, A.L.,
Johnston, M., Vickers, L., Yan, Y.H., and Zhang, C. (2015).
Precision and Disclosure in Text and Voice Interviews on
Smartphones. PLoS ONE 10(6): e0128337.
doi:10.1371/journal.pone.0128337
Singer, E. and Ye, C.
(2013). The Use and Effects of Incentives in Surveys. Annals. Annals
of the American Academy of Political and Social Science, 645(1):
112-141.
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
Author | Wilson, Ashley |
File Modified | 0000-00-00 |
File Created | 2023-10-22 |