High Frequency Surveys HTOPS June and August Supporting Statement B (1)

High Frequency Surveys HTOPS June and August Supporting Statement B (1).docx

High Frequency Surveys Program Household Trends and Outlook Pulse Survey (June, August)

OMB: 0607-1029

Document [docx]
Download: docx | pdf

OMB Information Collection Request

Supporting Statement B

U.S. Department of Commerce

U.S. Census Bureau


High Frequency Surveys Program

Household Trends and Outlook Pulse Survey

OMB Control Number 0607-1029


B. Collections of Information Employing Statistical Methods

  1. Describe (including a numerical estimate) the potential respondent universe and any sampling or other respondent selection method to be used. Data on the number of entities (e.g., establishments, State and local government units, households, or persons) in the universe covered by the collection and in the corresponding sample are to be provided in tabular form for the universe as a whole and for each of the strata in the proposed sample. Indicate expected response rates for the collection as a whole. If the collection had been conducted previously, include the actual response rate achieved during the last collection.

In February 2025, we conducted a sample replenishment operation. Approximately 110,000 additional households were invited to join the HTOPS panel. Of those households, 11,296 households responded to the baseline questionnaire, increasing the topical monthly sample size from 17,917 to approximately 29,213 housing units. Previous topicals yielded, on average, a response rate of approximately 55%. Subsequent topicals are expected to have a similar response rate, resulting in approximately 16,000 households responding to the June and August surveys.


  1. Describe the procedures for the collection of information including:

    • Statistical methodology for stratification and sample selection,

    • Estimation procedure,

    • Degree of accuracy needed for the purpose described in the justification,

    • Unusual problems requiring specialized sampling procedures, and

    • Any use of periodic (less frequent than annual) data collection cycles to reduce burden.

The sample design is a stratified systematic sample of all eligible HUs from the Census Bureau’s Master Address File (MAF), which covers all 50 states and the District of Columbia. Auxiliary data from the Demographic Frame (DF) and Planning Data Base (PDB) will be linked to the MAF to stratify the housing units into stratum based on demographic variables within the nine Census Bureau divisions. MAF records not stratified into a stratum based on the DF or PDB will be defined as their own strata. The sample will be distributed proportionately within divisions of the country to each stratum based on the number of housing units within the stratum. We will conduct a subsampling operation in stratum that, based on results of other demographic surveys, have higher response rates. Thus, the stratum where no subsampling occurs will be oversampled.

Future refreshment samples will be drawn from a frame that uses updated MAF, DF and PDB information, and those samples may be targeted at the geographic or domain level, to maintain representativeness of the Household Panel Survey, adjust sample sizes based on observed nonresponse, and account for sample units that are rotating out of the panel.

The final HTOPS survey weights are designed to produce estimates for the total persons aged 18 and older living within HUs (based on the person weight); and occupied household level estimates (based on the household weight). We will create these weights by adjusting the household-level sampling base weights by various factors to account for nonresponse, adults per household, and coverage. The final HTOPS survey weights are created by applying a Housing Unit adjustment, which converts the person level weight back into a housing unit (HU) weight by dividing the person level weight by the number of persons age 18 and older that were reported to live within the household, and the Occupied HU ratio adjustment, which ensures that the final Household Panel Survey weights will sum to the American Community Survey (ACS) one-year, state-level estimates of occupied HUs.

  1. Describe methods to maximize response rates and to deal with issues of non-response. The accuracy and reliability of information collected must be shown to be adequate for intended uses. For collections based on sampling, a special justification must be provided for any collection that will not yield "reliable" data that can be generalized to the universe studied.

Enrolled HTOPS participants are invited to respond to monthly topical surveys. Invitations will be sent by email, text message (opt-in), and for those panelists with no email or mobile phone contact information, outbound telephone calling. Using a unique login, panelists can access a topical questionnaire by computer, tablet, or smartphone. Phone-only panelists will complete topical surveys via inbound or outbound CATI.

Data collection for each topical survey will take place in a 2-week window. Each topical survey will be approximately 20 minutes long and panelists will receive up to two reminders to complete a topical survey. Panelists who complete a topical survey will be mailed a thank you letter with a $5 incentive (either digital or cash) about 2 to 6 weeks after the topical survey field period closes (depending on the type of incentive).

Future topical surveys can be sponsored by other Census Bureau survey programs. Each topical survey will offer panelists an opportunity to update contact information and verify their address for incentive mailing. Content governance will initially follow policies developed for the Household Pulse Survey and be amended as necessary.


Keeping panelists engaged will prevent attrition and maintain the representativeness of the panel. We will continue sending panelists one topical survey per month to keep them engaged. Panelists will not be eligible for more than one survey per month to keep burden low and reduce panel conditioning. Topical surveys may target specific groups of panelists depending on the topical survey sponsor. If panelists are not sampled for a particular month’s topical survey, they will be asked to respond to a pre-designed panel maintenance questionnaire that will also serve to verify demographic information and record any changes.



HTOPS Replacement and Replenishment

HTOPS panel members will be asked to complete approximately one questionnaire per month and will receive an incentive for each questionnaire. Panelists will be enrolled for three years and drop off after that period. In addition to this three-year limit, we expect attrition due to inactivity and requests to disenroll. Attrition can bias the panel estimates, making the development of a panel member replenishment plan of vital importance (Herzing & Blom, 2019; Lugtig et al., 2014; Schifeling et al., 2015; Toepoela & Schonlau, 2017).

Panelist requests to disenroll from the panel will be identified and processed according to forthcoming protocols. Periodic nonresponse or refusal to the monthly requests for otherwise active panelists is expected.

A particular questionnaire may be classified as “no response” due to unit nonresponse (i.e., no questionnaire initiation), item nonresponse resulting in an interview that is not usable for analyses (e.g., item nonresponse to questions deemed critical for analysis, high item nonresponse alone or after data review), and poor-quality data resulting in an unusable interview. Inactive panelists will remain members of the HTOPS panel if reengagement is desired by Census staff, especially for rare or historically undercounted populations. Definition of poor-quality responses is forthcoming.

We will assess on an ongoing basis (and no less than quarterly) the generalizability of the panel estimates to represent the target population. Evaluative methods will include precision within important demographic and geographic characteristics, R-indicators, propensity scores, and nonresponse bias analyses (Bianchi & Biffignandi, 2017; Eckman et al., 2021; Groves & Peytcheva, 2008; Peytcheva & Groves, 2009; Rosen et al., 2014).


Based on results from multiple analyses, we will identify any subgroups requiring replenishment. New members will be sampled and recruited using the same protocol as for initial enrollment.

Because incentives remain one of the most effective ways to encourage survey participation. The current incentive design includes the following:


  • Initial Invitation: $2 visible prepaid incentive with the initial invitation to complete the screener.

  • Baseline Questionnaire: $10 baseline contingent incentive after initial recruitment field period.

  • Topical Surveys: $5 for each topical survey (~20-minute average; once per month).

Respondents will be emailed digital incentives or mailed cash incentives (if unable or unwilling to accept digital incentives) for survey completion. The National Processing Center (NPC) and the Associate Director for Demographic Programs – Survey Operations (ADDP-SO) team will coordinate incentive distribution. The incentive structure could be amended to facilitate ongoing engagement of panelists, particularly for groups of panelists that are rare or historically undercounted.


  1. Describe any tests of procedures or methods to be undertaken. Testing is encouraged as an effective means of refining collections of information to minimize burden and improve utility. Tests must be approved if they call for answers to identical questions from 10 or more respondents. A proposed test or set of tests may be submitted for approval separately or in combination with the main collection of information.


Yes/No Radio Button and Select-all Check Box A/B Test



In Internet Self-Response, the way that respondents treat ‘forced choice’ radio buttons versus 'Select all that apply’ check boxes remains an open question. Research shows that a set of forced choice yes/no questions elicit more affirmative responses than a select all that apply question. Current research on these response formats generally focus on experiences and opinions. Fewer, if any, ask about factual questions typical in household surveys. Further, while some respondents treat forced choice questions like a select all that apply by ignoring the ‘no’ column, it does provide them the opportunity to explicitly say ‘no’, unlike the select all that apply format. Eliciting a ‘no’ provides more information beyond a ‘skipped/don’t know/refuse’ nonresponse. The comparison of item non-response for the two types of questions has not been explored in-depth. In this test conducted during the June data collection, respondents will receive questions about the same set of programs and who in the household participates in the programs, however they will be randomly assigned to either receive a select all that apply or forced choice format of some applicable questions. This will allow us to compare the prevalence of ‘Yes’ responses as well as the rate of item non-response in either setting.

  1. Provide the name and telephone number of individuals consulted on statistical aspects of the design and the name of the agency unit, contractor(s), grantee(s), or other person(s) who will actually collect and/or analyze the information for the agency.


Statistical Design:

Anthony Tersine

Demographic Statistical Methods Division

Demographic Programs Directorate

Anthony.g.tersine.jr@census.gov

Data Collection/Survey Design:

Jason Fields

Social Economic and Housing Statistics Division

Demographic Programs Directorate

jason.m.fields@census.gov


Jennifer Hunter Childs

Demographic Programs Directorate

jennifer.hunter.childs@census.gov

Statistical Analysis:


David Waddington

Social Economic and Housing Statistics Division

Demographic Programs Directorate

david.g.waddington@census.gov


Page | 3


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleHigh Frequency Surveys HTOPS June and August Supporting Statement B
AuthorMary Reuling Lenaiyasa (CENSUS/PCO FED)
File Modified0000-00-00
File Created2025-05-19

© 2025 OMB.report | Privacy Policy