Proposed Information Collection; Comment Request; 2015 National Content Test, 71377-71381 [2014-28247]
Download as PDF
Federal Register / Vol. 79, No. 231 / Tuesday, December 2, 2014 / Notices
the Commission’s Web site,
www.usccr.gov, or to contact the
Western Regional Office at the above
email or street address.
The meeting will be conducted
pursuant to the provisions of the rules
and regulations of the Commission and
FACA.
Dated in November 25, 2014.
David Mussatt,
Chief, Regional Programs Coordination Unit.
[FR Doc. 2014–28276 Filed 12–1–14; 8:45 am]
BILLING CODE 6335–01–P
DEPARTMENT OF COMMERCE
Census Bureau
Proposed Information Collection;
Comment Request; 2015 National
Content Test
U.S. Census Bureau,
Commerce.
ACTION: Notice.
AGENCY:
The Department of
Commerce, as part of its continuing
effort to reduce paperwork and
respondent burden, invites the general
public and other Federal agencies to
take this opportunity to comment on
proposed and/or continuing information
collections, as required by the
Paperwork Reduction Act of 1995,
Public Law 104–13 (44 U.S.C.
3506(c)(2)(A)).
SUMMARY:
To ensure consideration, written
comments must be submitted on or
before February 2, 2015.
ADDRESSES: Direct all written comments
to Jennifer Jessup, Departmental
Paperwork Clearance Officer,
Department of Commerce, Room 6616,
14th and Constitution Avenue NW.,
Washington, DC 20230 (or via the
Internet at jjessup@doc.gov).
FOR FURTHER INFORMATION CONTACT:
Requests for additional information or
copies of the information collection
instrument(s) and instructions should
be directed to Erin Love, Census Bureau,
HQ–3H154E, Washington, DC 20233;
(301) 763–2034 (or via email at
erin.s.love@census.gov).
SUPPLEMENTARY INFORMATION:
DATES:
rljohnson on DSK3VPTVN1PROD with NOTICES
I. Abstract
The 2015 National Content Test (NCT)
is part of the research and development
cycle leading up to the re-engineered
2020 Census. The 2015 NCT will help
the Census Bureau achieve one of its
Strategic Goals—developing a census
that is cost-effective, improves coverage,
and reduces operational risk.
VerDate Sep<11>2014
15:30 Dec 01, 2014
Jkt 235001
The first objective of this test is to
evaluate and compare different census
content, including race and Hispanic
origin, relationship, and withinhousehold coverage. This will be the
primary mid-decade opportunity to
compare different content strategies
prior to making final decisions about the
content in the 2020 Census. The test
will include a reinterview to further
assess the accuracy and reliability of the
question alternatives for race, origin,
and within-household coverage.
The second objective is to test
different contact strategies for
optimizing self-response. This includes
nine different approaches to
encouraging households to respond and,
specifically, to respond using the less
costly and more efficient Internet
response option. These approaches
include altering the timing of the first
reminder, use of email as a reminder,
altering the timing for sending the mail
questionnaire, use of a third reminder,
and sending a letter in place of a paper
questionnaire to non-respondents.
The third objective is to test different
options for offering non-English
materials. The goal is to provide
language support for respondents with
limited English proficiency. Options
being explored include online Spanish
questionnaires, dual-language English
and Spanish paper questionnaires and
letters, and additional questionnaire
options and support in non-English
languages.
Regarding the first objective, the
classification of racial and ethnic
responses to the decennial census by the
Census Bureau adheres to the U.S.
Office of Management and Budget’s
(OMB) October 30, 1997 ‘‘Revisions to
the Standards for the Classification of
Federal Data on Race and Ethnicity’’
(see www.whitehouse.gov/omb/fedreg_
1997standards). There are five
minimum categories for data on race:
‘‘White,’’ ‘‘Black or African American,’’
‘‘American Indian or Alaska Native,’’
‘‘Asian,’’ and ‘‘Native Hawaiian or Other
Pacific Islander.’’ There are two
minimum categories for data on
ethnicity: ‘‘Hispanic or Latino’’ and
‘‘Not Hispanic or Latino.’’ The OMB
standards advise that respondents shall
be offered the option of selecting one or
more racial designations. The OMB
standards also advise that race and
ethnicity are two distinct concepts;
therefore, Hispanics or Latinos may be
any race.
The minimum categories for data on
race and ethnicity for Federal statistics,
program administrative reporting, and
civil rights compliance reporting are
defined by OMB as follows:
PO 00000
Frm 00002
Fmt 4703
Sfmt 4703
71377
• American Indian or Alaska Native—
A person having origins in any of the
original peoples of North and South
America (including Central America),
and who maintains tribal affiliation or
community attachment.
• Asian—A person having origins in
any of the original peoples of the Far
East, Southeast Asia, or the Indian
subcontinent including, for example,
Cambodia, China, India, Japan, Korea,
Malaysia, Pakistan, the Philippine
Islands, Thailand, and Vietnam.
• Black or African American—A
person having origins in any of the
black racial groups of Africa. Terms
such as ‘‘Haitian’’ or ‘‘Negro’’ can be
used in addition to ‘‘Black or African
American.’’
• Hispanic or Latino—A person of
Mexican, Puerto Rican, Cuban, South or
Central American, or other Spanish
culture or origin, regardless of race. The
term, ‘‘Spanish origin,’’ can be used in
addition to ‘‘Hispanic or Latino.’’
• Native Hawaiian or Other Pacific
Islander—A person having origins in
any of the original peoples of Hawaii,
Guam, Samoa, or other Pacific Islands.
• White—A person having origins in
any of the original peoples of Europe,
the Middle East, or North Africa.
The 1997 OMB standards state the
minimum categories that must be used
to collect and present federal data on
race and ethnicity. Additionally, the
1997 OMB standards permit the
collection of more detailed information
on population groups, provided that any
additional groups can be aggregated into
the minimum standard set of categories.
Currently, the Census Bureau collects
additional detailed information on
Hispanic or Latino groups, American
Indian and Alaska Native tribes, Asian
groups, and Native Hawaiian and Other
Pacific Islander groups.
For example, responses to the race
question such as ‘‘Navajo Nation,’’
‘‘Doyon,’’ and ‘‘Mayan’’ are collected
and tabulated in Census Bureau
censuses and surveys, and can be
aggregated into the total American
Indian or Alaska Native population.
Detailed responses to the race question
such as ‘‘Chinese,’’ ‘‘Asian Indian,’’ and
‘‘Vietnamese’’ are collected and
tabulated, and can be aggregated into
the total Asian population. Responses to
the ethnicity question such as
‘‘Mexican,’’ ‘‘Puerto Rican,’’ and
‘‘Cuban’’ are collected and tabulated in
Census Bureau censuses and surveys,
and can be aggregated into the total
Hispanic or Latino population.
Responses to the race question such as
‘‘Native Hawaiian,’’ ‘‘Chamorro,’’ or
‘‘Fijian’’ are collected and tabulated,
and can be aggregated into the total
E:\FR\FM\02DEN1.SGM
02DEN1
rljohnson on DSK3VPTVN1PROD with NOTICES
71378
Federal Register / Vol. 79, No. 231 / Tuesday, December 2, 2014 / Notices
Native Hawaiian or Other Pacific
Islander population.
The 2015 NCT will test ways to
collect and tabulate detailed
information for all groups, including
data for White groups, such as German,
Irish, and Lebanese, and data for Black
groups, such as African American,
Jamaican, and Nigerian, which have not
been tabulated previously from the
question on race. Responses to the race
question such as ‘‘African American,’’
‘‘Jamaican,’’ or ‘‘Nigerian’’ will be
collected and tabulated, and can be
aggregated to the total Black or African
American population. Responses to the
race question such as ‘‘German,’’
‘‘Irish,’’ or ‘‘Lebanese’’ will be collected
and tabulated, and can be aggregated
into the total White population.
The 2015 NCT will also test a separate
‘‘Middle Eastern or North African’’
category and the collection of detailed
groups such as ‘‘Lebanese,’’ ‘‘Egyptian,’’
and ‘‘Iranian.’’ Following the current
OMB standards, Middle Eastern and
North African responses are classified as
‘‘White.’’
The results of the 2015 NCT will
guide future collection and tabulation of
detailed information for all race and
ethnicity groups.
Plans for the 2020 Census call for the
use of less costly and more efficient
web-based response options to collect
information, as opposed to a previous
predominant reliance on paper-based
questionnaires. One benefit of the
online response mode is that it allows
for more functionality and greater
flexibility in designing questions
compared to paper, which is
constrained by space availability. With
the advantage of new technology, the
2015 National Content Test will utilize
web-based technology, such as internet,
smart phone, tablet, and telephone to
improve question designs and optimize
reporting of detailed racial and ethnic
groups (e.g., Samoan, Iranian, Blackfeet
Tribe, Filipino, Jamaican, Puerto Rican,
Irish, etc.).
The web-based designs provide much
more utility and flexibility for using
detailed checkboxes and write-in spaces
to elicit and collect data for detailed
groups than traditional paper
questionnaires, and will help collect
data for both the broader OMB
categories, as well as detailed responses
across all groups.
Components of the Test
A. Race and Origin Content
The Census Bureau conducted an
extensive research undertaking as part
of the 2010 Census—the 2010 Census
Race and Hispanic Origin Alternative
VerDate Sep<11>2014
15:30 Dec 01, 2014
Jkt 235001
Questionnaire Experiment (AQE) (for
details, see www.census.gov/
2010census/news/press-kits/aqe/
aqe.html). The 2010 AQE examined
alternative strategies for improving the
collection of data on a race and
Hispanic origin, with four goals in
mind:
1. Increasing reporting in the standard
race and ethnic categories as defined by
the U.S. Office of Management and
Budget;
2. Decreasing item non-response for
these questions;
3. Increasing the accuracy and
reliability of the results for this
question; and
4. Eliciting detailed responses for all
racial and ethnic communities (e.g.,
Chinese, Mexican, Jamaican, etc.).
The results of the AQE supported all
of these objectives. Additionally, many
individuals across communities liked
the combined question approach. They
believed it presented equity to the
different categories. Some of the
findings from this research include:
• Combining race and ethnicity into
one question did not change the
proportion of people who reported as
Hispanics, Blacks, Asians, American
Indians and Alaska Natives, or Native
Hawaiians and Other Pacific Islanders.
• The combined question yielded
higher response rates.
• The combined question increased
reporting of detailed responses for most
groups, but decreased reporting for
others.
• The combined question better
reflected self-identity.
The successful strategies from the
AQE research have been employed in
the design of the Census Bureau’s middecade research. Four key dimensions
of the questions on race and Hispanic
origin are being tested in the 2015 NCT.
These include question format, response
categories, wording of the instructions,
and question terminology.
Question Format
The 2015 NCT will evaluate the use
of two alternative question format
approaches for collecting data on race
and ethnicity. One approach uses two
separate questions: the first about
Hispanic origin and the second about
race (‘‘separate questions’’). The other
approach combines the two items into
one question about race and origin
(‘‘combined question’’). The 2015 middecade research will test the approaches
with new data collection methods,
including internet, telephone, and inperson response.
1. Separate race and origin questions:
This is a modified version of the race
and Hispanic origin format used in the
PO 00000
Frm 00003
Fmt 4703
Sfmt 4703
2010 Census. Updates since the 2010
Census include added write-in spaces
and examples for the ‘‘White’’ and
‘‘Black or African Am.’’ response
categories, removal of the term ‘‘Negro,’’
and an instruction to select one or more
boxes in the Hispanic origin question.
2. Combined question with
checkboxes and write-ins on same
screen: This is a modified version of the
combined question approaches found to
be successful in the 2010 AQE.
Checkboxes are provided for the U.S.
Office of Management and Budget
(OMB) standard categories (per the 1997
Standards for the Classification of
Federal Data on Race and Ethnicity)
with a corresponding write-in space for
each checkbox category. In this version,
all write-in spaces are visible at all
times. Each response category contains
six example origins, which represent the
diversity of the geographic definitions of
the OMB category. For instance, the
‘‘Asian’’ category examples of Chinese,
Filipino, Asian Indian, Vietnamese,
Korean, and Japanese represent the six
largest detailed Asian groups in the
United States, reflecting OMB’s
definition of Asian (‘‘A person having
origins in any of the original peoples of
the Far East, Southeast Asia, and the
Indian subcontinent.’’). Respondents do
not have to select an OMB checkbox,
but may enter a detailed response in the
write-in space without checking a
category.
a. Combined question with
checkboxes and write-ins on separate
screens (Internet-only): In this version,
the detailed origin groups are solicited
on subsequent screens after the OMB
response categories have been selected.
On the first screen, the OMB checkbox
categories are shown along with their
six representative example groups. Once
the OMB categories have been selected,
one at a time, subsequent screens solicit
further detail for each category that was
chosen (e.g., Asian), using a write-in
space to collect the detailed groups (e.g.,
Korean and Japanese). The intent is to
separate mouse click tasks (checkbox
categories) and typing tasks (write-ins)
in an attempt to elicit responses that are
more detailed. The same version was
used as one of three race and origin
Internet panels in the 2014 Census Test.
3. Combined question branching with
detailed checkbox screens (Internetonly): This version is an alternative
method of soliciting detailed origin
groups using separate screens, detailed
checkboxes, and write-in spaces. On the
first screen, the OMB checkbox
categories are shown along with their
six representative example groups. Once
the OMB categories have been selected,
one at a time, subsequent screens solicit
E:\FR\FM\02DEN1.SGM
02DEN1
Federal Register / Vol. 79, No. 231 / Tuesday, December 2, 2014 / Notices
that—mark a box—overlooking the
remainder of the instruction. The new
instruction (‘‘Mark all boxes that
apply’’) is an attempt to improve the
clarity of the question and make it more
apparent that more than one group may
be selected.
Race Response Categories
The 2015 National Content Test will
evaluate the use of the Middle Eastern
or North African (MENA) category in
the race question. There will be two
treatments for testing this dimension:
1. Use of MENA category: This
treatment tests the addition of a MENA
checkbox category to the race question.
The MENA category is placed within
the current category lineup, based on
estimates of population size, between
the categories for Native Hawaiians and
Other Pacific Islanders and ‘‘Some other
race.’’ With the addition of this new
category, the ‘‘White’’ example groups
are revised. The Middle Eastern and
North African examples of ‘‘Lebanese’’
and ‘‘Egyptian’’ are replaced with the
European examples of ‘‘Polish’’ and of
‘‘French.’’ The MENA checkbox
category will have the examples of
‘‘Lebanese, Iranian, Egyptian, Syrian,
Moroccan, Algerian, etc.’’ All other
checkbox categories and write-in spaces
remain the same.
2. No separate MENA category: This
treatment tests approaches without a
separate MENA checkbox category, and
represents the current OMB definition
of White (‘‘A person having origins in
any of the original peoples of Europe,
the Middle East, or North Africa.’’). Here
we will provide examples of Middle
Eastern and North African origins
(‘‘Lebanese’’ and ‘‘Egyptian’’) with
European origin groups as part of the
‘‘White’’ racial category.
rljohnson on DSK3VPTVN1PROD with NOTICES
further detail for each category, this
time using a series of additional
checkboxes for the six largest detailed
groups (e.g., Chinese, Filipino, Asian,
Indian, Vietnamese, Korean, and
Japanese) with a write-in space also
provided to collect additional groups.
Wording of the Instructions
1. ‘‘Mark [X] one or more boxes’’: The
current paper version of the instructions
on paper states, ‘‘Mark [X] one or more
boxes AND print your specific
origin(s).’’
2. ‘‘Mark all that apply/You may mark
multiple groups’’: In this version, the
instruction is modified to ‘‘Mark all
boxes that apply AND print the specific
[origin(s)/ethnicities] in the spaces
below. Note, you may report more than
one group.’’ Recent qualitative focus
groups and cognitive research (e.g., 2010
AQE research; 2013 Census Test
research) found that respondents
frequently overlook the instruction to
‘‘Mark’’ [X] one or more boxes. The
research found that some respondents
may have stopped reading the
instruction after noticing the visual cue
[X] and proceeded directly to do just
Question Terms
1. ‘‘Origin’’ term: The current version
of the race and Hispanic origin
questions use the terms ‘‘race’’ and/or
‘‘origin’’ to describe the concepts and
groups in the question stem,
instructions, and examples. For
instance, in the combined race and
Hispanic origin approach, the question
stem is ‘‘What is your race or origin?’’
In addition, prior to each write-in field,
respondents are instructed to ‘‘Print
specific origin(s), for example . . .’’
2–3. Alternative terms: Recent
qualitative focus groups and qualitative
research (e.g., 2010 AQE research; 2013
Census Test research; cognitive pretesting for 2016 American Community
Survey Content Test) found that the
term ‘‘origin’’ is confusing or misleading
to many respondents, who may think it
is asking about where they immigrated
from or where they were born. Two
alternative options are being explored in
cognitive testing and usability research.
One approach tests the use of the term
‘‘ethnicities’’ along with ‘‘race’’ (e.g.,
‘‘Print the specific races(s) and/or
ethnicities . . .’’). The other approach
tests the removal of the terms altogether
from the question stem, instructions,
and examples. Instead, a general
approach asks, ‘‘Which categories
describe this person?’’ The exact
terminology to be used for the
alternative version is pending cognitive
testing and usability results later this
year, which will inform the wording to
be used in the 2015 NCT.
B. Relationship Content
Two versions of the relationship
question will be tested. Both versions
are the same as those used in a splitsample in the 2014 Census Test, with no
changes. The new relationship
categories have also been tested in other
Census Bureau surveys including the
American Housing Survey, American
Community Survey, and the Survey of
Income and Program Participation
(currently used in production).
Although research to date has been
informative, leading to the development
of the revised relationship question,
additional quantitative testing is
needed. Since the incidence of some
household relationships—such as samesex couples—is relatively low in the
general population, the revised question
needs to be tested with large,
VerDate Sep<11>2014
15:30 Dec 01, 2014
Jkt 235001
PO 00000
Frm 00004
Fmt 4703
Sfmt 4703
71379
representative samples prior to
routinely including them in the 2020
Census questionnaire.
The first version uses the 2010 Census
relationship question response options,
but in a new order, starting with
‘‘husband or wife’’ and then the
‘‘unmarried partner’’ category. This
version also re-introduces the foster
child category, which was removed
from the 2010 Census form due to space
issues.
The second version includes the same
basic response options as the 2010
Census version, but modifies/expands
the ‘‘husband or wife’’ and ‘‘unmarried
partner’’ categories to distinguish
between same-sex and opposite-sex
relationships.
C. Coverage Content (Internet Only)
The 2012 National Census Test
experimented with several methods to
improve accurate within-household
coverage for Internet respondents. One
benefit of the online response mode is
that it allows for more functionality and
greater flexibility in designing questions
compared to paper, which is
constrained by space availability. The
2012 test included a coverage follow-up
reinterview to evaluate the different
Internet design options, but some results
were inconclusive. In the 2015 NCT,
two designs will be tested to compare
different approaches for helping
respondents provide a more accurate
roster of household residents.
The first approach is the ‘‘RulesBased’’ approach, and will allow us to
see whether the presence of a question
asking the number of people in the
household along with the residence rule
instructions helps respondents create an
accurate roster. This is similar to the
approach used across all modes in
Census 2000 and the 2010 Census,
where the respondent was expected to
understand our residence rules and
apply them to their household. This is
followed by a household-level question
that probes to determine if any
additional people not listed originally
should be included for consideration as
residents of the household (several
types of people and living situations are
shown in a bulleted list).
The ‘‘Question-Based’’ approach
allows us to ask guided questions to
help improve resident information.
Respondents are not shown the
residence rule instructions and are only
asked to create an initial roster of people
they consider to be living or staying at
their address on Census Day. This is
followed by several short householdlevel questions about types of people
and living situations that might apply to
E:\FR\FM\02DEN1.SGM
02DEN1
71380
Federal Register / Vol. 79, No. 231 / Tuesday, December 2, 2014 / Notices
rljohnson on DSK3VPTVN1PROD with NOTICES
someone in the household that was not
listed originally.
D. Optimizing Self Response
The nine proposed contact strategies
for optimizing self response (OSR) are
summarized as follows:
Internet Push (Control): This is the
standard Internet Push strategy used in
the most recent series of self response
tests, including the 2014 Census Test.
This panel will serve as a control panel
against which to compare the
experimental strategies. There will be
nine treatments as part of the OSR test.
Internet Push With Early Postcard:
The motivation for this panel is to study
the timing of reminders. The hypothesis
is that sending the first reminder sooner
(closer to the initial Internet push)
would provide for a better connection
between the two mailings, and could
increase response. A side benefit is that
this could also reduce the volume of
later targeted mailings since responses
may be quicker overall.
The motivation for the following
sequence of three panels is based on
recent American Community Survey
(ACS) research, which has found
depressed self response rates among
certain respondents/areas with lower
Internet usage. Testing the delivery of
the paper questionnaires at various
points in the response process will
allow us to have complete response
measures under several scenarios for the
cost/benefit analysis needed to inform
2020 Census planning. Although these
strategies may not make sense for
everyone in 2020, using a responsive
design and tailoring the contact strategy
for certain geographic areas or
populations may be beneficial.
• Internet Push With Early
Questionnaire: questionnaire sent at
third mailing, one week sooner
• Internet Push With Even Earlier
Questionnaire: questionnaire sent at
second mailing, two weeks sooner
• Internet Choice: questionnaire sent
at first mailing, providing a choice of
Internet or paper from the beginning
Internet Push With Postcard as Third
Reminder: The motivation for this panel
is to further encourage self response,
after the questionnaire mailing, prior to
nonresponse follow-up. Numerous
survey research studies have concluded
that, while there is a point of
diminishing returns, further reminders
will inevitably increase self response
rates.
Internet Push Postcard: The
motivation for this panel is to study the
impact of sending a postcard at the first
mailing instead of a letter. There are two
potential benefits. First is the possible
cost savings of printing and mailing a
VerDate Sep<11>2014
15:30 Dec 01, 2014
Jkt 235001
postcard compared to the envelope
package (with letter and instruction
card). Second is the potential for
increased self response because reading
a postcard requires less effort by a
respondent. In this panel, we send a
letter at the third contact (sent to nonrespondents only), in place of a
postcard, to vary the types of contacts
received.
Internet Push With Early Postcard and
Second Letter Instead of Mail
Questionnaire: The motivation for
testing an approach in which we do not
send a mail questionnaire is to address
the high-level goal of greatly reducing
paper responses in the 2020 Census. By
testing an approach in which we send
an Internet push letter in place of a
paper questionnaire at the fourth
mailing, we will have a more robust set
of response measures for informing cost/
benefit analyses.
Internet Push With Postcard and
Email as 1st Reminder (same time): The
motivation for this panel is to determine
if we can take advantage of the email
addresses in the supplemental contact
frame maintained by the Center for
Administrative Records Research and
Applications. The hypothesis is that by
sending a postcard and email at the
same time, we may be able to elicit
increased response.
E. Language
In the two mailings that contain a
letter for each Optimizing Self response
strategy, three different methods will be
used to encourage response. In
particular, by altering the language
support provided in the letter, the goal
is to increase response for respondents
with limited English proficiency.
The control panel is similar to the
2014 Census Test design, in which the
mailing materials are in English with a
single Spanish sentence directing
respondents to the Web site or the
telephone assistance line.
One of the goals of language research
is to maximize the number of nonEnglish speakers that receive the same
message as English speakers prior to
going online to respond. Two panels
provide equality between the English
and Spanish content in the letter and
test whether one method is better at
eliciting Spanish responses. The swimlane design has been used in the past,
such as with the bilingual questionnaire
in the 2010 Census. The dual-sided
letter provides English content on one
side and Spanish content on the other
side. In addition, because research has
shown that Spanish-speaking
respondents do not always open the
mailings because they may not know
that language resource information is
PO 00000
Frm 00005
Fmt 4703
Sfmt 4703
provided inside, the outgoing envelope
for both panels will include the census
test Web site URL and a brief message
in both languages.
This test will also explore additional
options for non-English speakers to
complete the questionnaires.
F. Content Reinterview
A sub-sample of respondents from the
2015 NCT will be selected for a content
reinterview, focused on race and origin
and within-household coverage, with a
goal of assessing accuracy and reliability
of the different designs. Reinterviews
are conducted with a sub-sample of
respondents, by asking more detailed
questions on question topics, in order to
assess the accuracy of the responses.
II. Method of Collection
The initial mail-out is planned for late
August 2015. This contact will explain
why we are conducting the mandatory
2015 NCT, assure respondents that their
answers are confidential, and inform
them of the measures we take to keep
their personal information secure. The
second mail-out is considered a
reminder and is sent to all housing
units. All contacts after the second
mailing are sent to non-respondents
only.
Respondents are encouraged to
respond to the 2015 NCT by Internet but
may also be able to provide information
by phone. Many will also receive a
paper questionnaire at some point in the
mail-out strategy. The test will be
conducted nationally in all 50 U.S.
states, the District of Columbia, and
Puerto Rico.
III. Data
OMB Control Number: None.
Form Number: TBD.
Type of Review: Regular submission.
Affected Public: Individuals or
households.
Estimated Number of Respondents:
1.3 million households. (1.2 million
initial response + 100,000 reinterview).
Estimated Time per Response: 10
minutes.
Estimated Total Annual Burden
Hours: 216,667.
Estimated Total Annual Cost: There is
no cost to respondents except for their
time to respond.
Respondent’s Obligation: Mandatory.
Legal Authority: Title 13 U.S.C. 141
and 193.
IV. Request for Comments
Comments are invited on: (a) Whether
the proposed collection of information
is necessary for the proper performance
of the functions of the agency, including
whether the information shall have
E:\FR\FM\02DEN1.SGM
02DEN1
Federal Register / Vol. 79, No. 231 / Tuesday, December 2, 2014 / Notices
practical utility; (b) the accuracy of the
agency’s estimate of the burden
(including hours and cost) of the
proposed collection of information; (c)
ways to enhance the quality, utility, and
clarity of the information to be
collected; and (d) ways to minimize the
burden of the collection of information
on respondents, including through the
use of automated collection techniques
or other forms of information
technology.
Comments submitted in response to
this notice will be summarized and/or
included in the request for OMB
approval of this information collection;
they also will become a matter of public
record.
Dated: November 25, 2014.
Glenna Mickelson,
Management Analyst, Office of the Chief
Information Officer.
[FR Doc. 2014–28247 Filed 12–1–14; 8:45 am]
BILLING CODE 3510–07–P
DEPARTMENT OF COMMERCE
Foreign-Trade Zones Board
[B–85–2014]
rljohnson on DSK3VPTVN1PROD with NOTICES
Foreign-Trade Zone 116—Port Arthur,
Texas; Expansion of Subzone 116B;
Total Petrochemicals & Refining USA,
Inc.; Port Arthur and Jefferson County,
Texas
An application has been submitted to
the Foreign-Trade Zones (FTZ) Board by
the Foreign-Trade Zone of Southeast
Texas, Inc., grantee of FTZ 116,
requesting an expansion of Subzone
116B on behalf of Total Petrochemicals
& Refining USA, Inc. The application
was submitted pursuant to the
provisions of the Foreign-Trade Zones
Act, as amended (19 U.S.C. 81a–81u),
and the regulations of the FTZ Board (15
CFR part 400). It was formally docketed
on November 25, 2014.
Subzone 116B was approved on
September 18, 1995 (Board Order 772,
60 FR 49564, 9/26/95). The subzone
(1,457 acres) currently consists of four
sites located in Port Arthur and
Jefferson County: Site 1 (1,244 acres)—
main refinery complex located along the
Neches River at State Farm to Market
Highway 366 and 32nd St., Port Arthur;
Site 2 (19 acres)—West Port Arthur
Tank Farm located at Roosevelt and
53rd Streets, Port Arthur; Site 3 (194
acres)—refinery expansion site, located
adjacent to the refinery at State Farm to
Market Hwy 366, Port Arthur; and, Site
4—Sun Marine Terminal-Nederland
tank storage facility (leased storage)
located along the Neches River in
Nederland.
VerDate Sep<11>2014
15:30 Dec 01, 2014
Jkt 235001
The current request would add a
pipeline that originates from the
subzone’s leased storage facility at Site
4 to the main refinery located at Site 1,
as described in the application. No
additional authorization for production
activity has been requested at this time.
In accordance with the FTZ Board’s
regulations, Camille Evans of the FTZ
Staff is designated examiner to review
the application and make
recommendations to the FTZ Board.
Public comment is invited from
interested parties. Submissions shall be
addressed to the FTZ Board’s Executive
Secretary at the address below. The
closing period for their receipt is
January 12, 2015. Rebuttal comments in
response to material submitted during
the foregoing period may be submitted
during the subsequent 15-day period to
January 26, 2015.
A copy of the application will be
available for public inspection at the
Office of the Executive Secretary,
Foreign-Trade Zones Board, Room
21013, U.S. Department of Commerce,
1401 Constitution Avenue NW.,
Washington, DC 20230–0002, and in the
‘‘Reading Room’’ section of the FTZ
Board’s Web site, which is accessible
via www.trade.gov/ftz.
For further information, contact
Camille Evans at Camille.Evans@
trade.gov or (202) 482–2350.
Dated: November 25, 2014.
Andrew McGilvray,
Executive Secretary.
[FR Doc. 2014–28416 Filed 12–1–14; 8:45 am]
BILLING CODE 3510–DS–P
DEPARTMENT OF COMMERCE
International Trade Administration
Proposed Information Collection;
Comment Request; Domestic and
International Clients Export Services
and Customized Forms
International Trade
Administration, Commerce.
ACTION: Notice.
AGENCY:
The Department of
Commerce, as part of its continuing
effort to reduce paperwork and
respondent burden, invites the general
public and other Federal agencies to
take this opportunity to comment on
proposed and/or continuing information
collections, as required by the
Paperwork Reduction Act of 1995.
DATES: Written comments must be
submitted on or before February 2, 2015.
ADDRESSES: Direct all written comments
to Jennifer Jessup, Departmental
Paperwork Clearance Officer,
SUMMARY:
PO 00000
Frm 00006
Fmt 4703
Sfmt 4703
71381
Department of Commerce, Room 6616,
14th and Constitution Avenue NW.,
Washington, DC 20230 (or via the
Internet at JJessup@doc.gov).
FOR FURTHER INFORMATION CONTACT:
Requests for additional information or
copies of the information collection
instrument and instructions should be
directed to Joe Carter, Office of Strategic
Planning, 1999 Broadway, Suite 2205
Denver, CO 80220, (303) 844–5656,
joe.carter@trade.gov.
SUPPLEMENTARY INFORMATION:
I. Abstract
The International Trade
Administration’s Global Markets (GM) is
seeking approval to renew the currently
approved OMB control number: 0625–
0143. These collections include all
client intake, events/activities and
export success forms. This
comprehensive information collection
will cover all aspects of a U.S.
organization’s life-cycle with GM.
GM is mandated by Congress to help
U.S. organizations, particularly small
and medium-sized organizations, export
their products and services to global
markets. As part of its mission, GM
provides market entry/expansion
services and trade events to U.S.
organizations.
The Domestic and International
Clients Export Services and Customized
Forms are needed to collect information
to enable, but not limited to small and
medium sized, U.S. organizations to
efficiently and effectively enhance their
ability to determine which international
organizations are most suited for their
exporting expansion efforts.
The key to effectively and efficiently
assist U.S. organizations export is
identifying and verifying potential
international buyers of U.S. goods and
services. The categories of questions are:
Contact information, organization
information, organization type,
agreements and confirmations,
objectives, products and services,
exporting experience, marketing, events
and activities, trade fair/show, certified
trade missions, trade missions,
advocacy, environment, and education.
GM asks only those questions that
provide the required information to
assist GM in fulfilling a client’s
objective for a requested service and/or
event/activity.
As GM moves forward, we understand
the importance and need for strategic
planning and integration of future
technology and initiatives that relate to
GM programs and metrics with the
types of information collected from
clients to conduct those programs.
Additionally, the most important,
E:\FR\FM\02DEN1.SGM
02DEN1
Agencies
[Federal Register Volume 79, Number 231 (Tuesday, December 2, 2014)]
[Notices]
[Pages 71377-71381]
From the Federal Register Online via the Government Printing Office [www.gpo.gov]
[FR Doc No: 2014-28247]
=======================================================================
-----------------------------------------------------------------------
DEPARTMENT OF COMMERCE
Census Bureau
Proposed Information Collection; Comment Request; 2015 National
Content Test
AGENCY: U.S. Census Bureau, Commerce.
ACTION: Notice.
-----------------------------------------------------------------------
SUMMARY: The Department of Commerce, as part of its continuing effort
to reduce paperwork and respondent burden, invites the general public
and other Federal agencies to take this opportunity to comment on
proposed and/or continuing information collections, as required by the
Paperwork Reduction Act of 1995, Public Law 104-13 (44 U.S.C.
3506(c)(2)(A)).
DATES: To ensure consideration, written comments must be submitted on
or before February 2, 2015.
ADDRESSES: Direct all written comments to Jennifer Jessup, Departmental
Paperwork Clearance Officer, Department of Commerce, Room 6616, 14th
and Constitution Avenue NW., Washington, DC 20230 (or via the Internet
at jjessup@doc.gov).
FOR FURTHER INFORMATION CONTACT: Requests for additional information or
copies of the information collection instrument(s) and instructions
should be directed to Erin Love, Census Bureau, HQ-3H154E, Washington,
DC 20233; (301) 763-2034 (or via email at erin.s.love@census.gov).
SUPPLEMENTARY INFORMATION:
I. Abstract
The 2015 National Content Test (NCT) is part of the research and
development cycle leading up to the re-engineered 2020 Census. The 2015
NCT will help the Census Bureau achieve one of its Strategic Goals--
developing a census that is cost-effective, improves coverage, and
reduces operational risk.
The first objective of this test is to evaluate and compare
different census content, including race and Hispanic origin,
relationship, and within-household coverage. This will be the primary
mid-decade opportunity to compare different content strategies prior to
making final decisions about the content in the 2020 Census. The test
will include a reinterview to further assess the accuracy and
reliability of the question alternatives for race, origin, and within-
household coverage.
The second objective is to test different contact strategies for
optimizing self-response. This includes nine different approaches to
encouraging households to respond and, specifically, to respond using
the less costly and more efficient Internet response option. These
approaches include altering the timing of the first reminder, use of
email as a reminder, altering the timing for sending the mail
questionnaire, use of a third reminder, and sending a letter in place
of a paper questionnaire to non-respondents.
The third objective is to test different options for offering non-
English materials. The goal is to provide language support for
respondents with limited English proficiency. Options being explored
include online Spanish questionnaires, dual-language English and
Spanish paper questionnaires and letters, and additional questionnaire
options and support in non-English languages.
Regarding the first objective, the classification of racial and
ethnic responses to the decennial census by the Census Bureau adheres
to the U.S. Office of Management and Budget's (OMB) October 30, 1997
``Revisions to the Standards for the Classification of Federal Data on
Race and Ethnicity'' (see www.whitehouse.gov/omb/fedreg_1997standards).
There are five minimum categories for data on race: ``White,'' ``Black
or African American,'' ``American Indian or Alaska Native,'' ``Asian,''
and ``Native Hawaiian or Other Pacific Islander.'' There are two
minimum categories for data on ethnicity: ``Hispanic or Latino'' and
``Not Hispanic or Latino.'' The OMB standards advise that respondents
shall be offered the option of selecting one or more racial
designations. The OMB standards also advise that race and ethnicity are
two distinct concepts; therefore, Hispanics or Latinos may be any race.
The minimum categories for data on race and ethnicity for Federal
statistics, program administrative reporting, and civil rights
compliance reporting are defined by OMB as follows:
American Indian or Alaska Native--A person having origins
in any of the original peoples of North and South America (including
Central America), and who maintains tribal affiliation or community
attachment.
Asian--A person having origins in any of the original
peoples of the Far East, Southeast Asia, or the Indian subcontinent
including, for example, Cambodia, China, India, Japan, Korea, Malaysia,
Pakistan, the Philippine Islands, Thailand, and Vietnam.
Black or African American--A person having origins in any
of the black racial groups of Africa. Terms such as ``Haitian'' or
``Negro'' can be used in addition to ``Black or African American.''
Hispanic or Latino--A person of Mexican, Puerto Rican,
Cuban, South or Central American, or other Spanish culture or origin,
regardless of race. The term, ``Spanish origin,'' can be used in
addition to ``Hispanic or Latino.''
Native Hawaiian or Other Pacific Islander--A person having
origins in any of the original peoples of Hawaii, Guam, Samoa, or other
Pacific Islands.
White--A person having origins in any of the original
peoples of Europe, the Middle East, or North Africa.
The 1997 OMB standards state the minimum categories that must be
used to collect and present federal data on race and ethnicity.
Additionally, the 1997 OMB standards permit the collection of more
detailed information on population groups, provided that any additional
groups can be aggregated into the minimum standard set of categories.
Currently, the Census Bureau collects additional detailed information
on Hispanic or Latino groups, American Indian and Alaska Native tribes,
Asian groups, and Native Hawaiian and Other Pacific Islander groups.
For example, responses to the race question such as ``Navajo
Nation,'' ``Doyon,'' and ``Mayan'' are collected and tabulated in
Census Bureau censuses and surveys, and can be aggregated into the
total American Indian or Alaska Native population. Detailed responses
to the race question such as ``Chinese,'' ``Asian Indian,'' and
``Vietnamese'' are collected and tabulated, and can be aggregated into
the total Asian population. Responses to the ethnicity question such as
``Mexican,'' ``Puerto Rican,'' and ``Cuban'' are collected and
tabulated in Census Bureau censuses and surveys, and can be aggregated
into the total Hispanic or Latino population. Responses to the race
question such as ``Native Hawaiian,'' ``Chamorro,'' or ``Fijian'' are
collected and tabulated, and can be aggregated into the total
[[Page 71378]]
Native Hawaiian or Other Pacific Islander population.
The 2015 NCT will test ways to collect and tabulate detailed
information for all groups, including data for White groups, such as
German, Irish, and Lebanese, and data for Black groups, such as African
American, Jamaican, and Nigerian, which have not been tabulated
previously from the question on race. Responses to the race question
such as ``African American,'' ``Jamaican,'' or ``Nigerian'' will be
collected and tabulated, and can be aggregated to the total Black or
African American population. Responses to the race question such as
``German,'' ``Irish,'' or ``Lebanese'' will be collected and tabulated,
and can be aggregated into the total White population.
The 2015 NCT will also test a separate ``Middle Eastern or North
African'' category and the collection of detailed groups such as
``Lebanese,'' ``Egyptian,'' and ``Iranian.'' Following the current OMB
standards, Middle Eastern and North African responses are classified as
``White.''
The results of the 2015 NCT will guide future collection and
tabulation of detailed information for all race and ethnicity groups.
Plans for the 2020 Census call for the use of less costly and more
efficient web-based response options to collect information, as opposed
to a previous predominant reliance on paper-based questionnaires. One
benefit of the online response mode is that it allows for more
functionality and greater flexibility in designing questions compared
to paper, which is constrained by space availability. With the
advantage of new technology, the 2015 National Content Test will
utilize web-based technology, such as internet, smart phone, tablet,
and telephone to improve question designs and optimize reporting of
detailed racial and ethnic groups (e.g., Samoan, Iranian, Blackfeet
Tribe, Filipino, Jamaican, Puerto Rican, Irish, etc.).
The web-based designs provide much more utility and flexibility for
using detailed checkboxes and write-in spaces to elicit and collect
data for detailed groups than traditional paper questionnaires, and
will help collect data for both the broader OMB categories, as well as
detailed responses across all groups.
Components of the Test
A. Race and Origin Content
The Census Bureau conducted an extensive research undertaking as
part of the 2010 Census--the 2010 Census Race and Hispanic Origin
Alternative Questionnaire Experiment (AQE) (for details, see
www.census.gov/2010census/news/press-kits/aqe/aqe.html). The 2010 AQE
examined alternative strategies for improving the collection of data on
a race and Hispanic origin, with four goals in mind:
1. Increasing reporting in the standard race and ethnic categories
as defined by the U.S. Office of Management and Budget;
2. Decreasing item non-response for these questions;
3. Increasing the accuracy and reliability of the results for this
question; and
4. Eliciting detailed responses for all racial and ethnic
communities (e.g., Chinese, Mexican, Jamaican, etc.).
The results of the AQE supported all of these objectives.
Additionally, many individuals across communities liked the combined
question approach. They believed it presented equity to the different
categories. Some of the findings from this research include:
Combining race and ethnicity into one question did not
change the proportion of people who reported as Hispanics, Blacks,
Asians, American Indians and Alaska Natives, or Native Hawaiians and
Other Pacific Islanders.
The combined question yielded higher response rates.
The combined question increased reporting of detailed
responses for most groups, but decreased reporting for others.
The combined question better reflected self-identity.
The successful strategies from the AQE research have been employed
in the design of the Census Bureau's mid-decade research. Four key
dimensions of the questions on race and Hispanic origin are being
tested in the 2015 NCT. These include question format, response
categories, wording of the instructions, and question terminology.
Question Format
The 2015 NCT will evaluate the use of two alternative question
format approaches for collecting data on race and ethnicity. One
approach uses two separate questions: the first about Hispanic origin
and the second about race (``separate questions''). The other approach
combines the two items into one question about race and origin
(``combined question''). The 2015 mid-decade research will test the
approaches with new data collection methods, including internet,
telephone, and in-person response.
1. Separate race and origin questions: This is a modified version
of the race and Hispanic origin format used in the 2010 Census. Updates
since the 2010 Census include added write-in spaces and examples for
the ``White'' and ``Black or African Am.'' response categories, removal
of the term ``Negro,'' and an instruction to select one or more boxes
in the Hispanic origin question.
2. Combined question with checkboxes and write-ins on same screen:
This is a modified version of the combined question approaches found to
be successful in the 2010 AQE. Checkboxes are provided for the U.S.
Office of Management and Budget (OMB) standard categories (per the 1997
Standards for the Classification of Federal Data on Race and Ethnicity)
with a corresponding write-in space for each checkbox category. In this
version, all write-in spaces are visible at all times. Each response
category contains six example origins, which represent the diversity of
the geographic definitions of the OMB category. For instance, the
``Asian'' category examples of Chinese, Filipino, Asian Indian,
Vietnamese, Korean, and Japanese represent the six largest detailed
Asian groups in the United States, reflecting OMB's definition of Asian
(``A person having origins in any of the original peoples of the Far
East, Southeast Asia, and the Indian subcontinent.''). Respondents do
not have to select an OMB checkbox, but may enter a detailed response
in the write-in space without checking a category.
a. Combined question with checkboxes and write-ins on separate
screens (Internet-only): In this version, the detailed origin groups
are solicited on subsequent screens after the OMB response categories
have been selected. On the first screen, the OMB checkbox categories
are shown along with their six representative example groups. Once the
OMB categories have been selected, one at a time, subsequent screens
solicit further detail for each category that was chosen (e.g., Asian),
using a write-in space to collect the detailed groups (e.g., Korean and
Japanese). The intent is to separate mouse click tasks (checkbox
categories) and typing tasks (write-ins) in an attempt to elicit
responses that are more detailed. The same version was used as one of
three race and origin Internet panels in the 2014 Census Test.
3. Combined question branching with detailed checkbox screens
(Internet-only): This version is an alternative method of soliciting
detailed origin groups using separate screens, detailed checkboxes, and
write-in spaces. On the first screen, the OMB checkbox categories are
shown along with their six representative example groups. Once the OMB
categories have been selected, one at a time, subsequent screens
solicit
[[Page 71379]]
further detail for each category, this time using a series of
additional checkboxes for the six largest detailed groups (e.g.,
Chinese, Filipino, Asian, Indian, Vietnamese, Korean, and Japanese)
with a write-in space also provided to collect additional groups.
Race Response Categories
The 2015 National Content Test will evaluate the use of the Middle
Eastern or North African (MENA) category in the race question. There
will be two treatments for testing this dimension:
1. Use of MENA category: This treatment tests the addition of a
MENA checkbox category to the race question. The MENA category is
placed within the current category lineup, based on estimates of
population size, between the categories for Native Hawaiians and Other
Pacific Islanders and ``Some other race.'' With the addition of this
new category, the ``White'' example groups are revised. The Middle
Eastern and North African examples of ``Lebanese'' and ``Egyptian'' are
replaced with the European examples of ``Polish'' and of ``French.''
The MENA checkbox category will have the examples of ``Lebanese,
Iranian, Egyptian, Syrian, Moroccan, Algerian, etc.'' All other
checkbox categories and write-in spaces remain the same.
2. No separate MENA category: This treatment tests approaches
without a separate MENA checkbox category, and represents the current
OMB definition of White (``A person having origins in any of the
original peoples of Europe, the Middle East, or North Africa.''). Here
we will provide examples of Middle Eastern and North African origins
(``Lebanese'' and ``Egyptian'') with European origin groups as part of
the ``White'' racial category.
Wording of the Instructions
1. ``Mark [X] one or more boxes'': The current paper version of the
instructions on paper states, ``Mark [X] one or more boxes AND print
your specific origin(s).''
2. ``Mark all that apply/You may mark multiple groups'': In this
version, the instruction is modified to ``Mark all boxes that apply AND
print the specific [origin(s)/ethnicities] in the spaces below. Note,
you may report more than one group.'' Recent qualitative focus groups
and cognitive research (e.g., 2010 AQE research; 2013 Census Test
research) found that respondents frequently overlook the instruction to
``Mark'' [X] one or more boxes. The research found that some
respondents may have stopped reading the instruction after noticing the
visual cue [X] and proceeded directly to do just that--mark a box--
overlooking the remainder of the instruction. The new instruction
(``Mark all boxes that apply'') is an attempt to improve the clarity of
the question and make it more apparent that more than one group may be
selected.
Question Terms
1. ``Origin'' term: The current version of the race and Hispanic
origin questions use the terms ``race'' and/or ``origin'' to describe
the concepts and groups in the question stem, instructions, and
examples. For instance, in the combined race and Hispanic origin
approach, the question stem is ``What is your race or origin?'' In
addition, prior to each write-in field, respondents are instructed to
``Print specific origin(s), for example . . .''
2-3. Alternative terms: Recent qualitative focus groups and
qualitative research (e.g., 2010 AQE research; 2013 Census Test
research; cognitive pre-testing for 2016 American Community Survey
Content Test) found that the term ``origin'' is confusing or misleading
to many respondents, who may think it is asking about where they
immigrated from or where they were born. Two alternative options are
being explored in cognitive testing and usability research. One
approach tests the use of the term ``ethnicities'' along with ``race''
(e.g., ``Print the specific races(s) and/or ethnicities . . .''). The
other approach tests the removal of the terms altogether from the
question stem, instructions, and examples. Instead, a general approach
asks, ``Which categories describe this person?'' The exact terminology
to be used for the alternative version is pending cognitive testing and
usability results later this year, which will inform the wording to be
used in the 2015 NCT.
B. Relationship Content
Two versions of the relationship question will be tested. Both
versions are the same as those used in a split-sample in the 2014
Census Test, with no changes. The new relationship categories have also
been tested in other Census Bureau surveys including the American
Housing Survey, American Community Survey, and the Survey of Income and
Program Participation (currently used in production). Although research
to date has been informative, leading to the development of the revised
relationship question, additional quantitative testing is needed. Since
the incidence of some household relationships--such as same-sex
couples--is relatively low in the general population, the revised
question needs to be tested with large, representative samples prior to
routinely including them in the 2020 Census questionnaire.
The first version uses the 2010 Census relationship question
response options, but in a new order, starting with ``husband or wife''
and then the ``unmarried partner'' category. This version also re-
introduces the foster child category, which was removed from the 2010
Census form due to space issues.
The second version includes the same basic response options as the
2010 Census version, but modifies/expands the ``husband or wife'' and
``unmarried partner'' categories to distinguish between same-sex and
opposite-sex relationships.
C. Coverage Content (Internet Only)
The 2012 National Census Test experimented with several methods to
improve accurate within-household coverage for Internet respondents.
One benefit of the online response mode is that it allows for more
functionality and greater flexibility in designing questions compared
to paper, which is constrained by space availability. The 2012 test
included a coverage follow-up reinterview to evaluate the different
Internet design options, but some results were inconclusive. In the
2015 NCT, two designs will be tested to compare different approaches
for helping respondents provide a more accurate roster of household
residents.
The first approach is the ``Rules-Based'' approach, and will allow
us to see whether the presence of a question asking the number of
people in the household along with the residence rule instructions
helps respondents create an accurate roster. This is similar to the
approach used across all modes in Census 2000 and the 2010 Census,
where the respondent was expected to understand our residence rules and
apply them to their household. This is followed by a household-level
question that probes to determine if any additional people not listed
originally should be included for consideration as residents of the
household (several types of people and living situations are shown in a
bulleted list).
The ``Question-Based'' approach allows us to ask guided questions
to help improve resident information. Respondents are not shown the
residence rule instructions and are only asked to create an initial
roster of people they consider to be living or staying at their address
on Census Day. This is followed by several short household-level
questions about types of people and living situations that might apply
to
[[Page 71380]]
someone in the household that was not listed originally.
D. Optimizing Self Response
The nine proposed contact strategies for optimizing self response
(OSR) are summarized as follows:
Internet Push (Control): This is the standard Internet Push
strategy used in the most recent series of self response tests,
including the 2014 Census Test. This panel will serve as a control
panel against which to compare the experimental strategies. There will
be nine treatments as part of the OSR test.
Internet Push With Early Postcard: The motivation for this panel is
to study the timing of reminders. The hypothesis is that sending the
first reminder sooner (closer to the initial Internet push) would
provide for a better connection between the two mailings, and could
increase response. A side benefit is that this could also reduce the
volume of later targeted mailings since responses may be quicker
overall.
The motivation for the following sequence of three panels is based
on recent American Community Survey (ACS) research, which has found
depressed self response rates among certain respondents/areas with
lower Internet usage. Testing the delivery of the paper questionnaires
at various points in the response process will allow us to have
complete response measures under several scenarios for the cost/benefit
analysis needed to inform 2020 Census planning. Although these
strategies may not make sense for everyone in 2020, using a responsive
design and tailoring the contact strategy for certain geographic areas
or populations may be beneficial.
Internet Push With Early Questionnaire: questionnaire sent
at third mailing, one week sooner
Internet Push With Even Earlier Questionnaire:
questionnaire sent at second mailing, two weeks sooner
Internet Choice: questionnaire sent at first mailing,
providing a choice of Internet or paper from the beginning
Internet Push With Postcard as Third Reminder: The motivation for
this panel is to further encourage self response, after the
questionnaire mailing, prior to nonresponse follow-up. Numerous survey
research studies have concluded that, while there is a point of
diminishing returns, further reminders will inevitably increase self
response rates.
Internet Push Postcard: The motivation for this panel is to study
the impact of sending a postcard at the first mailing instead of a
letter. There are two potential benefits. First is the possible cost
savings of printing and mailing a postcard compared to the envelope
package (with letter and instruction card). Second is the potential for
increased self response because reading a postcard requires less effort
by a respondent. In this panel, we send a letter at the third contact
(sent to non-respondents only), in place of a postcard, to vary the
types of contacts received.
Internet Push With Early Postcard and Second Letter Instead of Mail
Questionnaire: The motivation for testing an approach in which we do
not send a mail questionnaire is to address the high-level goal of
greatly reducing paper responses in the 2020 Census. By testing an
approach in which we send an Internet push letter in place of a paper
questionnaire at the fourth mailing, we will have a more robust set of
response measures for informing cost/benefit analyses.
Internet Push With Postcard and Email as 1st Reminder (same time):
The motivation for this panel is to determine if we can take advantage
of the email addresses in the supplemental contact frame maintained by
the Center for Administrative Records Research and Applications. The
hypothesis is that by sending a postcard and email at the same time, we
may be able to elicit increased response.
E. Language
In the two mailings that contain a letter for each Optimizing Self
response strategy, three different methods will be used to encourage
response. In particular, by altering the language support provided in
the letter, the goal is to increase response for respondents with
limited English proficiency.
The control panel is similar to the 2014 Census Test design, in
which the mailing materials are in English with a single Spanish
sentence directing respondents to the Web site or the telephone
assistance line.
One of the goals of language research is to maximize the number of
non-English speakers that receive the same message as English speakers
prior to going online to respond. Two panels provide equality between
the English and Spanish content in the letter and test whether one
method is better at eliciting Spanish responses. The swim-lane design
has been used in the past, such as with the bilingual questionnaire in
the 2010 Census. The dual-sided letter provides English content on one
side and Spanish content on the other side. In addition, because
research has shown that Spanish-speaking respondents do not always open
the mailings because they may not know that language resource
information is provided inside, the outgoing envelope for both panels
will include the census test Web site URL and a brief message in both
languages.
This test will also explore additional options for non-English
speakers to complete the questionnaires.
F. Content Reinterview
A sub-sample of respondents from the 2015 NCT will be selected for
a content reinterview, focused on race and origin and within-household
coverage, with a goal of assessing accuracy and reliability of the
different designs. Reinterviews are conducted with a sub-sample of
respondents, by asking more detailed questions on question topics, in
order to assess the accuracy of the responses.
II. Method of Collection
The initial mail-out is planned for late August 2015. This contact
will explain why we are conducting the mandatory 2015 NCT, assure
respondents that their answers are confidential, and inform them of the
measures we take to keep their personal information secure. The second
mail-out is considered a reminder and is sent to all housing units. All
contacts after the second mailing are sent to non-respondents only.
Respondents are encouraged to respond to the 2015 NCT by Internet
but may also be able to provide information by phone. Many will also
receive a paper questionnaire at some point in the mail-out strategy.
The test will be conducted nationally in all 50 U.S. states, the
District of Columbia, and Puerto Rico.
III. Data
OMB Control Number: None.
Form Number: TBD.
Type of Review: Regular submission.
Affected Public: Individuals or households.
Estimated Number of Respondents: 1.3 million households. (1.2
million initial response + 100,000 reinterview).
Estimated Time per Response: 10 minutes.
Estimated Total Annual Burden Hours: 216,667.
Estimated Total Annual Cost: There is no cost to respondents except
for their time to respond.
Respondent's Obligation: Mandatory.
Legal Authority: Title 13 U.S.C. 141 and 193.
IV. Request for Comments
Comments are invited on: (a) Whether the proposed collection of
information is necessary for the proper performance of the functions of
the agency, including whether the information shall have
[[Page 71381]]
practical utility; (b) the accuracy of the agency's estimate of the
burden (including hours and cost) of the proposed collection of
information; (c) ways to enhance the quality, utility, and clarity of
the information to be collected; and (d) ways to minimize the burden of
the collection of information on respondents, including through the use
of automated collection techniques or other forms of information
technology.
Comments submitted in response to this notice will be summarized
and/or included in the request for OMB approval of this information
collection; they also will become a matter of public record.
Dated: November 25, 2014.
Glenna Mickelson,
Management Analyst, Office of the Chief Information Officer.
[FR Doc. 2014-28247 Filed 12-1-14; 8:45 am]
BILLING CODE 3510-07-P