Document 263794

PREFACE
2002 NSAF Sample Design is the second report in a series describing the methodology of the
2002 National Survey of America’s Families (NSAF). The NSAF is part of the Assessing the
New Federalism project at the Urban Institute, conducted in partnership with Child Trends. Data
collection for the NSAF was conducted by Westat.
The NSAF is a major household survey focusing on the economic, health, and social
characteristics of children, adults under the age of 65, and their families. During the third round
of the survey in 2002, interviews were conducted with over 40,000 families, yielding information
on over 100,000 people. The NSAF sample is representative of the nation as a whole and of 13
states, and therefore has an unprecedented ability to measure differences between states.
About the Methodology Series
This series of reports has been developed to provide readers with a detailed description of the
methods employed to conduct the 2002 NSAF. The 2002 series of reports include:
No. 1:
An overview of the NSAF sample design, data collection techniques, and
estimation methods
No. 2:
A detailed description of the NSAF sample design for both telephone and inperson interviews
No. 3:
Methods employed to produce estimation weights and the procedures used to
make state and national estimates for Snapshots of America’s Families
No. 4:
Methods used to compute and results of computing sampling errors
No. 5:
Processes used to complete the in-person component of the NSAF
No. 6:
Collection of NSAF papers
No. 7:
Studies conducted to understand the reasons for nonresponse and the impact of
missing data
No. 8:
Response rates obtained (taking the estimation weights into account) and methods
used to compute these rates
No. 9:
Methods employed to complete the telephone component of the NSAF
No. 10:
Data editing procedures and imputation techniques for missing variables
No. 11:
User’s guide for public use microdata
No. 12:
2002 NSAF questionnaire
About This Report
Report No. 2 describes the sample design for the 2002 NSAF. As in previous rounds of the
survey, the 2002 NSAF sample consists of a random digit dial (RDD) telephone sample
supplemented by an area probability sample of nontelephone households. While the
nontelephone sample for previous NSAF rounds were both nationally and state representative,
the 2002 NSAF used only a nationally representative nontelephone sample. The report covers
both the telephone and nontelephone sample design, adjustments made to the sample design
during the field period, within household sampling procedures and achieved sample sizes.
For More Information
For more information about the National Survey of America’s Families, contact:
Assessing the New Federalism
Urban Institute
2100 M Street, NW
Washington, DC 20037
E-mail: nsaf@ui.urban.org
Web site: http://anf.urban.org/nsaf
Adam Safir and Tim Triplett
ii
CONTENTS
Chapter
Page
1
OVERVIEW ..........................................................................................
1-1
2
PRINCIPAL FEATURES OF SAMPLE DESIGN BY ROUND .........
2-1
2.1
2.2
2.3
2.4
The Survey .................................................................................
Survey Components ...................................................................
Number of Completed Interviews..............................................
Projected Effective Sample Size................................................
2-1
2-3
2-6
2-8
RANDOM DIGIT DIAL HOUSEHOLD SAMPLING ........................
3-1
3.1
3.2
3.3
3.4
3.5
3.6
3-1
3-2
3-7
3-8
3-9
3-13
3
4
Sampling Telephone Numbers...................................................
Subsampling Households...........................................................
Subsampling Adult-Only Households .......................................
Subsampling High-Income Households ....................................
Household Sampling Revisions during Data Collection ...........
Achieved Response and Eligibility Rates ..................................
AREA SAMPLE....................................................................................
4-1
4.1
4.2
4-1
4-4
First-Stage Sampling..................................................................
Second-Stage Sampling .............................................................
4.2.1
Exclusion of Block Groups with High Telephone
Coverage Rates............................................................
Segment Stratification and Selection...........................
4-5
4-5
Chunk Selection .........................................................................
Achieved Response and Eligibility Rates ..................................
4-6
4-7
WITHIN-HOUSEHOLD SAMPLING AND ACHIEVED
SAMPLE SIZES ....................................................................................
5-1
4.2.2
4.3
4.4
5
5.1
5.2
5.3
5.4
Sampling Children .....................................................................
Sample Selection of Other Adults in Households with
Children......................................................................................
Sample Selection of Adults from Adult-Only Households........
Achieved Sample Sizes and Response Rates.............................
iii
5-1
5-1
5-2
5-3
CONTENTS (continued)
Chapter
6
Page
CONCLUSION......................................................................................
6-1
REFERENCES ......................................................................................
R-1
Number of Completed Interviews by Round, Sample Type,
and Interview Type ................................................................................
2-7
Projected Design Effects for the RDD Sample, by Interview Type
and Study Area.......................................................................................
2-10
Projected Effective Sample Sizes for the RDD Sample, by Interview
Type and Study Area .............................................................................
2-10
Assumed Proportion of Households by Household Type
and Poverty Status..................................................................................
3-5
3-2
Assumed Misclassification Rates, by Income Categories .....................
3-6
3-3
Assumed Residential and Response Rates.............................................
3-6
3-4
Subsampling or Household Retention Rates..........................................
3-7
3-5
Income Levels for Determining Less than 200 Percent
of the Poverty Level...............................................................................
3-8
Revised Subsampling Rates for Households with Children
Screening as High-Income, by Release Group ......................................
3-12
Revised Subsampling Rates for Adult-Only Households,
by Release Group...................................................................................
3-12
3-8
Reserve Sample Released, by Study Area .............................................
3-13
3-9
Screening for Residential Status and Presence of Children...................
3-14
Tables
Table
2-1
2-2
2-3
3-1
3-6
3-7
iv
CONTENTS (continued)
Tables (continued)
Table
Page
3-10
Subsampling Screener Refusals and Response Rates............................
3-15
3-11
Outcomes of Household Screening of Telephone Households..............
3-17
3-12
Outcomes of Income Screening of Telephone Households
with Children .........................................................................................
3-18
Outcomes of Income Screening of Adult-Only Telephone
Households.............................................................................................
3-19
4-1
Number of Round 3 Primary Sampling Units........................................
4-3
4-2
Maximum Telephone Service Rates Allowed in Covered
Block Groups .........................................................................................
4-5
Segment Counts, by Planned and Unplanned Chunking (Including
Sample Supplement) ..............................................................................
4-7
4-4
Outcomes of Area Listing......................................................................
4-8
4-5
Outcomes of Area Prescreening and Screening.....................................
4-8
5-1
Proportion of Random Digit Dial Adult-Only Households
with Three Other Adults under Age 65 in which Just One Adult
Is Selected ..............................................................................................
5-3
Within-Household Sampling and Extended Interviews of Children in
Telephone Low-Income Households .....................................................
5-6
Within-Household Sampling and Extended Interviews of Children in
All Subsampled Telephone Households ................................................
5-7
Within-Household Sampling and Extended Interviews of Other
Adults in Subsampled Telephone Households with Children ...............
5-8
3-13
4-3
5-2
5-3
5-4
v
CONTENTS (continued)
Tables (continued)
Table
5-5
Page
Subsampling and Extended Interviews of Adults in Subsampled
Adult-Only Telephone Households .......................................................
5-9
5-6
Sources of Adult Telephone Interviews.................................................
5-10
5-7
Within-Household Sampling and Extended Interviews of Children
in Nontelephone Low-Income Households ...........................................
5-11
Within-Household Sampling and Extended Interviews of Children
in All Nontelephone Households ...........................................................
5-12
Within-Household Sampling and Extended Interviews of Other
Adults in Nontelephone Households with Children ..............................
5-13
Subsampling and Extended Interviews of Adults in Adult-Only
Nontelephone Households .....................................................................
5-14
5-11
Sources of Adult Nontelephone Interviews ...........................................
5-15
5-12
Within-Household Sampling and Extended Interviews of Children
in Telephone and Nontelephone Low-Income Households...................
5-16
Within-Household Sampling and Extended Interviews of Children
in All Telephone and Nontelephone Households ..................................
5-17
Within-Household Sampling and Extended Interviews of Other
Adults in Telephone and Nontelephone Households with Children......
5-18
Subsampling and Extended Interviews of Adults in Adult-Only
Telephone and Nontelephone Households.............................................
5-19
Sources of Adult Telephone and Nontelephone Interviews ..................
5-20
5-8
5-9
5-10
5-13
5-14
5-15
5-16
vi
CONTENTS (continued)
Figures
Figure
Page
2-1
Study Areas............................................................................................
2-1
2-2
Sampling Frame Inclusions and Exclusions ..........................................
2-2
3-1
Household Subsampling Operations......................................................
3-3
vii
1. OVERVIEW
This report describes the sample design for the 2002 National Survey of America’s Families
(NSAF). This survey is the third round of the NSAF, and the objective is to estimate both the
characteristics of households and persons in 2002 and changes in those characteristics since the
1997 and 1999 NSAF. While the designs for all three rounds of the NSAF are similar, several
important differences exist. The first survey in 1997 was a dual-frame survey of both households
with telephones and those without telephones developed to serve as a baseline for evaluating
changes over time. The Round 2 survey was designed to improve estimates of change between
1997 and 1999 by retaining a substantial portion of the Round 1 sample. Analysis of the Round 2
data showed that the design changes did not improve the precision for estimates of change
between rounds as well as expected. Furthermore, the retention of a portion of the sample
resulted in additional operational and design complications.
Based on the findings from Round 2, the sample design for Round 3 was developed to be similar
to the Round 1 design, in the sense that the sample was mainly independent of the sample from
previous rounds. However, the sample design for Round 3 did include important modifications
from the previous rounds’ designs that were intended to reduce data collection costs. The most
important design change was the reduction of the sample size for nontelephone households in the
study areas. This change also has important implications for the estimation strategy, which is
discussed in 2002 NSAF Sample Estimation Survey Weights, Report No. 3.
This report describes the sample design and how it relates to the designs from previous rounds. It
also provides the details needed to appreciate the considerations that went into the decisions that
resulted in the features of this large and complex survey.
Chapter 2 summarizes the survey goals and the sampled units, and introduces its two major
components, the telephone and in-person surveys. One of the main objectives of Chapter 2 is to
describe the similarities and differences between the Round 3 sample design and the designs for
the previous rounds. The remaining chapters focus primarily on the Round 3 design. Chapter 3
describes the random digit dial (RDD) telephone sample design and the sampling of households
in the telephone component. The subsampling procedures for households without children and
for high-income households are included in this chapter. Chapter 4 gives a detailed account of
the sampling for the in-person survey component. It discusses the changes in the sampling
needed to move from a sample for each study area to an overall national sample for this
component of the survey. Chapter 5 presents the methods used to sample children and adults
from within the sampled households. It contains tables on the number of sampled and
interviewed persons from the survey. Chapter 6 provides some concluding remarks.
1-1
2. PRINCIPAL FEATURES OF SAMPLE DESIGN BY ROUND
The samples for all three rounds of NSAF—Round 1 in 1997, Round 2 in 1999, and Round 3 in
2002—have similar designs. A sample design Report No. 2 similar to this one is available for
each earlier survey. These reports describe the details of the sample design for the specific
survey. In this chapter we discuss the design features for Round 3 in relation to the features from
the previous rounds. We focus on how the Round 3 design differs from that used in Round 1 and
Round 2. The specifics of the Round 3 design are given in subsequent chapters.
2.1
The Survey
The NSAF collected information on the economic, health, and social dimensions of the wellbeing of children, adults under the age of 65, and their families in 13 states and the balance of the
nation. The Urban Institute selected these study areas (see figure 2-1) in 1996 prior to the first
survey because they represent a broad range of fiscal capacity, child well-being, and approaches
to government programs. Data were also collected in the balance of the nation to permit
estimates for the United States as a whole.
Figure 2-1.
Study Areas
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Balance of nation
In Round 1 and Round 2, Milwaukee County in Wisconsin was a separate study area that had its
own sample. A separate sample from the balance of Wisconsin was selected in these rounds to
produce estimates for the entire state of Wisconsin by combining Milwaukee and the balance of
Wisconsin. In Round 3, the separate Milwaukee County study area was eliminated and the entire
state of Wisconsin was treated as a single study area, as shown in figure 2-1.
The primary goal of the survey in all three rounds was to obtain social and economic information
about children in low-income households (those with incomes below 200 percent of the poverty
threshold), since the impact of New Federalism was likely to be greatest on these children.
Secondary goals included obtaining similar data on children in higher-income households, plus
adults under age 65 (with and without children).
With few exceptions, the decision was made to limit the survey to children, adults, and families
living in regular housing. Figure 2-2 explains the concept of regular housing through examples
of specific inclusions and exclusions. Although one impact of New Federalism could be the
displacement of persons from regular housing, including the population that lives outside of
2-1
Figure 2-2.
Sampling Frame Inclusions and Exclusions
Inclusions
Houses, apartments, and mobile homes occupied by individuals, families, multiple families,
or extended families where at least one occupant is under the age of 65
Houses, apartments, and mobile homes occupied by multiple unrelated persons, provided that
the number of unrelated persons is less than nine and at least one occupant is under the age of
65
People in workers’ dormitories and camps
Military personnel living on post with their families, as well as military personnel living off
post with or without their families
Included Persons in Excluded Structures
People living temporarily away from home were enumerated at their usual residences. This
includes college students in dormitories, patients in hospitals, vacationers, business travelers,
snowbirds, and so on. Structures that were expected to primarily include only such people
were excluded
Exclusions
The institutionalized population. Examples of institutions include prisons, jails, juvenile
detention facilities, psychiatric hospitals and residential treatment programs, and nursing
homes for the disabled and aged
Noninstitutional group quarters, including communes, monasteries, convents, group homes
for the mentally or physically disabled, shelters, halfway houses, dormitories, and dwelling
units with nine or more unrelated persons
The homeless
People in transient hotel/motel rooms, tents, recreational vehicles, trailers, and other similar
temporary arrangements
Military barracks and ships
2-2
regular housing was considered infeasible within the survey context. The elderly population was
also excluded. College students were enumerated at their parents’ homes. Most of these
inclusions and exclusions are typical of those made in other household surveys. For example, the
Current Population Survey (CPS) has essentially the same rules (U.S. Bureau of the Census
2000). The major difference is that the CPS includes military personnel living on post with
families but excludes those living in noninstitutional group quarters.
2.2
Survey Components
The sample for each round of NSAF has two separate components: an RDD survey of
households with telephones, and an area survey of households without telephones. The RDD
component provides a cost-effective method to collect the desired data on a large number of
households for each study area and nationally. The area component enables the survey to cover
households without telephones. The area component is important in this survey because lowincome households are more highly concentrated in nontelephone households than the entire
universe of households. For example, Giesbrecht et al. (1996) estimated that about 20 percent of
families in poverty live in households without a telephone and that about 10 percent of families
with one child 3 years old or under have no telephone based on CPS data. Even though these
results are out-of-date given the changes in technology in the past 5 to 10 years, the only recent
data on the relationship of poverty and having a telephone in the household published are from
the 2000 Census and are discussed below. Even these data do not distinguish between landline
and wireless telephones. After the area sample was selected, households were screened to find
households without telephones. Only these nontelephone households were interviewed from the
area sample.
When Round 1 was designed, the dual-frame approach to this type of problem was relatively
new and data to optimize various aspects of the design were not available. Despite the lack of
data, it was clear that the dual-frame approach produced more precise estimates than a pure areasampling approach of the same cost. In addition, it produced less biased estimates than a pure
RDD approach of the same cost. Waksberg et al. (1997) describe some of the options that were
considered early in the process of designing the Round 1 survey.
The Round 2 sample design used data from Round 1 to improve survey efficiency. The research
supporting revisions to the Round 2 design are described in chapter 3 of 1999 NSAF Sample
Design, Report No. 2. The primary emphasis in Round 2 shifted from producing estimates of the
current level (the Round 1 design objective) to estimating changes from 1997 to 1999. As a
result, a substantial proportion of the sample from Round 1 was retained and included in the
Round 2 sample. For the RDD sample, the proportion of the sample of telephone numbers
retained depended on the outcome of the Round 1 interview. Round 1 telephone numbers that
were residential and cooperative were sampled at higher rates than those numbers that were
either nonresidential or uncooperative. In addition to retaining a portion of the Round 1 sample,
telephone numbers not in existence in Round 1 were sampled to provide complete coverage of
telephone households in 1999. The Round 2 area sample was largely the same as used in
2-3
Round 1, but an additional sample from the balance of the nation was added to improve the
precision of national estimates.
Analysis of the Round 2 data found that the expected improvements in the precision of change
estimates due to retaining a portion of the Round 1 sample were not as large as had been
expected. The reduction in variances for change estimates depended on retaining a substantial
fraction of the Round 1 households and the characteristics of the retained households being
highly positively correlated across time. The design assumptions regarding both of these key
factors were somewhat optimistic. In the component of the sample in which an interview was
completed in Round 1 and the telephone number was retained for Round 2, the proportion of
identical households was less than assumed. In addition, even when the same household
responded in both rounds, the correlation for important estimates was lower than expected.
Given these Round 2 findings, it was decided that it would be unwise to retain a sample of
telephone numbers from the previous rounds for Round 3. Retaining telephone numbers has its
own costs in terms of introducing differential sampling fractions that decrease the precision of
the estimates and lower response rates. Report No. 8 in the 1999 series showed the screener
response rate in the retained sample was about 2 or 3 points lower than the response rate in the
newly sampled numbers from the same sampling frame. This result is consistent with panel
surveys (see Kalton, Kaspryzk, and McMillen, 1989). We examine the 2002 response rates in
Report No. 8 and provide a more direct assessment of the effect of retaining telephone numbers
in the sample there. The retained sample also added some complexity operationally. Thus, it was
decided that the Round 3 telephone sample should be independent of the samples selected in
Round 1 and Round 2. Although the sample for Round 3 was selected independently, the
allocation of the sample to the study areas and the sampling procedures were very similar to
those used in Round 1.
On the other hand, the design for the Round 3 area sample was very different from that used in
the previous rounds. In Round 1 and Round 2, separate area samples were selected in each study
area and for the balance of the nation. The area and RDD samples were combined to produce
estimates of both telephone and nontelephone households for each study area and nationally.
Because of the cost of screening a large sample of households to find nontelephone households,
the sample size for the area sample was relatively small in each of the study areas. The small
sample size of nontelephone households caused some instability in the estimates of the precision
for the study areas (see 1997 Report No. 4 for more details). At the national level, the sample
size was sufficiently large that the variance estimates were reliable.
These concerns about the area sample for producing reliable estimates of nontelephone
households at the study area level led to research to find a lower-cost alternative. Ferraro and
Brick (2001) studied various methods for adjusting a sample of telephone households to account
for the undercoverage of households without telephones. This research found that an approach
called “modified poststratification” had better statistical properties for NSAF study areas than
previously considered approaches. In addition, new data from the 2000 Census of Population
showed the percentage of households without telephones was much smaller than reported in the
2-4
CPS for the same time period. The census estimates only 2.4 percent of households did not have
a telephone, compared with the 4.9 percent reported in the CPS for the same time.
Based on cost, stability of the variance estimates, new research on statistical adjustment methods,
and lower estimates of the percentage on households without telephones, the sample design for
the area sample for Round 3 was changed significantly from that used in previous rounds. The
Round 3 design included a sufficient sample to produce reliable national estimates of all
households, using the area sample to represent households without telephones. At the same time,
the area sample size for the study areas was reduced, and the plan was to rely on the modified
poststratification approach for estimating characteristics of all households for the study areas.
Thus, the sample size for the area sample and the cost of collecting these data were significantly
reduced. Tables later in this chapter show the changes in sample size at the study area level and
overall.
The other basic features of the sample design were very consistent across the three rounds of data
collection. In all three rounds, costs were reduced through the use of screener-based subsampling
of households contacted in the RDD component. In this approach the RDD screening interview
includes a very short income question. Those households that report no children in the household
or reported incomes above 200 percent of the poverty threshold were subsampled. The extended
interview has more detailed and reliable income questions for those included in the subsample.
In Round 1, the inconsistency between the responses to the short and detailed versions of the
income questions was greater than anticipated, and the Round 2 subsampling rates were revised
in Round 2 to account for this difference. In Round 3, subsampling rates similar to those
suggested by the Round 2 research (see 1999 NSAF Sample Design, Report No. 2) were used.
In all the rounds and across both the RDD and the area samples, the number of household
members that could be sampled and interviewed were limited. The main reason to impose these
limits was to reduce the respondent burden for the household as a whole. Even if there were
several children under age 6 in a household, only one was randomly selected. Similarly, only one
child age 6 to 17 was sampled in a household. The most knowledgeable adult (MKA) in the
household for the child was interviewed about the sample child. During the MKA interview,
additional data were collected about the MKA and about the MKA’s spouse/partner, if that
person was living in the same household. The MKA provided all the data about the
spouse/partner. Generally, every question about the MKA was repeated with reference to the
spouse/partner. However, some questions on health insurance and health care usage were asked
about only one of the two. The appropriate person targeted for these questions was randomly
assigned as either the MKA or the spouse/partner. Some questions were asked only about the
MKA, related to feelings, religious activities, and opinions. These items were not repeated for
the spouse/partner because proxy response did not seem sufficiently valid or reliable, and
because self-response on these few questions was operationally impractical. All these rules for
subsampling persons within sample households were the same for all three rounds.
2-5
Two other within-household subsampling steps were used in all three rounds of data collection.
Other adults in households with children (adults who were not the MKA of any children in the
household) and adults in adult-only households were subsampled. The rules for the subsampling
were complex and are described in detail in the 1997 NSAF Sample Design, Report No. 2. Selfresponse was required for sample adults. During the interview with a sample adult, additional
data were collected about the sample adult’s spouse/partner, if they were living in the same
household. As in the MKA interview, the data were always collected by asking the sample
person to respond for the spouse/partner. No attempt was made to collect these data directly from
the spouse of a sample adult. As in the MKA interview, some questions were asked only about
the sample adult, related to feelings, religious activities, and opinions.
In all three rounds of NSAF, the sample design set in place at the beginning of the survey was
revised somewhat as data on the outcomes of the data collection became available. For example,
the expected sample sizes depended on assumptions about residency rates, response rates,
poverty rates, and other parameters. Since the observed rates differed from the assumed rates,
sometimes revisions in sampling rates or the number of sampled units had to be implemented as
the data collection proceeded. For the first two rounds, the adjustments that took place are
documented in the previous sample design reports. In Round 3, the outstanding changes that took
place during the field period involved releasing more telephone numbers than originally planned
due to lower-than-expected residency and response rates, and using a refusal subsampling
procedure for the RDD sample to speed up the data collection and reduce costs slightly. These
changes to the original sample design are discussed in later chapters, along with the other more
minor changes.
2.3
Number of Completed Interviews
The number of completed interviews for each round of NSAF is given in table 2-1. The table
shows the number of completed interview by type of sample (RDD or area sample) and by type
of interview (various types of adult interviews and child interviews). Note that one MKA
interview can provide data on up to two sample children and that there were a few MKA
interviews with parents under the age of 18. In addition to the total number of interviews, the
table also shows the range of the number of completed interviews in each study area and in the
balance of the nation by round.
The table shows that the number of completed RDD interviews is about the same for all three
rounds. The big difference is in the number of completed area interviews, where the number
completed in Round 3 is lower than the previous rounds. This difference is particularly obvious
in the study areas. This difference is a consequence of the sample design change described
earlier.
2-6
Table 2-1.
Number of Completed Interviews by Round, Sample Type, and Interview Type
2-7
Type of Sampled Interview
RDD sample
All adults
Adult MKAs
Other adults in households
with children
Adults in households
without children
Children under age 6
Low-income children under
age 6
Children age 6 to 17
Low-income children age 6
to 17
Area sample
All adults
Adult MKAs
Other adults in households
with children
Adults in households
without children
Children under age 6
Low-income children under
age 6
Children age 6 to 17
Low-income children age 6
to 17
Total
Range for Study Areas
Round 1 Round 2 Round 3 Round 1
Round 2
Round 3
46,621
27,248
2,407
45,025
29,054
3,054
43,133
28,208
2,872
16,966
13,917
12,053
651–1562
342–1,354
466–1,346
12,067
7,246
11,990
4,923
12,088
5,331
613–914
398–539
515–958
258–407
500–941
264–491
21,210
11,813
22,841
8,560
21,864
8,887
1,682
915
121
1,678
876
113
648
294
46
36–299
18–150
2–28
33–236
11–128
1–15
2–90
0–45
0–6
646
689
308
15–121
19–92
548
497
525
463
180
149
9–94
6–87
615
550
582
509
199
168
13–105
11–98
Balance of the Nation
Round 1 Round 2 Round 3
2,347–3,771 2,085–3,746 1,933–3,798 4,913
1,431–2,025 1,309–2,382 1,254–2,196 2,669
102–218
104–279
101–310
225
6,168
3,610
496
7,251
4,603
521
2,019
2,062
2,127
1,179
746
1,462
633
1,987
873
1,103–1,557 1,056–1,881 1,001–1,708 2,064
624–844
470–627
475–779 1,182
2,872
1,097
3,557
1,501
191
120
7
326
178
31
320
143
24
1–38
64
117
153
8–72
7–63
0–32
0–28
69
66
104
86
84
64
6–91
6–80
0–26
0–23
80
69
119
95
103
84
2.4
Projected Effective Sample Size
In surveys like NSAF that sample study areas and specific subgroups of the population at
different rates, the number of completed interviews does not give a full and accurate picture of
the precision of the estimates. Sampling with differential probabilities generally reduces the
precision of estimates aggregated over groups with different rates, even though it improves the
precision for those subgroups sampled at higher rates. For example, households in the study
areas are sampled at higher rates than are households in the balance of the nation, and the
precision of national estimates is lower than it would be if the rates for the groups were identical.
Of course, the higher sampling rates in the study areas provide reliable estimates for these areas
that would not be possible otherwise.
In designing the sample, the losses in efficiency of the samples due to the differential sampling
rates were taken into account in determining the needed sample sizes. The other source of
efficiency losses in typical household surveys is clustering of sampled households in geographic
areas. List-assisted RDD surveys do not cluster households, so the only clustering of sampled
households in the NSAF is for nontelephone households. Since the nontelephone households are
a small proportion of the total sample, we developed our projections of the precision based only
on the RDD sample. The design effect or deff (the ratio of the variance of an estimate under the
actual design to what would be obtained with a simple random sample of the same size) is one
method of accounting for the efficiency of the sample. Another way of thinking about the design
effect is that it is the inverse of the efficiency of the sample, so that a sample with a deff of 2 is
equivalent to a simple random sample of half the sample size. Design effects were estimated
from the samples for the previous two rounds and were reported in the corresponding Report No.
4 for the round. These average deffs were instrumental in projecting design effects for Round 3.
The primary goal of the sample design was to obtain approximately the same number of lowincome child interviews as achieved in Round 2, with a secondary objective of obtaining about
the same number of low-income adult interviews. Since the efficiency of the sample is directly
tied to the rates used for subsampling the different groups, keeping the subsampling rates as
consistent as possible with the optimal rates developed in Round 2 while still completing about
the same number of interviews was also important. Other factors also affected the sample design.
The sample size allocated to the Milwaukee study area could now be used in other areas because
this was dropped as a stand-alone study area for Round 3. The sample size for this area was
reallocated to California, the balance of the nation, and Michigan to improve the precision for
these areas and the national estimates.
Once the sample sizes were determined, projected design effects and effective sample sizes for
the key interview groups could be computed. The interview groups are children, other adults
(who are not MKAs), and all adults. Since the survey is critically interested in people in families
with incomes less than 200 percent of the poverty level, table 2-2 has the projected deffs for both
low-income and all persons. Using these projected design effects and an additional adjustment
due to other losses of efficiency in the sample design, the effective sample sizes for each study
area and for the nation were computed. These projected effective RDD sample sizes are shown in
table 2-3 for all groups.
2-8
The projected design effects and effective sample size shown in tables 2-2 and 2-3 were
computed prior to the start of data collection. Thus, changes in the design that occurred during
data collection and the actual sample sizes achieved are not reflected in these tables. The average
design effects based on the data collected in Round 3 are given in Report No. 4.
The next two chapters describe the procedures used to select the RDD and area sample of
households.
2-9
Table 2-2.
Projected Design Effects for the RDD Sample, by Interview Type and Study Area
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
National
Children
Low-Income
1.08
1.03
1.05
1.04
1.04
1.04
1.05
1.08
1.05
1.04
1.05
1.05
1.02
1.06
1.99
All
1.11
1.04
1.04
1.04
1.03
1.03
1.04
1.12
1.04
1.04
1.07
1.05
1.02
1.06
2.11
Other Adults
Low-Income
1.47
1.49
1.59
1.59
1.51
1.54
1.38
1.37
1.43
1.50
1.34
1.42
1.33
1.68
2.76
All
1.37
1.33
1.32
1.39
1.27
1.30
1.19
1.31
1.19
1.28
1.22
1.22
1.17
1.44
2.56
All Adults
Low-Income
1.52
1.43
1.73
1.73
1.51
1.42
1.32
1.46
1.73
1.60
1.55
1.53
1.32
1.50
3.34
All
1.58
1.52
1.82
1.88
1.61
1.47
1.30
1.58
1.72
1.72
1.58
1.60
1.37
1.60
3.14
Table 2-3.
Projected Effective Sample Sizes for the RDD Sample, by Interview Type and Study Area
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
National
Children
Low-Income
663
845
634
662
594
665
562
690
619
747
802
629
563
1,699
5,225
All
1,322
2,006
1,892
1,525
1,990
1,986
1,935
1,242
2,258
1,692
1,619
1,813
1,660
4,177
12,854
Other Adults
Low-Income All
207
508
235
703
135
472
137
411
165
652
209
732
247
995
172
426
113
587
150
500
179
491
160
581
203
730
518
1,546
1,025
3,643
2-10
All Adults
Low-Income All
593
1,207
734
1,720
440
1,236
459
1,012
498
1,546
629
1,804
635
2,216
583
1,080
405
1,530
532
1,198
599
1,269
504
1,422
564
1,647
1,588
3,729
3,960
11,337
3. RANDOM DIGIT DIAL HOUSEHOLD SAMPLING
This chapter describes the sample design and implementation for the RDD component of the
survey. The first section describes the sampling telephone numbers using the list-assisted sample
design for each study area and the balance of the nation. The second section describes the overall
plan for subsampling households. The third and fourth sections go into detail about the main
subsampling procedures, subsampling adult-only households and high-income households,
respectively. The fifth section describes changes to the sampling parameters made during the
data collection period and based on monitoring the progress of the earlier stages of the sample.
The final section of the chapter presents tables on the sample sizes achieved using the methods
discussed.
3.1
Sampling Telephone Numbers
The RDD sample for all three rounds of NSAF used a list-assisted approach to select the sample
of telephone households. These households were screened to identify low-income households
with children and other households of interest as described later in this chapter. Casady and
Lepkowski (1993) describe list-assisted sampling and a recent update to the application of this
method is Tucker, Lepkowski, and Piekarski (2002). In list-assisted sampling, the set of all
possible residential telephone numbers is divided into 100-banks. Each 100-bank contains the
100 telephone numbers with the same first eight digits (i.e., the identical area code, telephone
prefix, and first two of the last four digits of the telephone number). The frame consists of all
100-banks with at least one residential number listed in a published telephone directory. Any
household telephone number that is not in these 100-banks is not covered by the sample. A
simple random or a systematic sample of telephone numbers is selected from this frame.
List-assisted RDD sampling is now the standard sampling procedure for telephone surveys. The
key advantages of this method relative to the Mitofsky-Waksberg method (Waksberg 1978) is
that the sample is unclustered, and the full sample of telephone numbers can be released to
interviewers without the sequential impediment in the Mitofsky-Waksberg method. A
disadvantage of list-assisted RDD is a small amount of undercoverage due to excluding
households in 100-banks with no listed households. Studies of the undercoverage due to this
exclusion have shown that only a small percentage of households are excluded, and households
excluded from the frame are not very different from those that are included (Brick, et al. 1995;
Tucker et al. 2002). Together, these attributes indicate the undercoverage bias from excluding
100-banks with no listed households is small.
Once the sample was selected from the 100-banks, two steps were used to improve the working
residential rate and thereby reduce costs. The first procedure eliminated numbers listed only in
the yellow pages. The second procedure used a vendor to dial the household automatically to
eliminate many nonworking numbers. If a tritone signal was detected, then the telephone was
classified as a nonworking number and was never dialed by Westat interviewer. Both steps were
used in all three rounds of NSAF data collection. The procedures were enhanced by the vendor
3-1
over time, and a larger percentage of numbers was classified as nonresidential or business in
Round 3 compared to earlier rounds. When the vendor’s new procedures were introduced,
Westat examined the accuracy of the results by comparing the outcomes from the vendor’s
process with those from a recently conducted survey (the 2001 National Household Education
Survey). Based on this study, Westat obtains the results of the purging from the vendor and
specifies which cases will not be dialed so that less than 1 percent of the purged numbers are
residential.
The sample was selected all at once, when the newest quarterly update of the sampling frame
was available in December 2001. The sample was then assigned to waves for data collection as
described in section 3.5. Consideration was given to sampling for later releases from the next
quarterly frame update that was expected by March 2002. However, the entire sample was
selected at once because most of the sample was planned to be in process before a sample from
the new frame could be prepared. Furthermore, the differences in the frames by quarter are
typically very small. Because of the magnitude of the number of telephone numbers in the
sample, the purging of the sample of telephone numbers was done in five batches, each
containing about 100,000 telephone numbers. The dates the batches were purged were January
7–21, January 14–28, January 22–February 5, January 28–February 11, and February 4–18. The
order of the batches corresponded to the waves of releasing the numbers.
Westat interviewers dialed any telephone number not eliminated by the tritone and yellow page
purges to determine its working residential status. When a residential telephone number was
reached, the interviewer asked about the age composition and income of the household. These
questions were used to subsample households, as discussed in the next section.
3.2
Subsampling Households
As in Rounds 1 and 2, households were subsampled for the NSAF interview using subsampling
rates that depended on whether there were children in the household and whether the household
income was below 200 percent of the poverty threshold. Specifically, households with children
under 18 and incomes below 200 percent of the poverty threshold were subsampled at 100
percent (in other words, they were retained with certainty). Households with all members 65 and
older were not subsampled (none were retained). All other households were retained for the
interview with subsampling rates greater than zero and less than one. The rationale and
procedures for the subsampling are the same as in previous rounds. Essentially, the subsampling
rates were developed to obtain the desired sample sizes and effective sample sizes for the
targeted subgroups needed for analysis. The rationale is discussed in more detail in the 1997 and
1999 sample design reports.
Two distinct subsampling steps were implemented. The first subsampled households with no
children. The second subsampled households with and without children if the reported income
was above 200 percent of the poverty threshold or if the response to income screener question
was missing. Thus, four strata with different subsampling fractions were created, as illustrated in
figure 3-1. The two subsampling steps are discussed in more detail in the next sections.
3-2
Figure 3-1.
Household Subsampling Operations
Children
present?
Yes
Keep
No
Low
income?
Yes
Keep
No
Subsample
Anyone
under age
65?
No
Yes
Subsample
Yes
Low
income?
No
Drop
Subsample
3-3
Keep
The subsampling rates were computed based on a variety of assumptions. Most assumptions
were computed using data from Round 2. The key assumptions used to compute the subsampling
rates are as follows:
Type of household—Table 3-1 gives the proportion of child, adult, and elderly households
estimated from Round 2.
Low-income rates—Table 3-1 also gives the estimated proportion of households that are
poor (less than 200 percent of the poverty threshold), nonpoor, and unknown, for both
child and adult households. These proportions are the expected rates for households that
respond to the screener income item. The Round 2 rates were used for Round 3 despite the
recent economic downturn to be conservative.
Misclassification by income—The rates of households being misclassified by income are
shown in table 3-2. The false negative rate is the proportion of households that report high
income when they are actually low income. The false positive rate is the proportion of
households that report low income when they are actually high-income. The proportions in
the table are the observed (weighted) rates from Round 2 and are discussed more fully later
in the chapter.
Residential rates—The assumed residency rates are given in the first column of table 3-3.
The rates are three points less than the residency rates estimated from Round 2. Lower
rates were used because the residency rate has been decreasing rather steeply during this
time period. The ring–no answer (NA) and answering machine (NM) cases are treated as
nonresidential for this purpose since they do not result in completed interviews.
Response rates—The assumed response rates are shown in the remaining columns of
table 3-3. The screener response rates are 2 percentage points less than the estimated
response rate from Round 2, again without any allocation for NA and NM cases. We also
assumed the adult and child extended response rates would be 2 percentage points less than
the observed Round 2 rates. Once again, the lower rates were assumed because response
rates to most RDD surveys have been decreasing in recent years and the Round 2 rates
were some of the highest response rates we have seen in RDD surveys in the past three
years.
The rates for subsampling adult-only households and households by income were derived using
these assumptions. These subsampling or retention rates are shown in table 3-4. The required
number of completed number of screeners and sampled telephone numbers was then derived
from these assumptions.
3-4
Table 3-1.
Assumed Proportion of Households, by Household Type and Poverty Status
3-5
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Household Type
Child
Adult
0.366
0.421
0.389
0.418
0.376
0.462
0.324
0.410
0.337
0.457
0.367
0.427
0.361
0.441
0.373
0.397
0.366
0.434
0.350
0.437
0.397
0.415
0.368
0.459
0.351
0.434
0.360
0.432
0.363
0.430
Elderly
0.213
0.193
0.163
0.266
0.206
0.205
0.198
0.230
0.200
0.213
0.188
0.174
0.214
0.208
0.207
Child Households–Screening
Poor
Nonpoor
Unknown
0.339
0.617
0.045
0.306
0.628
0.066
0.233
0.725
0.041
0.311
0.631
0.057
0.203
0.735
0.062
0.239
0.706
0.054
0.202
0.761
0.037
0.390
0.557
0.053
0.183
0.763
0.054
0.309
0.632
0.059
0.348
0.601
0.051
0.241
0.712
0.047
0.253
0.703
0.044
0.271
0.684
0.044
0.279
0.671
0.050
Adult-Only Households–Screening
Poor
Nonpoor Unknown
0.227
0.696
0.077
0.171
0.751
0.078
0.136
0.810
0.055
0.167
0.762
0.071
0.123
0.801
0.077
0.151
0.788
0.061
0.130
0.825
0.045
0.235
0.701
0.064
0.100
0.839
0.061
0.159
0.774
0.067
0.179
0.752
0.069
0.139
0.807
0.055
0.148
0.796
0.057
0.172
0.769
0.060
0.165
0.771
0.064
Table 3-2.
Assumed Misclassification Rates, by Income Categories
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Child Households
False
False
Positive Negative
0.166
0.139
0.195
0.109
0.249
0.118
0.216
0.149
0.218
0.086
0.245
0.105
0.269
0.104
0.184
0.142
0.267
0.083
0.176
0.138
0.179
0.148
0.198
0.086
0.200
0.119
0.195
0.121
0.182
0.123
Unknown to
Low-Income
0.315
0.597
0.462
0.541
0.401
0.503
0.358
0.536
0.442
0.562
0.547
0.361
0.332
0.495
0.537
Adult-Only Households
False
False
Unknown to
Positive Negative Low-Income
0.321
0.142
0.385
0.358
0.129
0.385
0.362
0.112
0.427
0.380
0.128
0.455
0.354
0.073
0.325
0.360
0.106
0.275
0.373
0.098
0.401
0.322
0.120
0.510
0.473
0.085
0.329
0.332
0.118
0.205
0.298
0.158
0.521
0.282
0.086
0.243
0.320
0.110
0.286
0.327
0.104
0.279
0.337
0.111
0.332
Table 3-3.
Assumed Residential and Response Rates
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Residency Rate
(%)
45.6
39.5
38.5
39.9
43.4
39.1
39.6
45.9
40.0
41.9
37.4
39.0
40.9
40.1
40.1
Screener
Response Rate
(%)
80.3
69.6
75.1
72.3
69.3
74.1
80.3
80.3
65.7
66.8
74.6
74.7
79.5
76.7
74.3
3-6
Adult Extended
Response Rate
(%)
77.4
70.3
76.0
71.8
72.1
75.0
82.9
77.9
67.5
69.0
74.0
77.6
82.8
77.1
75.0
Child Extended
Response Rate
(%)
78.7
72.9
80.9
74.6
76.7
79.9
84.5
81.2
72.8
72.9
77.9
82.2
83.0
80.6
78.5
Table 3-4.
Subsampling or Household Retention Rates
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
3.3
Child Households
Above
Unknown
Poverty
Income
0.507
0.750
0.652
0.750
0.629
0.750
0.636
0.750
0.646
0.750
0.654
0.750
0.619
0.750
0.498
0.750
0.606
0.750
0.651
0.750
0.576
0.750
0.593
0.750
0.750
0.750
0.570
0.750
Adult-Only Households
Above
Unknown
Adult
Poverty
Income
0.643
0.255
0.397
0.803
0.247
0.471
0.611
0.224
0.424
0.607
0.220
0.432
0.708
0.236
0.429
0.859
0.243
0.406
0.877
0.295
0.435
0.579
0.288
0.400
0.534
0.264
0.429
0.600
0.257
0.388
0.550
0.312
0.402
0.584
0.280
0.402
0.777
0.327
0.408
0.838
0.206
0.419
Subsampling Adult-Only Households
Since one goal of the survey was to estimate characteristics of all adults under age 65 and of the
subset of these adults who lived in households without children, the same procedure for
subsampling adult-only households used in the previous rounds was also used in Round 3.
Together with the subsampling of persons within households with children, these subsampling
procedures provide a sufficient sample size for making reliable estimates for these adults. As in
previous rounds, the MKAs and their spouse/partners are called “Option A” adults. Other adults
are called “Option B” adults. “Option B” adults who live in households with children are called
“Option B stragglers.” Statistics about nonelderly adults are formed using data from the “Option
A” interviews about the MKAs and their spouses and using data from the “Option B” interviews
about all other adults under age 65.1 The group of all other adults under age 65 consists of
nonelderly adults in adult-only households and some nonelderly adults in households with
children. In households with children, an adult was eligible for sampling if the adult did not have
any children under age 18 living in the household and if the adult had not already been identified
as the MKA for a focal child or as the spouse/partner of an MKA.
1
“Option A” interviews were administered to the MKA about the focal child(ren). It also obtained income, earnings, health insurance, and other
information about the MKA and his/her spouse/partner. The “Option B” interview obtained the same information about the sample adult and
his/her spouse/partner as obtained by the “Option A” interview—only questions about children were missing from the “Option B” version of
the questionnaire.
3-7
The target sample sizes for adults were set for the combined group of “Option B” adults and
“Option B” stragglers. The targets for each study area were approximately equal to the observed
sample sizes in Round 2. Since the Round 3 sample had a much smaller sample of nontelephone
households in Round 3 compared with previous rounds, the target sample sizes were close to the
number of telephone interviews from Round 2.
The retention rates for adult-only households in the RDD sample are shown in table 3-4. The
retention rates ranged from 53 percent to 88 percent across the study areas. These rates are
generally higher than they were in earlier rounds in order to reduce the variability in the
estimates due to subsampling adults at differential rates. The research supporting this approach is
described in 1999 NSAF Sample Design, Report No. 2.
The subsampling was implemented by loading the retention rate table into the computer-assisted
telephone interviewing (CATI) system. Each telephone number was randomly assigned as either
a “child-only” household or a “child and adult” household using a random sampling procedure
before the case was loaded. Households subsampled as “child and adult” were classified as
“adult-only” households if there were no children under age 18 present but at least one person
was under the age 65.
3.4
Subsampling High-Income Households
The same procedure used in the previous rounds was used in Round 3 to subsample households
by income. In households with children and in subsampled adult-only households, all households
were asked a simple question on household income. Those households reporting income above
200 percent of the poverty threshold were subsampled. The income levels for determining 200
percent of poverty are shown in table 3-5. The subsampling rates are given in table 3-4 for both
types of households.
Table 3-5.
Income Levels for Determining Less than 200 Percent of the Poverty Level
Household Size by Person
1
2
3
4
5
6
7
8
9+
Without Children
$18,300
$23,500
$27,500
$36,200
$43,700
$50,200
$57,800
$64,600
$77,700
a. This type of household can occur only if an emancipated minor is living alone.
3-8
With Children
$18,300a
$24,200
$28,300
$35,800
$42,400
$47,900
$54,500
$60,500
$71,400
After asking whether there were children in the household, the interviewer asked whether
household income was above or below the poverty threshold given in table 3-5. The household
was then either retained or subsampled out, depending on the response. As in the previous
rounds, we expected some household responses to the simple income item to be incorrect
compared with what they would say in response to the income items in the extended interview.
This misclassification error introduces an additional design effect on low-income households
with children (the household subsampling is only done for the RDD sample of households). The
benefit of asking the question and taking the loss in the effective sample size is that it sharply
reduces costs for completing the extended interviews.
The switching rates for households that do and do not report low income on the screener are
functions of the false negative and false positive error rates. The false negative and false positive
rates that were observed in Round 2 and assumed to apply to Round 3 are given in table 3-2.
Report No. 2 in the 1999 series gives a full discussion of all these issues. The subsampling rates
that were designed to be close to the optimum values for Round 2 were fully implemented for
Round 3.
3.5
Household Sampling Revisions during Data Collection
As noted in section 3.2, the sample design depended on a number of assumptions. When data
collection began, the outcomes were monitored and used to project eventual sample yields. If the
revised projections differed substantially from the target, the sampling rates and number of
telephone numbers could be adjusted to be consistent with the new projections. The same
monitoring process was used in both Round 1 and Round 2. In both rounds the monitoring led to
revisions in the sampling procedures during data collection.
We assumed that changes in sampling parameters might be needed and designed the sampling to
facilitate making changes during the data collection period. The sample for each study area was
divided into a main sample and a reserve sample. Within the main sample cases were further
partitioned into release groups. The overall sample of telephone numbers was selected, and then
the telephone numbers were assigned randomly to the main and reserve samples for each study
area independently. Within the main sample, all telephone numbers in a study area were
randomly assigned to one of 103 release groups. Each release group had 5,000 telephone
numbers and was a microcosm of the main sample. The only exception is the last release group
(103) that was smaller because it contained the remainder of the sample. Release groups 1
through 60 were used for a variety of methodological experiments. To support the analysis of
these experiments, it was decided that no revisions in the sampling or other procedures would be
implemented in the first 60 release groups.
Relatively early in the data collection period we observed several deviations from the assumed
rates. The proportion of households with children that was low-income differed from the
assumed rate in several study areas. Similarly, the proportion of adult-only households by
income differed from the assumed rate. These differences resulted in deviations from the
expected number of interviews for specific types of households. The deviations from the
3-9
assumed rates were concentrated in specific study areas rather than being uniformly distributed
over the entire sample. To deal with the differences, the subsampling rates were revised in later
release groups to achieve sample yields that were closer to the targets.
Monitoring the data collection process also found that the assumed residency rates and response
rates for the screener were lower than expected. The consequence of these deviations from the
assumptions was the completion of fewer screener interviews than planned and consequently the
sampling of a smaller number of people for the extended interviews. The magnitude of the
deviations of the residency and response rates from the assumed rates varied by study area. The
losses in sample yield due to these lower rates were modest in most study areas, but in some
study areas the expected consequences on the sample yield were substantial and changes were
necessary. The obvious way to deal with the lower-than-expected residency and response rates
was to release reserve sample in selected study areas. However, releasing reserve sample would
have some disadvantages, such as increasing data collection costs, extending the data collection
period, and delaying the release of estimates from the survey.
To mitigate the cost and time implications associated with releasing more sample telephone
numbers, a subsampling approach was instituted. The plan was to subsample telephone numbers
that resulted in an initial refusal at the screener. A random subsample of telephone numbers with
screener refusals was instituted and only those subsampled were included in refusal conversion
efforts. The screener refusal cases that were not subsampled were excluded from subsequent data
collection efforts. The plan included a provision for weighting the subsampled refusal cases to
account for all cases that were eligible for subsampling for the preparation of estimates and
response rates. The subsampling was limited to screener refusal cases and did not apply to
extended interview work.
Data collection costs were reduced because expenses related to refusal cases could be avoided
for cases that were not subsampled for conversion. When a household refuses the screener, the
interviewer records demographic information about the refusing respondent and the respondent’s
reasons for refusing to participate. Interviewers also rate the strength of the refusal as mild, firm,
or hostile. In Round 3, up to two refusal conversion attempts were made for refusing households
at the screener level. In addition, Telephone Research Center (TRC) supervisors reviewed all
cases coded as hostile to determine whether that designation was merited. Cases rated as hostile
that supervisors judged inappropriately coded were recoded and then were eligible to be released
for an additional conversion attempt. Truly hostile (profane or abusive) refusal cases were never
released for conversion. Before attempting to convert the screener refusal, an express mailing
was made to the household if an address was available for that household, adding further to the
cost of refusal conversion.
The reduction in field time was realized by implementing the subsampling so that screeners
released later in the field period were not subsampled for refusal conversion, while those
released earlier were converted. The details on implementing the subsampling to achieve this
goal are given below. Thus, the refusal subsampling allowed the sample to be worked more
efficiently while still including all of the appropriate scheduling procedures (including hold
periods for refusal cases). The usual practice of continuing a smaller interviewer workforce to
3-10
cover the calls to late refusals is not needed to the same extent and the data collection period can
end more quickly.
Since the telephone numbers had been randomly assigned to the release groups prior to data
collection, these release groups were well suited for the subsampling approach. Release groups 1
to 82 and release group 103 were designated for subsampling (i.e., refusal conversions were
scheduled for all initial screener refusals in these groups). Telephone numbers in release groups
83 to 102 were designated as not subsampled. The same designation was used for all study areas
and for the balance of the nation.
The number of release groups included in the subsample was determined by evaluating the effect
on the sample yield and the increase in variance in the estimates due to weighting the refusal
conversion cases for the subsampling. The sample yield obtained by subsampling was
determined using the initial screener refusal rate observed in the early waves of the Round 3 data
collection. Since only about 20 percent of the full sample of telephone numbers was excluded by
the subsampling, the weighting factor on the retained screener initial refusal cases was not very
large. The increase in variance due to the weighting was expected to have a negligible effect
(less than 2 percent) on the variance of the estimates.
Since the revised data collection plan had two components (changes in the subsampling rates
dealing with screening by income and the introduction of refusal subsampling), the
implementation of the two components was coordinated. The goal was to use the same
subsampling rate by income for the cases irrespective of the subsampling for refusal conversion.
To accomplish this goal, the average rate of subsampling by income was computed for release
groups 61 through 82 and 103 (those subsampled for refusal conversion). This subsampling rate
was applied to telephone numbers in release groups 83 through 102 and any reserve sample.
The revised subsampling rates for households with children that were classified as high-income
in the screener are given in table 3-6. The first column shows the subsampling rate applied for
release groups 61 through 82 plus 103, and the second column shows the rate for release groups
83 through 102 plus any reserve sample. The bold entries are the only ones that changed from the
corresponding planned rates given in table 3-4. Table 3-7 gives the revised subsampling rates for
household with adults-only by the release groups. The bolded entries indicate revisions from the
rates applied in release groups 1 through 60.
Table 3-8 gives the number of reserve telephone numbers released in the study areas where the
reserve was needed. In total, the sample of telephone numbers increased by 42,411 numbers.
This represented an increase of about 8 percent over the originally planned sample size of
telephone numbers.
3-11
Table 3-6.
Revised Subsampling Rates for Households with Children Screening
as High-Income, by Release Group
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Release Groups
61–82 and 103
0.507
0.000
0.629
0.100
0.000
0.654
0.619
0.300
0.000
0.500
0.576
0.593
0.750
0.570
Release Groups
83–102
0.507
0.473
0.629
0.490
0.468
0.654
0.619
0.447
0.440
0.610
0.576
0.593
0.750
0.570
Table 3-7.
Revised Subsampling Rates for Adult-Only Households, by Release Group
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Subsampling Rate for Adult-Only
Households prior to Income Item
Release Groups Release Groups
61–82 and 103
83–102
0.643
0.643
0.803
0.803
0.611
0.611
0.400
0.550
0.500
0.650
0.400
0.733
0.877
0.877
0.579
0.579
0.100
0.415
0.600
0.600
0.550
0.550
0.420
0.539
0.415
0.676
0.350
0.704
3-12
Subsampling Rate for Adult-Only
High-Income Households
Release Groups Release Groups
61–82 and 103
83–102
0.255
0.255
0.247
0.247
0.224
0.224
0.220
0.220
0.236
0.236
0.459
0.302
0.295
0.295
0.288
0.288
0.500
0.329
0.257
0.257
0.020
0.231
0.280
0.280
0.315
0.323
0.400
0.260
Table 3-8.
Reserve Sample Released, by Study Area
Study Area
Total
Colorado
Florida
Massachusetts
Michigan
Minnesota
New Jersey
Wisconsin
Bal. of nation
3.6
Number Released
42,411
6,700
2,952
2,322
4,589
7,498
6,324
6,608
5,418
Achieved Response and Eligibility Rates
The RDD household screening component for Round 3 is summarized in table 3-9. The sample
of 556,651 telephone numbers includes the 42,411 telephone numbers released from the reserve
sample. The table shows the numbers for each study area and the total across all the study areas
and the balance of the nation.
Overall about 54 percent of the sample phone numbers were determined to be nonworking or
nonresidential, and another 9 percent were never resolved. The never resolved numbers are those
that had only ring–no answer or answering machine responses, despite repeated attempts to reach
someone at the number. The survival method (see Brick et al. 2002 or Report No. 2 for 1999)
was used to estimate the residential status for the never resolved numbers. About 32 percent of
these numbers were estimated as residential and the rest were not residential. The percentage
varies by study area. Using the survival method allocation approach for the never-resolved
numbers, the estimated residency rate for the nation is 39 percent (this is a weighted number that
accounts for the population of the study area and is discussed in more detail in Report No. 4).
The last column of the table gives the number of households that responded to the questionnaire
item on the presence of children in the household.
Since screeners that were initially refused were subsampled, the screener response rate cannot be
computed by dividing the number of completed screeners by the number of residential numbers.
The response rate must account for the subsampling. The details of the computations and other
discussion of response rates are given in Report No. 7, but essentially the completed screeners
that were subsampled were also weighted by the inverse of the subsampling rate in the
computations. Table 3-10 shows the number of sampled telephone numbers that were initial
refusal screener cases and the number of these that were subsampled. As can be seen, about 20 to
25 percent of the initial refusal cases were sampled in each study area. The last column of the
table shows the estimated screener response rate. This response rate is based on weighting cases
by the inverse of their probability of selection, including the refusal subsampling probability.
3-13
Table 3-9.
Screening for Residential Status and Presence of Children
Verified Nonworking
No Contact
or Nonresidentiala
(after 14
Released
Study Area
Sample Number Rate (%) Calls)
Alabama
22,608
11,766
52.0
1,165
California
37,554
19,610
52.2
3,249
Colorado
41,180
23,840
57.9
3,077
Florida
36,637
19,665
53.7
2,629
Massachusetts
42,142
20,666
49.0
3,203
Michigan
40,858
23,159
56.7
2,816
Minnesota
41,035
24,792
60.4
2,015
Mississippi
19,325
9,591
49.6
1,022
New Jersey
60,479
29,038
48.0
7,145
New York
35,108
19,044
54.2
2,689
Texas
29,653
17,330
58.4
1,926
Washington
34,137
19,699
57.7
2,129
Wisconsin
30,927
16,670
53.9
1,977
Bal. of nation
85,008
47,153
55.5
5,395
Total
556,651 302,023
54.3
40,437
Only
Answering
Machine
Contacts
341
682
649
802
758
623
444
206
1,241
590
369
473
357
1,289
8,824
Known and
Imputed
Known
b
Residential Residential
9,336
9,934
14,009
15,314
13,614
14,534
13,537
14,697
17,512
18,706
14,258
15,298
13,783
14,702
8,506
8,903
23,051
25,979
12,782
13,728
10,028
10,715
11,835
12,732
11,923
12,712
31,168
33,378
205,342
221,333
Residency Complete Age
c
Rate (%) Screening
43.9
6,400
40.8
8,901
35.3
9,008
40.1
8,364
44.4
10,465
37.4
9,534
35.8
9,747
46.1
5,947
43.0
13,136
39.1
7,518
36.1
6,476
37.3
8,329
41.1
8,563
39.3
21,115
39.4
133,503
a. These include those telephone numbers that were classified as nonresidential on the basis of the automated tritone detection and White Page matching, as well as those verified by
interviewers.
b. Survival analysis was used to allocate phones that were never answered and those where the only contact was with an answering machine by residency status.
c. The percent of released sample that was either verified or imputed to be working phone numbers for private residences.
3-14
Table 3-10.
Subsampling Screener Refusals and Response Rates
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Known
Residential
9,336
14,009
13,614
13,537
17,512
14,258
13,783
8,506
23,051
12,782
10,028
11,835
11,923
31,168
205,342
Initial Refusals
Not
Subsampled Subsampled
3,533
798
5,382
1,303
4,376
2,032
5,258
1,631
7,075
2,173
4,707
1,752
3,780
1,835
3,144
694
8,799
3,251
5,342
1,254
4,265
909
4,291
949
3,043
1,592
10,692
3,279
73,687
23,452
Subsampled
(%)
81.6
80.5
68.3
76.3
76.5
72.9
67.3
81.9
73.0
81.0
82.4
81.9
65.7
76.5
75.9
Screener
Response Rate
(%)
68.2
61.9
68.8
61.6
60.8
67.8
72.8
70.6
55.5
58.3
64.3
69.1
73.9
67.9
65.4
Note: Response rates are weighted and account for refusal subsampling and use the survival allocation method for handling unknown residency
numbers.
Table 3-11 shows the results of screening households by age. The screening was used to classify
households into three categories:
Households with children under 18;
Households with no children but at least one person under the age of 65; and
Households where everyone was 65 or older.
The first category included the very small number of households that had no adults but did have
emancipated minors (persons under 18). The largest overall sampling rates were applied to the
households with children, consequently the sample size requirements for this group determined
the size of the sample needed for screening. Overall, 37 percent of the screened sample of
households had children. All households with children were retained for income screening.
Adult-only households, those with someone under the age of 65 but with no children, constituted
44 percent of the sampled households. As described in section 3.3, adult-only households were
subsampled prior to income screening. The average retention rate for adult-only households at
this stage was 63 percent. Households with only elderly members were 20 percent of screened
households. Since these households were out of the scope of the survey they were immediately
eliminated as ineligible.
3-15
Table 3-12 shows the results of the next stage of screening, income screening, for telephone
households with children. Of households with children, 95 percent answered the income
screening question. Of those households that did answer, 28 percent reported low income
(income 200 percent of the poverty threshold or less) and 68 percent reported income above 200
percent of the poverty threshold. All the low-income households were retained, and 56 percent of
the others with known income were retained. For the households that did not respond to the
income items, 69 percent were retained.
Table 3-13 shows comparable information for adult-only telephone households retained for
income screening. The income question was answered by 94 percent of the households. Among
those that answered, the low-income rate was 16 percent and all of these low-income households
were retained. About 25 percent of those households who responded but did not have low
income were retained. Finally, 35 percent of those with unknown income were retained.
Combining the subsampling of households with children and households with adults only who
were retained for income screening, 42 percent of all households that screened with income
above 200 percent of the poverty threshold were retained for extended interviews.
3-16
Table 3-11.
Outcomes of Household Screening of Telephone Households
3-17
Households with Children
Screened
Number of Households
Households with
Children
Total Age with
(%)
Study Area Screeners Children
Alabama
6,400
2,244
35.1
California
8,901
3,702
41.6
Colorado
9,008
3,367
37.4
Florida
8,364
2,738
32.7
Massachusetts 10,465
3,670
35.1
Michigan
9,534
3,452
36.2
Minnesota
9,747
3,523
36.1
Mississippi
5,947
2,201
37.0
New Jersey
13,136
4,922
37.5
New York
7,518
2,705
36.0
Texas
6,476
2,654
41.0
Washington
8,329
3,079
37.0
Wisconsin
8,563
2,942
34.4
Bal. of nation 21,115
7,482
35.4
Total
133,503
48,681
36.5
Adult-Only Households
65 and Over Households
Screened
Adult-Only Retention
Households
Households Rate of
Number of Screened
Adult-Only Households Retained for Adult-Only Number of with Only
Households 65 and Over Adults Over
(under 65) with Adults Income
Households 65 (%)
Screening (%)
Households Only (%)
2,842
44.4
1,773
62.4
1,314
20.5
3,775
42.4
2,968
78.6
1,424
16.0
4,209
46.7
2,508
59.6
1,432
15.9
3,564
42.6
1,871
52.5
2,062
24.7
4,684
44.8
3,021
64.5
2,111
20.2
4,110
43.1
3,010
73.2
1,972
20.7
4,298
44.1
3,737
86.9
1,926
19.8
2,472
41.6
1,380
55.8
1,274
21.4
5,726
43.6
2,265
39.6
2,488
18.9
3,335
44.4
1,902
57.0
1,478
19.7
2,723
42.0
1,453
53.4
1,099
17.0
3,782
45.4
2,036
53.8
1,468
17.6
3,865
45.1
2,608
67.5
1,756
20.5
9,445
44.7
6,498
68.8
4,188
19.8
58,830
44.1
37,030
62.9
25,992
19.5
Table 3-12.
Outcomes of Income Screening of Telephone Households with Children
Not Low-Income Households
3-18
Households
with
Children
Study Area
Alabama
2,244
California
3,702
Colorado
3,367
Florida
2,738
Massachusetts
3,670
Michigan
3,452
Minnesota
3,523
Mississippi
2,201
New Jersey
4,922
New York
2,705
Texas
2,654
Washington
3,079
Wisconsin
2,942
Bal. of nation
7,482
Total
48,681
a. According to the simple screener question.
Complete
Income
Screening
2,160
3,496
3,226
2,596
3,471
3,297
3,383
2,094
4,645
2,534
2,510
2,964
2,824
7,132
46,332
Low-Income
Householdsa Identified Selected
836
1,324
703
1,247
2,249
1,042
784
2,442
1,566
913
1,683
817
731
2,740
1,250
876
2,421
1,587
735
2,648
1,654
909
1,185
545
983
3,662
1,622
789
1,745
1,096
922
1,588
918
812
2,152
1,272
716
2,108
1,569
2,145
4,987
2,857
13,398
32,934
18,498
Households with Unknown
Income
Selection
Rate (%) Identified Selected
53.1
84
56
46.3
206
145
64.1
141
103
48.5
142
96
45.6
199
138
65.6
155
110
62.5
140
99
46.0
107
69
44.3
277
195
62.8
171
113
57.8
144
97
59.1
115
70
74.4
118
84
57.3
350
237
56.2
2,349
1,612
Selection
Rate (%)
66.7
70.4
73.0
67.6
69.3
71.0
70.7
64.5
70.4
66.1
67.4
60.9
71.2
67.7
68.6
Total
Sample
Households
with
Children
1,595
2,434
2,453
1,826
2,119
2,573
2,488
1,523
2,800
1,998
1,937
2,154
2,369
5,239
33,508
Table 3-13.
Outcomes of Income Screening of Adult-Only Telephone Households
3-19
Low-Income
Households with Unknown
Complete Households Not Low-income Households
Income
Retained
for Income Income
Identified
Selection
Selection
Study Area Screening Screening and Selected Identified Selected Rate (%) Identified Selected Rate (%)
Alabama
1,773
1,642
415
1,227
331
27.0
131
49
37.4
California
2,968
2,784
478
2,306
547
23.7
184
53
28.8
Colorado
2,508
2,362
352
2,010
444
22.1
146
50
34.2
Florida
1,871
1,736
351
1,385
282
20.4
135
48
35.6
Massachuset 3,021
2,820
384
2,436
565
23.2
201
75
37.3
Michigan
3,010
2,831
449
2,382
650
27.3
179
71
39.7
Minnesota
3,737
3,558
514
3,044
910
29.9
179
62
34.6
Mississippi
1,380
1,285
382
903
251
27.8
95
33
34.7
New Jersey
2,265
2,108
253
1,855
547
29.5
157
54
34.4
New York
1,902
1,753
313
1,440
346
24.0
149
51
34.2
Texas
1,453
1,338
271
1,067
223
20.9
115
40
34.8
Washington
2,036
1,931
302
1,629
454
27.9
105
31
29.5
Wisconsin
2,608
2,465
364
2,101
638
30.4
143
47
32.9
Bal. of
6,498
6,092
1,156
4,936
1,095
22.2
406
143
35.2
Total
37,030
34,705
5,984
28,721
7,283
25.4
2,325
807
34.7
a. According to the simple screener question.
Total
Sample
Adult-Only
Households
795
1,078
846
681
1,024
1,170
1,486
666
854
710
534
787
1,049
2,394
14,074
4. AREA SAMPLE
This chapter describes the in-person or area sample that is designed to yield a national sample of
households without current telephone service. Unlike in 1997 and 1999, only a national area
sample was selected for 2002. Since the design of the area sample for 2002 was a modification of
the sample design used in 1999 (which was a modification of the 1997 design), the sections that
follow briefly explain key features of the earlier design as an introduction to the 2002 design.
More details on the sample design for the area sample for 1997 and 1999 are in Report No. 2 for
each of those years.
The first section of this chapter covers selecting the primary sampling units (PSUs) for 2002 by
subsampling PSUs from the 1999 sample. The second section describes with the sampling of
segments. The third section reviews the subsampling segments using “chunks,” which are
compact subsets of segments. The last section summarizes the outcomes of these sampling
procedures as they were implemented in the field.
The sampling procedures are those typically used in area probability samples, with two
exceptions. The first special procedure was the elimination from the sampling frame of blocks in
census block groups (BGs) with very high telephone service rates. This was done to lower the
cost of data collection even though it was realized that this would leave a certain group of
households unrepresented. The second special procedure concerned sampling of dwelling units
(DUs) within segments. Traditionally, a constant expected number of DUs is sampled from all
sample segments after a preliminary listing operation. In this survey, listing and screening were
carried out simultaneously so that the number of DUs varied by segment. Segments with a large
number of expected DUs were thus “chunked” so that only a portion of the selected segment was
listed and prescreened. Briefly, “chunking” involves obtaining a rough count of DUs within a
segment dividing the segment into “chunks” of roughly the same number of DUs. One “chunk”
is then selected for listing and prescreening. These two special procedures are described in more
detail in the sections that follow.
4.1
First-Stage Sampling
When the sample of PSUs were selected originally, the process was categorized into four distinct
phases. The first was to define the PSUs in terms of geographic units. Starting with a standard set
of PSU definitions used by Westat as the sampling frame for many surveys, one modification
was made to optimize the definitions for NSAF. The second phase was to decide how many
PSUs to select in each targeted state and in the balance of the nation. The third phase was to
stratify the PSUs to reduce between-PSU variance as much as possible for statistics of interest.
The fourth phase was to actually draw the sample PSUs from the strata. These phases are each
described here in terms of the original sampling and the subsampling for 2002.
4-1
PSU Definition. Since the 2002 sample of PSUs is a subsample of those sampled in 1999, the
definitions of the PSUs were identical in all three rounds. This section summarizes those
definitions. The standard Westat PSUs were formed in 1991. These PSUs were defined to follow
several rules. Each metropolitan statistical area (MSA) defined by the Census Bureau in 1990
was generally defined as a separate PSU. This procedure has the effect of minimizing betweenPSU variance while adding only modestly to within-PSU travel costs. The between-PSU
variance was reduced compared to a plan that established separate PSUs for the central cities and
the suburbs, because heterogeneity is maximized within PSUs rather than across them. The
within-PSU travel cost was not much higher despite leaving the central cities and suburbs
together because of the generally high-quality transportation networks in metropolitan areas.
Number of Sample PSUs. The number of sample PSUs was reduced from the previous rounds
because the 2002 area sample is a national sample only rather than also providing study area
estimates. In consultation with the Urban Institute, the total number of PSUs was reduced so that
only approximately half the total number of PSUs from 1999 were retained. To increase the
efficiency of the sample and reduce the variability in the sampling rates for households without
telephones, all the PSUs in the balance of the nation were retained (37), and the number of
sampled PSUs in the study areas was reduced.
The allocation of the sample was determined by first setting the total number of PSUs for the
sample using by the proportion of the number of 1999 eligible nontelephone households in the
balance of the nation as a guide. The estimated number of 1999 eligible nontelephone
households by study area is shown in column one of table 4-1. Because the balance of the nation
had 56 percent of the nontelephone population in 1999, a total of 66 PSUs (37/0.56) was chosen
so that the number of PSUs could be allocated roughly proportional to size by study area. Thus,
29 PSUs were to be subsampled from the 96 PSUs for 1999 in the study areas.
The next step was allocating these additional 29 PSUs to the study areas proportional to the
percent of 1999 eligible nontelephone households. This allocation is shown in column three.
Because some sites such as Milwaukee and Massachusetts have such small nontelephone
populations, the proportional allocation did not assign a full PSU to these study areas. To have
the study areas as strata, at least one PSU had to be sample from each of the study area. The
number of PSUs shown in column four uses the proportional allocation but requires at least one
PSU in each study area. This revision increased the number of sampled PSUs from 66 to 68. The
number of PSUs sampled in Round 2 is shown in column five for comparison.
Before we selected the subsample of PSUs using this allocation, we investigated the effects the
allocation would have on the variability in the weights for the nontelephone households by
examining the average segment weights. The distribution of average segment weights within
each study area indicated a great deal of variability in the weights, with the main problems
arising because of the supplemental sample of segments selected in Round 1 for six study areas
(Colorado, Florida, Massachusetts, New Jersey, New York, and Washington). In particular, the
average segment weight in Florida and New York was much higher than the average weight. To
4-2
Table 4-1.
Number of Round 3 Primary Sampling Units
Study Area
Alabama
Bal. Wisconsina
California
Colorado
Florida
Mass.
Michigan
Milwaukeea
Minnesota
Miss.
New Jersey
New York
Texas
Bal. of nation
Washington
Total U.S.
Focal Sites
1999 Eligible Percent Proportion
Nontelephone by Study Allocation of
Households
PSUs
Area
60,879
2.9
1.9
19,105
0.9
0.6
199,315
9.6
6.3
15,168
0.7
0.5
70,550
3.4
2.2
10,702
0.5
0.3
57,664
2.8
1.8
7,831
0.4
0.2
15,182
0.7
0.5
54,050
2.6
1.7
30,663
1.5
1.0
66,053
3.2
2.1
265,790
12.8
8.4
1,167,908
56.3
37.0
35,028
1.7
1.1
2,075,889
100.0
66
907,980
43.7
29
Number of
Round 3
PSUs
2
1
6
1
2
1
2
1
1
2
1
2
8
37
1
68
31
Number of
Round 2
PSUs
10
8
6
6
8
4
6
1
7
12
6
7
9
37
6
133
96
Final
Allocation
2
1
6
1
3
1
2
1
1
2
1
3
6
37
1
68
31
a. Balance of Wisconsin and Milwaukee are shown separately but are combined into one site in the 2002 designs.
reduce the variability in the weights, the number of allocated PSUs in Round 3 was revised.
Since Texas had a low average segment weight and a relatively large number of allocated PSUs,
the allocation for Texas was decreased by two PSUs and the allocations for Florida and New
York were each increased by one. The final number of PSUs allocated to each study area is
shown in column six.
Stratification. Since the 2002 sample was a subsample of the Round 2 PSUs, the 1999 PSU
stratification was maintained. The following is a short description of the stratification procedure.
The general idea of stratification is to group similar PSUs into the same stratum and then select
just one or two PSUs per stratum. This procedure reduces between-PSU variance for most
statistics, particularly those related to the statistics used in the grouping. If enough information
related to the outcomes is available for stratification purposes, having a large number of strata
can increase the precision of point estimates. However, forming the maximum possible number
of strata makes estimation of variances more difficult. Given the small number of PSUs for each
state in the NSAF, the original sample design called for creating the maximum possible number
of strata, and then sampling one PSU per stratum. For efficiency, strata are formed so that they
are of nearly equal size in terms of population, whenever possible. When some PSUs are larger
than the average population per desired stratum, these PSUs are selected with certainty and
4-3
called self-representing (SR), or certainty, PSUs. These procedures were carried out for the
original stratification of PSUs in Round 1.
In Round 3, the subsampling of PSUs from the study areas reduced the effectiveness of most of
the stratification from the earlier rounds. Two notable exceptions are for the entire balance of the
nation sample of PSUs and for California. In these two study areas, the entire sample of PSUs
was retained and thus all the previous stratification was carried forward into Round 3. In the
other study areas, the study area was considered a separate stratum and a small subsample of
PSUs was selected. Thus, much of the stratification of the PSUs within the study area used in the
previous rounds was of little utility in these study areas.
Selecting PSUs. When the original sample of PSUs was selected, a probability proportional to
size (PPS) (population in the eligible BGs in the strata) selection method was used. Since the
goal of the subsampling was to retain this proportionality, the subsample of PSUs was selected
with equal probability from all the originally sampled PSUs in the study area.
Prior to subsampling, the PSUs were sorted in order of PSU weight (the inverse of PSU
probability of selection within each stratum). The sort order was alternating (ascending,
descending) across strata or study areas. The sort was used to control the variability in the sizes
of the subsampled PSUs, thereby reducing the chance of obtaining a subsample that had
predominately large- or small-sample PSUs. The last step was to select the PSUs with equal
probability within the study area stratum.
4.2
Second-Stage Sampling
The second-stage was the sampling of segments from the subsampled PSUs. Since all the
sampled segments were retained in the subsampled PSUs, only a short description of the 1999
second-stage sampling is given here. One exception was the sampling of segments in
Milwaukee, and this procedure is described in detail later in this section.
As mentioned previously, two special procedures were used in the sampling of segments for this
survey. The first concerned the exclusion of areas with very high rates of telephone coverage.
This exclusion had little effect on segment formation and selection rules. It is discussed in
section 4.2.1. The second special procedure was to screen compact chunks without listing. This
feature had a major effect on segment formation and selection rules. In surveys where there will
be subsampling within each segment to yield a uniform household sample size for all segments,
it is fairly straightforward to decide on a minimum size for each segment and on how many
segments to select from each sample PSU. For NSAF, natural variation in block size made it
advantageous to allow the minimum segment size and the number of sample segments to vary by
PSU while keeping the number of sample households nearly uniform across PSUs. Section 4.2.2
focuses on how the segments were defined and on the determination of the number of segments
to select in each PSU. The selection procedures are described in 4.2.3.
4-4
4.2.1
Exclusion of Block Groups with High Telephone Coverage Rates
Extending coverage to every nontelephone household was judged as an expensive undertaking
with uncertain benefits. The 1990 Decennial Census showed that there are BGs where virtually
every household had a telephone. The details from the 2000 Decennial Census were not available
at the time of sampling for 2002, but nontelephone households are even more rare using these
data. The prospect of having to screen hundreds or even thousands of households in some areas
to find a single nontelephone household seemed to be carrying the concept of coverage for every
nontelephone household to an unreasonable extreme. Accordingly, the decision was made to
restrict the area sample to BGs where the percentage of households with a phone in 1990 was
below 92 to 98 percent, with the exact limit varying from state to state. Table 4-2 shows the
cutoffs by state. In addition to the exclusion of blocks in BGs with high telephone coverage, all
blocks with zero year-round housing as of 1990 were excluded.
Table 4-2.
Maximum Telephone Service Rates Allowed in Covered Block Groups
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
4.2.2
Maximum
Telephone Service
Rate Allowed in
Covered BGs (%)
95
98
97
97
98
97
98
92
98
97
95
98
98
97
Nontelephone
Households
Excluded (%)
7.3
7.3
8.8
9.1
9.4
9.8
9.1
9.7
5.6
7.5
7.7
6.1
9.2
8.0
BGs Excluded
(%)
37.4
59.2
56.6
48.0
70.5
56.1
57.6
34.5
68.2
58.7
42.2
53.3
56.7
54.9
All Households
Excluded (%)
40.9
59.0
57.9
54.6
70.4
59.9
60.7
35.9
66.8
58.7
45.3
53.7
59.5
57.1
Segment Stratification and Selection
The segments were stratified by size in the original sampling procedure. Size was defined as the
ratio of the number of year-round housing units in the segment to the desired chunk size, c, for
the PSU. The two cutpoints in the stratification were 0.75 and 10.0. This means that a low
stratum was established for segments with DU counts more than 25 percent below the desired
4-5
chunk size, a high stratum was established for segments with 10 or more chunks, and a middle
stratum was left for all the remaining segments in the PSU.
Within each stratum, the segments were sorted by the segment-level telephone coverage rate.
(This was computed as the average telephone rate in 1990 for census BGs intersecting the
segment.) The sort order was reversed in the middle stratum so that small segments with high
nontelephone rates were close to medium segments with high nontelephone rates and so that
medium segments with low nontelephone rates were close to large segments with low
nontelephone rates. A systematic PPS sample was then drawn where the measure of size for a
segment was its assigned chunk count. Sampling was independent across PSUs, both NSR and
SR.
Segments were sampled PPS to the number of chunks assigned to the segment. When a chunk
was subsampled within a segment, the chunk was selected PPS to the number of housing units in
the chunk (based on the quick count) in Round 3. The chunks were created of equal size
whenever feasible. In Round 3, chunks were sampled with equal probabilities.
As mentioned earlier, the segments in the subsampled PSUs were retained for the 2002 sample
except for Milwaukee. The Milwaukee and balance-of-Wisconsin study areas were combined
into a single study area (Wisconsin). Since Milwaukee was the only nonstate area in the original
sample, it had been divided into segments (without the intermediate sampling of PSUs) and a
sample of 50 segments had been selected. In the average sampled PSU outside of Milwaukee,
PSUs were sampled and then about eight segments per PSU were selected. With the new unified
Wisconsin study area and the national design in Round 3, it was neither practical nor necessary
to retain all 50 segments for Milwaukee. A subsample of 20 segments of the 50 segments
sampled from Milwaukee was selected and retained for Round 3. To control the sample of
nontelephone households, the 50 segments were partitioned into two strata, based on the percent
of nontelephone households found in each segment in 1999. The two strata represent a high
(more than 4 percent nontelephone households) and a low density of having nontelephone
households. An equal probability sample was drawn from each stratum. A subsample of 7
segments was selected from the low eligibility stratum and 13 from the high eligibility stratum.
4.3
Chunk Selection
The number of chunks associated with each sampled segment was the same as the number used
in 1999. The method of associating the number to the segment is described in section 4.2.2 of
Report No. 2 for 1999. The only difference is that in previous rounds the number depended on
information from the Decennial Census, while in Round 3 the number was a function of the
chunk selection in Round 2. Essentially, if the interviewer’s initial count of the number of DUs
in the segment fell outside a range based on Round 2 listing, special instructions were required.
If the initial count for the segment contained less than 60 DUs, the entire segment was listed. If
the initial count for the segment contained more than 60 DUs, the interviewer would call the
home office. If the segment had been chunked in Round 2 and the count was within 20 percent of
the 1999 count, the same chunking instructions were used as in 1999. If the count was not within
4-6
20 percent, the segment was rechunked (divided into chunks based on the new counts) and a new
chunk was selected.
A few segments were identified as single-chunk segments but were discovered, during cruising,
to have grown considerably since 1999 due to new construction. Rather than screen many more
DUs than planned (most of which would be fairly new and would thus likely have phone
service), the decision was made to chunk some of these. Also, there were segments where
chunking was planned, but growth made it advisable to form more chunks than originally
planned. Lastly, in a few segments so much housing had been demolished that the planned
chunking was not carried out. Table 4-3 provides some information on planned and unplanned
chunking.
Table 4-3.
Segment Counts, by Planned and Unplanned Chunking (Including Sample Supplement)
Chunking
Chunking Expected
Done As but Not
Expected Done
Study Area
Alabama
1
0
California
24
2
Colorado
2
1
Florida
7
0
Massachusetts
0
1
Michigan
5
0
Minnesota
1
0
Mississippi
2
0
New Jersey
13
1
New York
33
0
Texas
14
0
Washington
7
1
Wisconsin
9
0
Bal. of nation
64
5
Total
182
11
4.4
More
Chunking
Done than
Expected
0
5
2
1
0
1
0
2
0
3
0
4
1
10
29
Less
Chunking
Done than
Expected
1
5
4
1
1
0
0
0
0
1
0
0
0
1
14
Chunking
Not
Expected
but Done
Anyway
1
1
0
3
0
1
0
2
0
1
1
1
0
4
15
Chunking
Not
Expected
and Not
Done
15
29
6
21
7
11
6
7
10
7
35
7
19
151
331
Total
Segments
18
66
15
33
9
18
7
13
24
45
50
20
29
235
582
Achieved Response and Eligibility Rates
The results of the area listing operation are shown in table 4-4. With only a national sample and
no reliable site-level estimates, the tables shown in this chapter in the previous reports at the site
level are only shown nationally for Round 3. Over 18,000 households were listed and 89 percent
were occupied. This is slightly less than half the number of households listed in Round 2.
4-7
Table 4-4.
Outcomes of Area Listing
Category
Listed Households
Occupied Households
Rate
Total, U.S.
18,299
16,305
89.1%
The next step in the survey was to prescreen and screen the households to determine if they were
eligible (i.e., there was no phone in the household and at least one occupant was under age 65).
The prescreening attempted to determine eligibility, and the screening was the initial screener
interview. If the respondent to the prescreener reported that there was no phone and at least one
occupant under age 65, then the interviewer immediately segued into the cell phone procedures
so that the screener and extended interviews could be conducted by telephone interviewers at the
TRC in Maryland. However, at some DUs, the information on the presence of a telephone may
have come from a teenager who was not eligible to make the cell phone call to the TRC. In
addition, there were instances where a qualified respondent was too busy to participate. In such
cases, the interviewer returned later for the screener interview. In most DUs, there was no time
lapse between prescreening and screening.
The results of the prescreening and screening operations are shown in table 4-5. Response on the
prescreener level (phone ownership and age only) was very high, at 99 percent. At the screener
level, it was 84.5 percent. The response rate for the screener in Round 2 was 80 percent. The
overall eligibility rate was 4.5 percent, which is slightly lower than the 4.8 percent eligibility rate
in Round 2. This means that roughly 22 households had to be contacted to find a single
household without a phone (and with at least one adult under 65).
Table 4-5.
Outcomes of Area Prescreening and Screening
Category
Complete
Response Rate
Total, U.S. Prescreening
16,109
98.8%
Screening
616
84.5%
Eligible
Eligibility Rate
Households Eligible for WithinHousehold Subsampling
730a
4.5%
721b
4.4%
NA
607
a. No telephone in household and at least one occupant under age 65.
b. Nine households switched from reporting no telephone and at least one occupant under age 65 at the prescreener to having either a phone or
only elderly residents at the screener.
4-8
5. WITHIN-HOUSEHOLD SAMPLING AND ACHIEVED SAMPLE SIZES
In both the RDD and area components, a sample of people living in the sampled households was
selected to reduce the burden of participation for the household and to improve response rates.
Different methods were used to sample children, adults in households with children, and adults
in households without children. These methods were the same as those used in Round 1 and
Round 2. The first three sections of this chapter describe the sampling methods. The fourth
section summarizes the outcomes of sampling people within households.
5.1
Sampling Children
After selecting the households, children under 18 were sampled from the selected households for
the child sample. If a sampled household had exactly one child, then that child was always
selected. In households with more than one child under 18, either one child or two children were
selected. One child was randomly selected from all the children 5 years old or younger.
Similarly, one child was randomly sampled from all children between the ages of 6 and 17 years.
Thus, in households with both children under 6 years old and children 6 and older, two children
were selected. The procedures were the same for households sampled from the RDD and the area
components.
The interviewer asked the MKA about each sample child. The MKA was most often the mother
or father of the child, but people with other relationships to the child were also interviewed if
they were most knowledgeable about a sample child.2 During the interview about the child,
questions were also asked about the MKA and his/her spouse/partner if the spouse/partner also
lived within the household. The MKA responded by proxy for all items about the spouse/partner.
However, some questions were asked only about the MKA and others were randomly asked
about only the MKA or the spouse/partner when both were present. The MKA or the
spouse/partner was randomly chosen as the subject for the items with equal probability. The
strategy was applied uniformly in the RDD and area components. As mentioned earlier, these
rules were the same as those used in Rounds 1 and 2.
5.2
Sample Selection of Other Adults in Households with Children
If the household was subsampled for adults as discussed in chapter 3, then a sample of adults
under age 65 (other than the MKA and the MKA’s spouse/partner) was selected from households
with children. The telephone numbers designated as being “in sample” if they were adult-only
households were also designated for sampling other adults. The procedure was used if the
household contained both children and other adults. Adults in the other-adult category include
2
If a person under the age of 18 was identified as a focal child, but did not have an MKA or the MKA was the spouse or unmarried partner of this
person, then this individual was considered an emancipated minor and regarded as an adult in the interview. There were 26 such interviews in
Round 3.
5-1
adult siblings, grandparents, aunts, uncles, and other relatives of sample children, boarders, and
live-in servants. The purpose of interviewing some of these adults was to ensure their
representation in estimates about the entire group of all adults under the age of 65.
The sample selection of other adults was performed at the close of the MKA interview so that
relationship data were available on all the adults in the household. To be eligible, the adults had
to be under age 65 and not be the parent of a child under age 18 living in the household. Ideally,
the eligibility rule would have included only adults who were not potential MKAs or potential
MKA spouse/partners for nonsample children under age 18 living in the household. However,
determining this would have involved asking detailed questions for each child in the household
about who was the MKA for the child and which adults in the household might be
spouse/partners of the potential MKAs. Such an approach would have been too burdensome and
any small biases that were induced by the procedure used were deemed acceptable. The biases
involved multiple chances of selection for nonparent MKAs and no chance of selection for nonMKA parents. To clarify this last point, a parent had zero chance of selection if he/she resided
with his/her child but was not viewed by the household respondent as the MKA or as the
spouse/partner of the MKA. Such situations are exceedingly rare, so any bias was of little
consequence.
To reduce the variability in the sampling rates of other adults in households with large numbers
of adults, the sample size for other adults in a household depended upon the number of such
adults present. If there were only one or two, then one of these adults was randomly selected. If
there were three or more present, then two were randomly selected.
Once the random selection of other adults was completed, the sampled adults were interviewed.
During each such interview, data were collected about both the sample adult and his/her
spouse/partner if the spouse/partner also lived in the household.3 All this information was
collected by proxy through the sample adult. Questions on health care insurance and usage were
asked about both the selected adult and the spouse/partner, but proxy responses from the selected
adult were accepted for both of them. As with the MKA interview, some questions were asked
only of the sample adult.
5.3
Sample Selection of Adults from Adult-Only Households
Sampling methods used for adults under age 65 in adult-only households were similar to those
used for other adults in households with children. A sample of either one or two adults was
selected depending upon how many adults were present. Unlike other adults in households with
children, the within-household subsampling rule was different for RDD households with three
adults under age 65 than in corresponding area households. A decision was made based on a
random number whether to sample one or two in the RDD component. The probabilities for
deciding whether to take one or two adults from adult-only RDD households with three adults
3
Since the sample selection was done at the close of the MKA interview, it was not possible to accidentally select both an adult and his/her
spouse/partner and then ask them both about themselves and each other.
5-2
varied by study area and are given in table 5-1. These probabilities were jointly set with the
subsampling rate for adult-only telephone households in order to achieve targeted sample sizes
for adults from adult-only households discussed in chapter 3. The rates are the same as used in
Rounds 1 and 2. Note that the rates do not vary by household income.
Table 5-1.
Proportion of RDD Adult-Only Households with Three Other Adults under Age 65 in
which Just One Adult Is Selected
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Proportion
0.54
0.65
0.49
0.51
0.56
0.53
0.53
0.68
0.62
0.65
0.56
0.53
0.61
0.53
In the area component, two adults were always selected when the household had three adults.
The rules for households with other numbers of adults did not vary by component. As with the
sample of other adults from households with children, data were collected from the sampled
adult about the spouse/partner of each sample adult if that adult lived in the same household. The
only exceptions were the set of questions that were asked only of the sampled adult.
Since no household relationship data were available at the time of sampling in these households
(in households with children the data were available), it was possible to select both an adult and
the spouse of that adult. If the relationships revealed a sampled adult was the spouse/partner of
the sampled adult, then the interview for that adult was automatically deleted.
5.4
Achieved Sample Sizes and Response Rates
This section gives a series of tables showing the outcomes of the within-household sampling
operations for children and adults by survey component (RDD and area) and by household
income (low and overall). The tables give achieved nominal sample sizes and response rates by
site. Effective sample sizes will be smaller due to design effects that arise from clustering in the
area sample and the variability of the sampling rates and weights. Effective sample sizes will be
5-3
given in the 2002 NSAF Variance Estimation, Report No. 4. Some groups that were sampled at
lower rates than other groups are as follows:
Households without phones,
Households with screener income above 200 percent of the poverty threshold,
Households outside of the study areas,
Children in households with multiple children in the same age range,
Adults in households without children, and
Other adults in households with children and multiple adults besides the MKA and spouse
(e.g., grown children living with parents and young siblings).
The first five tables (table 5-2 through 5-6) are the sample sizes from the RDD component. The
first two tables are sample sizes of children and the remaining are sample sizes of adults. The
next five tables (table 5-7 through 5-11) are for the area component. The same pattern is
followed, with the first two tables in the set giving child sample sizes and the other three tables
giving adult sample sizes. The last five tables (table 5-12 through 5-16) give the combined RDD
and area sample sizes.
The response rates reported in the tables are the simple unweighted ratios of the number of
interviewed persons to the number of eligible sampled persons. These rates reflect the
operational aspects of interviewing for different groups. Response rates that include differential
sampling rates and combine the screener and extended rates are discussed in detail in the 2002
NSAF Data Editing, Report No. 8.
Tables 5-2 and 5-3 show that in the RDD sample, children under age 6 were selected at a higher
rate than those 6 to 17 years old. This is a direct result of the subsampling rules that specified
that only one child in the age range was to be sampled per household. Since the mean number of
children per household is greater for 6- to 17-year-olds, the subsampling rate is lower for this
group. The response rate in table 5-2 is about the same as the response rate in table 5-3,
indicating the extended interview response rates were relatively constant by income level. This
pattern is consistent with previous rounds of the NSAF.
Tables 5-4 and 5-5 give the results for the sampling and extended interview response rates for
other adults (adults who are not MKAs or their spouse/partner) in telephone households. Nearly
80 percent of the other adults listed in households with children were sampled, but only about 60
percent in households without children were sampled. The extended response rate was higher in
adult-only households than in households with children. The higher response rate may have been
a consequence of having more difficulty locating and interviewing other adults (many of whom
were older siblings still living at home) in households with children. In adult-only households, it
is often possible to continue the interview since the screener respondent may be the sampled
adult, and this tends to increase the response rate in these households.
5-4
Table 5-6 shows that 65 percent of the adult interviews were conducted with the MKA of the
sampled child (28,208 of 43,158). Only 28 percent of the adult interviews were conducted with
adults in households without children (12,053), and the remaining 7 percent were other adults in
households with children (2,872).
Tables 5-7 to 5-11 are the corresponding counts for nontelephone households. In general, the
results are very similar to those noted earlier for the telephone sample. The main difference is
that the extended response rates are much higher—for children the response rates are about 95
percent, for other adults in households with children the rates are around 75 percent, and for
adults in adult-only households the rates are about 94 percent. Again, the likely reason for the
lower rates for other adults in households with children has to do with locating the sampled
person to conduct the interview. Table 5-11 shows that the percent of interviews with adults in
households without children is higher than in the RDD sample, and the percent of MKA
interviews is lower.
Tables 5-12 to 5-16 combine the telephone and nontelephone samples, producing counts for the
entire effort. While these tables are valuable in the sense that they give the overall counts from
Round 3, they provide less information about the interviewing operations than the previous
tables because they combine units with different attributes. For example, the extended response
rates in these tables are very similar to those in the telephone tables because the sample size from
the RDD sample was so much larger that it dominates the combined rates. Thus, these tables do
not provide much insight into the survey operations.
5-5
Table 5-2.
Within-Household Sampling and Extended Interviews of Children in Telephone Low-Income Households
Children under Age 6
5-6
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Listed
556
910
597
571
465
649
499
570
645
536
697
626
549
1,401
9,271
Selected
398
642
414
401
336
435
342
405
464
390
483
419
366
1,001
6,496
Children Age 6 to 17
Average
Selection
Rate (%)
71.6
70.5
69.3
70.2
72.3
67.0
68.5
71.1
71.9
72.8
69.3
66.9
66.7
71.4
70.1
Interviewed
336
491
337
309
264
365
282
331
350
305
409
351
328
873
5,331
Ineligible
0
0
0
3
0
0
0
0
0
0
0
0
0
0
3
Response
Rate (%)
84
76
81
78
79
84
82
82
75
78
85
84
90
87
82
Listed
1,168
1,955
1,186
1,336
1,122
1,386
1,149
1,320
1,420
1,123
1,356
1,175
1,128
3,283
20,107
Selected
677
1,015
624
735
603
708
578
742
796
634
744
628
570
1,712
10,766
Average
Selection
Rate (%)
58.0
51.9
52.6
55.0
53.7
51.1
50.3
56.2
56.1
56.5
54.9
53.4
50.5
52.1
53.5
Interviewed
574
779
530
545
475
595
502
607
596
505
620
542
516
1,501
8,887
Ineligible
0
0
1
1
0
0
0
0
0
0
0
0
0
0
2
Response
Rate (%)
85
77
85
74
79
84
87
82
75
80
83
86
91
88
83
Table 5-3.
Within-Household Sampling and Extended Interviews of Children in All Subsampled Telephone Households
Children under Age 6
5-7
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Listed
891
1,591
1,486
1,007
1,255
1,552
1,449
849
1,671
1,228
1,212
1,340
1,383
3,107
20,021
Selected
667
1,169
1,084
749
910
1,115
1,046
625
1,231
916
879
946
974
2,284
14,595
Children Age 6 to 17
Average
Selection
Rate (%)
74.9
73.5
72.9
74.4
72.5
71.8
72.2
73.6
73.7
74.6
72.5
70.6
70.4
73.5
72.9
Interviewed
564
909
909
596
726
935
898
500
941
712
732
813
866
1,987
12,088
Ineligible
0
0
0
3
0
0
0
0
0
0
0
0
0
0
3
Respons
e Rate
(%)
85
78
84
80
80
84
86
80
76
78
83
86
89
87
83
Listed
2,109
3,442
3,264
2,430
2,841
3,644
3,385
2,090
3,701
2,569
2,644
2,866
3,265
7,104
45,354
Selected
1,280
1,923
1,916
1,456
1,661
2,044
1,937
1,239
2,197
1,535
1,557
1,686
1,871
4,119
26,421
Average
Selection
Rate (%)
60.7
55.9
58.7
59.9
58.5
56.1
57.2
59.3
59.4
59.8
58.9
58.8
57.3
58.0
58.3
Interviewed
1,073
1,499
1,621
1,123
1,309
1,708
1,686
1,001
1,681
1,230
1,264
1,443
1,669
3,557
21,864
Ineligible
0
0
1
1
0
0
0
0
0
0
0
0
0
0
2
Response
Rate (%)
84
78
85
77
79
84
87
81
77
80
81
86
89
86
83
Table 5-4.
Within-Household Sampling and Extended Interviews of Other Adults
in Subsampled Telephone Households with Children
5-8
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Low-Income Telephone Households with Children
Listed
Average
Other
Selection
InterInAdults
Selected Rate (%) viewed
eligible
142
118
83.1
69
1
432
310
71.8
185
7
128
94
73.4
61
0
137
105
76.6
71
0
91
74
81.3
43
0
160
128
80.0
88
2
156
119
76.3
90
1
134
116
86.6
70
0
152
110
72.4
51
3
169
130
76.9
70
4
146
111
76.0
81
1
113
88
77.9
62
1
107
81
75.7
58
0
396
309
78.0
222
2
2,463
1,893
76.9
1,221
22
Response
Rate (%)
59
61
65
68
58
70
76
60
48
56
74
71
72
72
65
All Telephone Households with Children
Listed
Average
Other
Selection
InterAdults
Selected Rate (%) viewed
249
209
83.9
132
711
515
72.4
310
320
253
79.1
168
255
200
78.4
127
299
235
78.6
141
423
333
78.7
243
468
378
80.8
278
196
167
85.2
101
379
290
76.5
158
370
283
76.5
166
280
217
77.5
154
265
215
81.1
148
362
288
79.6
225
886
700
79.0
521
5,463
4,283
78.4
2,872
Note: Other adults are adults who are neither the MKA of a sample child nor the spouse of such a person, nor the parent of a child under age 18 in the household.
Ineligible
1
11
0
0
1
4
4
0
5
4
1
2
0
4
37
Response
Rate (%)
63
62
66
64
60
74
74
60
55
59
71
69
78
75
68
Table 5-5.
Subsampling and Extended Interviews of Adults in Subsampled Adult-Only Telephone Households
5-9
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Low-Income Adult-Only Households
Average
Listed
Selection
InterAdults
Selected Rate (%) viewed
735
463
63.0
350
916
569
62.1
402
624
392
62.8
306
619
389
62.8
297
734
453
61.7
313
747
490
65.6
367
866
564
65.1
472
667
427
64.0
321
519
304
58.6
197
592
368
62.2
263
515
314
61.0
241
552
338
61.2
263
615
400
65.0
328
1,997
1,275
63.8
1,046
10,698
6,746
63.1
5,166
Ineligible
11
14
6
6
10
10
16
12
6
11
9
8
6
22
147
Response
Rate (%)
77
72
79
78
71
76
86
77
66
74
79
80
83
83
78
All Adult-Only Households
Average
Listed
Selection
Adults
Selected Rate (%)
1,438
879
61.1
2,089
1,262
60.4
1,530
928
60.7
1,247
755
60.5
1,996
1,193
59.8
2,145
1,294
60.3
2,716
1,655
60.9
1,204
753
62.5
1,762
1,023
58.1
1,381
844
61.1
1,000
604
60.4
1,504
882
58.6
1,936
1,177
60.8
4,401
2,674
60.8
26,349
15,923
60.4
Interviewed
667
875
722
569
818
985
1,346
578
684
592
466
679
945
2,127
12,053
Ineligible
17
24
16
15
31
24
40
22
31
24
16
26
20
56
362
Response
Rate (%)
77
71
79
77
70
78
83
79
69
72
79
79
82
81
77
Table 5-6.
Sources of Adult Telephone Interviews
5-10
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
MKA for
Sample
Child
1,357
1,941
2,113
1,454
1,708
2,196
2,173
1,254
2,184
1,606
1,620
1,862
2,137
4,603
28,208
Spouse of
MKA
916
1,432
1,660
1,038
1,243
1,630
1,711
815
1,611
1,126
1,218
1,414
1,687
3,391
20,892
Emancipated
minors
2
6
2
4
1
0
1
0
1
3
0
0
0
5
25
Spouse
0
0
1
3
0
0
0
0
0
1
0
0
0
2
7
Other Adults in
Household with Children
Interviewed Spouse
132
12
310
51
168
11
127
10
141
11
243
30
278
13
101
7
158
22
166
15
154
25
148
10
225
14
521
61
2,872
292
Adults in Household
without Children
Interviewed
Spouse
667
288
875
341
722
317
569
261
818
336
985
429
1,346
642
578
255
684
302
592
230
466
202
679
317
945
463
2,127
963
12,053
5,346
Total Adult
Extended
Interviews
2,158
3,132
3,005
2,154
2,668
3,424
3,798
1,933
3,027
2,367
2,240
2,689
3,307
7,256
43,158
Table 5-7.
Within-Household Sampling and Extended Interviews of Children in Nontelephone Low-Income Households
5-11
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Children under Age 6
Average
Selection
Listed
Selected Rate (%)
6
5
83.3
25
15
60.0
0
0
0.0
14
10
71.4
4
3
75.0
16
6
37.5
0
0
0.0
8
5
62.5
2
1
50.0
3
3
100.0
57
30
52.6
1
1
100.0
18
10
55.6
102
68
66.7
256
157
61.3
Children Age 6 to 17
Interviewed
5
15
0
9
3
6
0
5
1
2
28
1
10
64
149
Ineligible
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Response
Rate (%)
100
100
0
90
100
100
0
100
100
67
93
100
100
94
95
Listed
12
39
4
10
5
13
0
24
0
5
49
6
24
168
359
Selected
8
15
2
5
3
6
0
9
0
3
25
3
10
88
177
Average
Selection
Rate (%)
66.7
38.5
50.0
50.0
60.0
46.2
0.0
37.5
0.0
60.0
51.0
50.0
41.7
52.4
49.3
Interviewed
8
15
2
4
3
6
0
8
0
3
23
2
10
84
168
Ineligible
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Response
Rate (%)
100
100
100
80
100
100
0
89
0
100
92
67
100
95
95
Table 5-8.
Within-Household Sampling and Extended Interviews of Children in All Nontelephone Households
5-12
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Children under Age 6
Average
Selection
Listed
Selected Rate (%)
7
6
85.7
25
15
60.0
0
0
0.0
15
11
73.3
4
3
75.0
16
6
37.5
0
0
0.0
9
6
66.7
2
1
50.0
4
4
100.0
65
34
52.3
2
2
100.0
21
12
57.1
135
89
65.9
305
189
62.0
Children Age 6 to 17
Interviewed
6
15
0
10
3
6
0
6
1
3
32
2
12
84
180
Ineligible
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Response
Rate (%)
100
100
0
91
100
100
0
100
100
75
94
100
100
94
95
Listed
12
39
4
14
5
13
0
25
0
6
52
8
28
201
407
Selected
8
15
2
7
3
6
0
10
0
4
28
5
13
107
208
Average
Selection
Rate (%)
66.7
38.5
50.0
50.0
60.0
46.2
0.0
40.0
0.0
66.7
53.8
62.5
46.4
53.2
51.1
Interviewed
8
15
2
6
3
6
0
9
0
4
26
4
13
103
199
Ineligible
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Response
Rate (%)
100
100
100
86
100
100
0
90
0
100
93
80
100
96
96
Table 5-9.
Within-Household Sampling and Extended Interviews of Other Adults in Nontelephone Households with Children
5-13
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Low-Income Telephone Households with Children
Listed
Average
Other
Selection
InterInAdults Selected Rate (%) viewed
eligible
0
0
0.0
0
0
9
7
77.8
6
0
0
0
0.0
0
0
1
1
100.0
1
0
0
0
0.0
0
0
0
0
0.0
0
0
0
0
0.0
0
0
5
4
80.0
3
0
0
0
0.0
0
0
3
2
66.7
0
0
9
8
88.9
6
0
1
1
100.0
1
0
3
3
100.0
3
0
32
25
78.1
19
0
63
51
81.0
39
0
Response
Rate (%)
0
86
0
100
0
0
0
75
0
0
75
100
100
76
76
All Telephone Households with Children
Listed
Average
Other
Selection
InterAdults
Selected
Rate (%) viewed
0
0
0.0
0
9
7
77.8
6
0
0
0.0
0
1
1
100.0
1
0
0
0.0
0
0
0
0.0
0
0
0
0.0
0
5
4
80.0
3
0
0
0.0
0
3
2
66.7
0
9
8
88.9
6
1
1
100.0
1
6
5
83.3
5
43
33
76.7
24
77
61
79.2
46
Note: Other adults are adults who are neither the MKA of a sample child nor the spouse of such a person, nor the parent of a child under age 18 in the household.
Ineligible
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
Response
Rate (%)
0
86
0
100
0
0
0
75
0
0
75
100
100
73
75
Table 5-10.
Subsampling and Extended Interviews of Adults in Adult-Only Nontelephone Households
5-14
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
Low-income Adult-Only Nontelephone Households
Listed
Selection
InterInAdults
Selected Rate (%) viewed
eligible
17
11
64.7
10
0
17
13
76.5
13
0
8
8
100.0
8
0
14
11
78.6
11
0
6
4
66.7
4
0
4
3
75.0
3
0
1
1
100.0
1
0
17
12
70.6
12
0
1
1
100.0
1
0
3
2
66.7
2
0
43
27
62.8
24
1
6
3
50.0
2
1
23
17
73.9
16
1
176
119
67.6
109
2
336
232
69.0
216
5
Response
Rate (%)
91
100
100
100
100
100
100
100
100
100
92
100
100
93
95
All Adult-Only Nontelephone Households
Listed
Selection
InterAdults
Selected Rate (%) viewed
20
14
70.0
13
30
21
70.0
21
9
9
100.0
9
19
14
73.7
13
10
7
70.0
7
8
5
62.5
4
9
6
66.7
5
20
14
70.0
14
1
1
100.0
1
5
4
80.0
4
60
41
68.3
38
8
5
62.5
4
31
23
74.2
22
257
172
66.9
153
487
336
69.0
308
Ineligible
0
0
0
1
0
0
0
0
0
0
1
1
1
4
8
Response
Rate (%)
93
100
100
100
100
80
83
100
100
100
95
100
100
91
94
Table 5-11.
Sources of Adult Nontelephone Interviews
5-15
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
MKA for
Sample Child
10
22
2
13
5
9
0
12
1
6
45
5
21
143
294
Spouse of
MKA
2
14
0
7
0
2
0
2
0
3
26
3
3
86
148
Emancipated
Minor
0
0
0
0
0
0
0
0
0
0
1
0
0
0
1
Spouse
0
0
0
0
0
0
0
0
0
0
1
0
0
0
1
Other Adults in
Household with Children
Adults in Household
without Children
Interviewed
Interviewed
0
6
0
1
0
0
0
3
0
0
6
1
5
24
46
Spouse
0
1
0
0
0
0
0
0
0
0
0
0
0
2
3
13
21
9
13
7
4
5
14
1
4
38
4
22
153
308
Spouse
3
1
0
2
0
2
1
4
0
0
12
1
6
41
73
Total Adult
Extended Interviews
23
49
11
27
12
13
5
29
2
10
90
10
48
320
649
Table 5-12.
Within-Household Sampling and Extended Interviews of Children in Telephone
and Nontelephone Low-Income Households
5-16
Children under Age 6
Average
Selection
Study Area
Listed Selected Rate (%)
Alabama
562
403
71.7
California
935
657
70.3
Colorado
597
414
69.3
Florida
585
411
70.3
Massachusetts
469
339
72.3
Michigan
665
441
66.3
Minnesota
499
342
68.5
Mississippi
578
410
70.9
New Jersey
647
465
71.9
New York
539
393
72.9
Texas
754
513
68.0
Washington
627
420
67.0
Wisconsin
567
376
66.3
Bal. of nation 1,503
1,069
71.1
Total
9,527
6,653
69.8
Children Age 6 to 17
Interviewed
341
506
337
318
267
371
282
336
351
307
437
352
338
937
5,480
Ineligible
0
0
0
3
0
0
0
0
0
0
0
0
0
0
3
Response
Rate (%)
85
77
81
78
79
84
82
82
75
78
85
84
90
88
82
Listed
1,180
1,994
1,190
1,346
1,127
1,399
1,149
1,344
1,420
1,128
1,405
1,181
1,152
3,451
20,466
Selected
685
1,030
626
740
606
714
578
751
796
637
769
631
580
1,800
10,943
Average
Selection
Rate (%)
58.1
51.7
52.6
55.0
53.8
51.0
50.3
55.9
56.1
56.5
54.7
53.4
50.3
52.2
53.5
Interviewed
582
794
532
549
478
601
502
615
596
508
643
544
526
1,585
9,055
Ineligible
0
0
1
1
0
0
0
0
0
0
0
0
0
0
2
Response
Rate (%)
85
77
85
74
79
84
87
82
75
80
84
86
91
88
83
Table 5-13.
Within-Household Sampling and Extended Interviews of Children in All
Telephone and Nontelephone Households
Children under Age 6
5-17
Study Area
Listed
Alabama
898
California
1,616
Colorado
1,486
Florida
1,022
Massachusetts 1,259
Michigan
1,568
Minnesota
1,449
Mississippi
858
New Jersey
1,673
New York
1,232
Texas
1,277
Washington
1,342
Wisconsin
1,404
Bal. of nation
3,242
Total
20,326
Selected
673
1,184
1,084
760
913
1,121
1,046
631
1,232
920
913
948
986
2,373
14,784
Average
Selection
Rate (%)
74.9
73.3
72.9
74.4
72.5
71.5
72.2
73.5
73.6
74.7
71.5
70.6
70.2
73.2
72.7
Interviewed
570
924
909
606
729
941
898
506
942
715
764
815
878
2,071
12,268
Ineligible
0
0
0
3
0
0
0
0
0
0
0
0
0
0
3
Children Age 6 to 17
Average
Response
Selection
Rate (%) Listed Selected Rate (%)
85
2,121
1,288
60.7
78
3,481
1,938
55.7
84
3,268
1,918
58.7
80
2,444
1,463
59.9
80
2,846
1,664
58.5
84
3,657
2,050
56.1
86
3,385
1,937
57.2
80
2,115
1,249
59.1
76
3,701
2,197
59.4
78
2,575
1,539
59.8
84
2,696
1,585
58.8
86
2,874
1,691
58.8
89
3,293
1,884
57.2
87
7,305
4,226
57.9
83
45,761 26,629
58.2
Interviewed
1,081
1,514
1,623
1,129
1,312
1,714
1,686
1,010
1,681
1,234
1,290
1,447
1,682
3,660
22,063
Ineligible
0
0
1
1
0
0
0
0
0
0
0
0
0
0
2
Response
Rate (%)
84
78
85
77
79
84
87
81
77
80
81
86
89
87
83
Table 5-14.
Within-Household Sampling and Extended Interviews of Other Adults in Telephone
and Nontelephone Households with Children
5-18
Low-Income Households with Children
Listed
Average
Other
Selection InterStudy Area
Adults Selected Rate (%) viewed
Alabama
142
118
83.1
69
California
441
317
71.9
191
Colorado
128
94
73.4
61
Florida
138
106
76.8
72
Massachusetts
91
74
81.3
43
Michigan
160
128
80.0
88
Minnesota
156
119
76.3
90
Mississippi
139
120
86.3
73
New Jersey
152
110
72.4
51
New York
172
132
76.7
70
Texas
155
119
76.8
87
Washington
114
89
78.1
63
Wisconsin
110
84
76.4
61
Bal. of nation
428
334
78.0
241
Total
2,526
1,944
77.0
1,260
Ineligible
1
7
0
0
0
2
1
0
3
4
1
1
0
2
22
Response
Rate (%)
59
62
65
68
58
70
76
61
48
55
74
72
73
73
66
All Households with Children
Listed
Average
Other
Selection
Adults Selected Rate (%)
249
209
83.9
720
522
72.5
320
253
79.1
256
201
78.5
299
235
78.6
423
333
78.7
468
378
80.8
201
171
85.1
379
290
76.5
373
285
76.4
289
225
77.9
266
216
81.2
368
293
79.6
929
733
78.9
5,540
4,344
78.4
Interviewed
132
316
168
128
141
243
278
104
158
166
160
149
230
545
2,918
Note: Other adults are adults who are neither the MKA of a sample child nor the spouse of such a person, nor the parent of a child under age 18 in the household.
Ineligible
1
11
0
0
1
4
4
0
5
4
1
2
0
4
37
Response
Rate (%)
63
62
66
64
60
74
74
61
55
59
71
70
78
75
68
Table 5-15.
Subsampling and Extended Interviews of Adults in Adult-Only
Telephone and Nontelephone Households
5-19
Low-Income Adult-Only Households
Average
Selection InterStudy Area
Listed
Selected Rate (%) viewed
Alabama
752
474
63.0
360
California
933
582
62.4
415
Colorado
632
400
63.3
314
Florida
633
400
63.2
308
Massachusetts
740
457
61.8
317
Michigan
751
493
65.6
370
Minnesota
867
565
65.2
473
Mississippi
684
439
64.2
333
New Jersey
520
305
58.7
198
New York
595
370
62.2
265
Texas
558
341
61.1
265
Washington
558
341
61.1
265
Wisconsin
638
417
65.4
344
Bal. of nation
2,173
1,394
64.2
1,155
Total
11,034
6,978
63.2
5,382
Ineligible
11
14
6
6
10
10
16
12
6
11
10
9
7
24
152
Response
Rate (%)
78
73
80
78
71
77
86
78
66
74
80
80
84
84
79
All Adult-Only Households
Average
Selection
Listed
Selected Rate (%)
1,458
893
61.2
2,119
1,283
60.5
1,539
937
60.9
1,266
769
60.7
2,006
1,200
59.8
2,153
1,299
60.3
2,725
1,661
61.0
1,224
767
62.7
1,763
1,024
58.1
1,386
848
61.2
1,060
645
60.8
1,512
887
58.7
1,967
1,200
61.0
4,658
2,846
61.1
26,836
16,259
60.6
Interviewed
680
896
731
582
825
989
1,351
592
685
596
504
683
967
2,280
12,361
Ineligible
17
24
16
16
31
24
40
22
31
24
17
27
21
60
370
Response
Rate (%)
78
71
79
77
71
78
83
79
69
72
80
79
82
82
78
Table 5-16.
Sources of Adult Telephone and Nontelephone Interviews
5-20
Study Area
Alabama
California
Colorado
Florida
Massachusetts
Michigan
Minnesota
Mississippi
New Jersey
New York
Texas
Washington
Wisconsin
Bal. of nation
Total
MKA for
Sample Child
1,367
1,963
2,115
1,467
1,713
2,205
2,173
1,266
2,185
1,612
1,665
1,867
2,158
4,746
28,502
Spouse of
MKA
918
1,446
1,660
1,045
1,243
1,632
1,711
817
1,611
1,129
1,244
1,417
1,690
3,477
21,040
Emancipated
Minor
2
6
2
4
1
0
1
0
1
3
1
0
0
5
26
Spouse
0
0
1
3
0
0
0
0
0
1
1
0
0
2
8
Other Adults in Household
with Children
Interviewed
Spouse
132
12
316
52
168
11
128
10
141
11
243
30
278
13
104
7
158
22
166
15
160
25
149
10
230
14
545
63
2,918
295
Adults in Household
without Children
Interviewed
Spouse
680
291
896
342
731
317
582
263
825
336
989
431
1,351
643
592
259
685
302
596
230
504
214
683
318
967
469
2,280
1,004
12,361
5,419
Total Adult
Extended
Interviews
2,181
3,181
3,016
2,181
2,680
3,437
3,803
1,962
3,029
2,377
2,330
2,699
3,355
7,576
43,807
6. CONCLUSION
The primary goal of this report has been to describe the features of the sampling procedures used
in the 2002 NSAF. Since this is the third round of data collection for NSAF we also pointed out
how the Round 3 design is similar to the design in the previous rounds but does contain a number
of important changes.
The design for the survey is complicated because it uses a dual-frame approach with a sample of
telephone households selected in each study area combined with a sample of nontelephone
households. While a dual-frame design was used in Round 1 and Round 2, a very important
change in the design in Round 3 was the elimination of separate nontelephone samples in each
study area. The nontelephone sample in Round 3 was designed to produce national estimates of
nontelephone households that could be combined with the RDD sample to produce national
estimates of all households. The study area estimates will be based entirely on the RDD sample,
using methods described in 2002 NSAF Sample Estimation Survey Weights, Report No. 3.
Another very important change in the sample design for Round 3 was the combining of the
Milwaukee, Wisconsin, and the balance-of-Wisconsin study areas into one study area for all of
Wisconsin.
As discussed in the report, many of the procedures used in Round 3 were based on experiences
from the earlier rounds. The data from Round 1 were used to optimize the sample design for
Round 2 and the research results were directly carried forward to Round 3. In addition, the
earlier data collection efforts provided parameters estimates that were the initial assumptions
used to design the Round 3 sample. Estimates of response rates, residency rates, nontelephone
eligibility rates, and incoming switching are just a few parameters derived from the earlier
rounds.
As in previous rounds, the data collection effort was monitored and deviations from the assumed
rates were tracked. Based on the tracking of these data changes were made in the sample during
the data collection period. The sample size for the RDD component was increased in some study
areas and the sampling rates were modified, as deemed necessary. One procedure that was
introduced in the RDD sample for Round 3 as a result of the monitoring was refusal
subsampling. Using this procedure, telephone numbers that resulted in refusals at the screener
level were subsampled and only the retained telephone numbers were included in refusal
conversion efforts.
The sampling procedures used in Round 3 have important implications for producing estimates
from Round 3 and for producing estimates of change from 1997 or 1999 to 2002. In particular,
the elimination of the nontelephone sample in the study areas complicates analyses in the sense
that two sets of weights are required in Round 3, one for producing national estimates and
another for producing study-area estimates. These estimation issues are addressed in 2002 NSAF
Sample Estimation Survey Weights, Report No. 3 and 2002 NSAF Variance Estimation, Report
No. 4.
6-1
REFERENCES
Methodology References
Brick, M. Forthcoming. 2002 NSAF Sample Estimation Survey Weights. Methodology
Report No. 3.
Brick, M. Forthcoming. 2002 NSAF Variance Estimation. Methodology Report No. 4.
Brick, M. Forthcoming. 2002 NSAF Response Rates and Methods Evaluation.
Methodology Report No. 7.
Flores-Cervantes, Ismael, J. Michael Brick, and Ralph DiGaetano. 1999. 1997 NSAF
Variance Estimation. Methodology Report No. 4.
Judkins, David, Gary Shapiro, J. Michael Brick, Ismael Flores-Cervantes, David Ferraro,
Teresa Strickler, and Joseph Waksberg. 1999. 1997 NSAF Sample Design. Methodology
Report No. 2.
Judkins, David, J. Michael Brick, Pam Broene, David Ferraro, and Teresa Strickler. 2001.
1999 NSAF Sample Design. Methodology Report No. 2.
General References
Brick, J. Michael, J. Montaquila, and Fritz Scheuren. 2002. “Estimating Residency Rates
for Undetermined Telephone Numbers.” Public Opinion Quarterly 66: 18–39.
Brick, J. Michael, J. Waksberg, D. Kulp, and A. Starer. 1995. “Bias in List-Assisted
Telephone Surveys.” Public Opinion Quarterly 59(2): 218–35.
Casady, R., and J. Lepkowki. 1993. “Stratified Telephone Survey Designs.” Survey
Methodology 19: 103–13.
Ferraro, D., and J. Michael Brick. 2001. “Weighting for Nontelephone Households in RDD
Surveys.” Proceedings of the Survey Research Methods Section of the American Statistical
Association [CD-ROM], Alexandria, VA: American Statistical Association.
Giesbrecht, L.H., D.W. Kulp, and A.W. Starer. 1996. “Estimating Coverage Bias in RDD
Samples with Current Population Survey Data.” Proceedings of the Survey Research
Methods Section, American Statistical Association (503–8).
Kalton, G., D. Kapryzk, and D. McMillen. 1989. “Nonsampling Errors in Panel Surveys.”
In Panel Surveys, edited by Kaspryzk, Duncan, Kalton, and Singh (249–70). New York:
John Wiley and Sons.
R-1
Tucker, C., J.M. Lepkowski, and L. Piekarski. 2002. “The Current Efficiency of ListAssisted Telephone Sampling Designs.” Public Opinion Quarterly 66: 321–38.
Waksberg, J. 1978. “Sampling Methods for Random Digit Dialing.” Journal of the
American Statistical Association 73: 40–46.
Waksberg, J., Brick, J.M., Shapiro, G., Flores-Cervantes, I., and Bell, B. 1997. Dual-frame
RDD and area sample for household survey with particular focus on low-income
population. Proceedings of the Survey Research Methods Section of the American
Statistical Association (713–18).
U.S. Bureau of the Census. 2000. Technical Paper 63, Current Population Survey: Design
and Methodology. http://www.bls.census.gov/cps/tp/tp63.htm.
R-2