These technical notes include information on sampling and weighting, survey methodology, sampling and nonsampling errors, and discussions of data comparisons to previous cycles of the National Survey of Recent College Graduates (NSRCG) and the Integrated Postsecondary Education Data System (IPEDS) data. For a more detailed discussion of survey methodology, readers are referred to the 1999 NSRCG Methodology Report.
The National Survey of Recent College Graduates (NSRCG) is sponsored by the National Science Foundation (NSF), Division of Science Resources Statistics (SRS). The NSRCG is one of three data collections covering personnel and graduates in science and engineering. The other two surveys are the National Survey of College Graduates (NSCG) and the Survey of Doctorate Recipients (SDR). Together, they constitute NSF's Scientists and Engineers Statistical Data System (SESTAT). These surveys serve as the basis for developing estimates and characteristics of the total population of scientists and engineers in the United States.
The first NSF-sponsored NSRCG (then known as New Entrants) was conducted in 1974. Subsequent surveys were conducted in 1976, 1978, 1979, 1980, 1982, 1984, 1986, 1988, 1990, 1993, 1995, 1997, and 1999. The initial survey collected data on only bachelor's degree recipients, but all subsequent surveys included both bachelor's and master's degree recipients.
For the 1999 NSRCG, a sample of 279 colleges and universities was asked to provide lists of eligible bachelor's and master's degree recipients. From these lists, a sample of 13,918 graduates (9,786 bachelor's and 4,132 master's recipients) was selected. These graduates were interviewed between May 1999 and March 2000. Computer-assisted telephone interviewing (CATI) served as the primary means of data collection. Mail data collection was used only for those who could not be reached by telephone. The weighted response rates were 99.5 percent for institutions and 78 percent for graduates.
The NSRCG questionnaire underwent relatively few revisions for the 1999 survey. These revisions consisted mainly of deleting a series of questions about alternative arrangements with employers that had been added for the 1997 cycle only. All revisions were done in coordination with similar revisions to the other SESTAT surveys. Topics covered in the survey include:
The NSRCG used a two-stage sample design. In the first stage, a stratified nationally representative sample of 279 institutions was selected with probability proportional to size. There were 106 self-representing institutions, also known as certainty units. For each institution, the measure of size was a composite related to both the number of graduates and the proportion of these who were black or Hispanic. The 173 noncertainty institutions were implicitly stratified by sorting the list by type of control (public, private), region, and the percentage of degrees awarded in science or engineering. Institutions were then selected by systematic sampling from the ordered list.
The second stage of the sampling process involved selecting graduates within the sampled institutions by cohort. Each sampled institution was asked to provide lists of graduates for sampling. Within graduation year (cohort), each eligible graduate was then classified into one of 40 strata based on the graduate's major field of study and degree level. While race was not an explicit stratification variable, black, Hispanic, and American Indian/Alaskan Native graduates were assigned a measure of size equal to three, while all other graduates were assigned a measure of size equal to one. This method had the same effect as oversampling black, Hispanic, and American Indian/Alaskan Native graduates by a factor of three. Table 1 lists the major fields and the corresponding sampling rates by cohort and degree.
These rates are overall sampling rates for the major field, and include the institution's probability of selection and the within-institution sampling rate. To achieve the within-institution sampling rate, the overall rate was divided by the institution's probability of selection. The sampling rates by stratum were applied within each eligible responding institution and resulted in sampling 13,918 graduates, slightly larger than the target sample size of 13,500 because persons with unknown majors were also included for complete population coverage.
To be included in the sample, the graduates had to meet all of the following criteria:
Prior to data collection from graduates, it was first necessary to obtain the cooperation of the sampled institutions that provided lists of graduates. All eligible sampled institutions except one provided graduate lists for the 1999 NSRCG. In addition, one sampled institution was ineligible because no S&E degrees were awarded during the two cohort years for the 1999 survey. The response rates for the institutional list collection were 99.6 percent unweighted and 99.5 percent weighted.
Graduate data collection took place between May 1999 and March 2000, with computer-assisted telephone interviewing as the primary means of data collection. Flyers were sent to all graduates announcing the study and asking for the phone numbers at which they could be reached during the survey period. Extensive tracing of graduates was required to obtain the desired response rate. Tracing activities included computerized telephone number searches, national change of address searches (NCOA), school alumni office contacts, school major field department contacts, directory assistance, military locators, post office records, personal referrals from parents or others who knew the graduate, and the use of professional tracing organizations.
Table 2 gives the response rates by cohort, degree, major, type of address, gender, and race/ethnicity. The overall unweighted graduate response rate was 79 percent; the weighted response rate was 78 percent. As can be seen from table 2, response rates varied somewhat by graduate characteristics. Rates were lowest for graduates with school sampling lists that provided no address, provided a foreign address, or identified the graduate as a nonresident alien. It is possible that many unlocated persons with foreign addresses or listed as nonresident aliens were actually ineligible for the survey due to living outside the United States during the survey reference week. However, a graduate was only classified as ineligible if his/her ineligibility status could be confirmed.
To produce national estimates, the data were weighted. The weighting procedures adjusted for unequal selection probabilities, for nonresponse at the institution and graduate level, and for duplication of graduates on the sampling file (graduates in both cohorts). In addition, a ratio adjustment was made at the institution level, using the number of degrees awarded as reported in IPEDS for specified categories of major and degree level. Because this adjustment was designed to reduce the variability associated with sampling institutions, it was not affected by the differences in target populations between NSRCG and IPEDS at the person level. These differences between NSRCG and IPEDS are discussed in a later section of these notes. The final adjustment to the graduate weights adjusted for responding graduates who could have been sampled twice. For example, a person who obtained an eligible bachelor's degree in 1997 could have obtained an eligible master's degree in 1998 and could have been sampled for either degree. To make the estimates from the survey essentially unbiased, the weights of all responding graduates who could have been sampled twice were divided by 2.
The weights of the graduates who were not eligible to be sampled twice were not adjusted. The weights developed for the 1999 NSRCG comprise both full sample weights for use in computing survey estimates, and replicate weights for variance estimation using a jackknife replication variance estimation procedure.
Most editing checks were included within the CATI system, including range checks, skip pattern rules, and logical consistency checks. Skip patterns were controlled by the CATI system so that inappropriate items were avoided and appropriate items were not missed. For logical consistency check violations, CATI screens appeared that explained the discrepancy and asked the respondent for corrections. Some additional logical consistency checks were added during data preparation. All of the edit checks discussed above were rerun after item nonresponse imputation.
Missing data occurred if the respondent cooperated with the survey but did not answer one or more individual questions. The level of item nonresponse in this study was very low (typically 1 percent or less) due to the use of CATI for data collection and of data retrieval techniques for missing key items. However, imputation for item nonresponse was performed for each survey item to make the study results simpler to present and to allow consistent totals to be obtained when analyzing different questionnaire items. "Not applicable" responses were not imputed because these represented respondents who were not eligible to answer the given item.
Imputation was performed using a hot-deck method. Hot-deck methods estimate the missing value of an item by using values of the same item from other record(s) in the same file. Using the hot-deck procedure, each missing questionnaire item was imputed separately. First, respondent records were sorted by items thought to be related to the missing item. Next, a value was imputed for each item nonresponse recipient from a respondent donor within the same subgroup. The results of the imputation procedure were reviewed to ensure that the plan had been followed correctly. In addition, all edit checks were run on the imputed file to be sure that no data inconsistencies were created in the imputation process.
The survey estimates provided in these tables are subject to two sources of error: sampling and nonsampling errors. Sampling errors occur because the estimates are based on a sample of individuals in the population rather than on the entire population and hence are subject to sampling variability. If the interviews had been conducted with a different sample, the responses would not have been identical; some figures might have been higher, while others might have been lower.
The standard error is the measure of the variability of the estimates due to sampling. It indicates the variability of a sample estimate that would be obtained from all possible samples of a given design and size. Standard errors can be used as a measure of the precision expected from a particular sample. Tables 3 and 4 contain standard errors for key statistics included in the detailed tables.
If all possible samples were surveyed under similar conditions, intervals within plus or minus 1.96 standard errors of a particular statistic would include the true population parameter being estimated in about 95 percent of the samples. This is the 95 percent confidence interval. For example, suppose the total number of 1997 and 1998 bachelor's degree recipients majoring in engineering is 114,612 and the estimated standard error is 4,297. In this case, the 95 percent confidence interval for the statistic would extend from:
114,612 - (4,297 x 1.96) to 114,612 + (4,297 x 1.96) = 106,190 to 123,034
This means that one can be confident that intervals constructed in this way contain the true population parameter for 95 percent of all possible samples.
Estimates of standard errors were computed using a technique known as jackknife replication. As with any replication method, jackknife replication involves constructing a number of subsamples (replicates) from the full sample and computing the statistics of interest for each replicate. The mean square error of the replicate estimates around their corresponding full sample estimate provides an estimate of the sampling variance of the statistic of interest. To construct the replicates, 86 stratified subsamples of the full sample were created. Eighty-six jackknife replicates were then formed by deleting one subsample at a time from the full sample. WesVar, a computer program developed at Westat, was used to calculate direct estimates of standard errors for a number of statistics from the survey.
Computing and printing standard errors for each estimate from the survey is a time consuming and costly effort. For this survey, a different approach was taken for estimating the standard errors of the estimates included in this report. First, the standard errors for a large number of different estimates were directly computed using the jackknife replication procedures described above. Next, models were fitted to the estimates and standard errors and the parameters of these models were estimated from the direct estimates. These models and their estimated parameters were used to approximate the standard error of an estimate from the survey. This process is called the development of generalized variance functions.
Models were fitted for the two types of estimates of primary interest: estimated totals and estimated percentages. It should be noted that the models used to estimate the generalized variance functions may not be completely appropriate for all estimates.
For estimated totals, the generalized variance function applied assumes that the relative variance of the estimate (the square of the standard error divided by the square of the estimate) is a linear function of the inverse of the estimate. Using this model, the standard error of an estimate can be computed as:
where se(y) is the standard error of the estimate y, and a and b are estimated parameters of the model. The parameters of the models were computed separately for 1997 bachelor's, 1997 master's, 1998 bachelor's, and 1998 master's recipients for important domains of interest. The estimates of the parameters are given in table 5.
The following steps should be followed to approximate the standard error of an estimated total:
For example, suppose that the number of 1997 bachelor's degree recipients in engineering who were currently working in an engineering-related job was 39,400 (y = 39,400). The most appropriate domain from Table 5 is engineering majors with bachelor's degrees from 1997 and the parameters are a = 0.001360 and b = 73.981. Approximate the standard error using equation (1) as:
The model used to approximate the standard errors for estimates of percentages was somewhat less complex. The generalized variance for estimated percentages assumed that the ratio of the variance of an estimate to the variance of the same estimate from a simple random sample of the same size was a constant. This ratio is called the design effect and is often labeled the DEFF. Since the variance for an estimated percentage, p, from a simple random sample is p(100-p) divided by the sample size, the standard error of an estimated percentage can be written as:
where n is the sample size or denominator of the estimated percentage. DEFFs were computed separately for 1997 bachelor's, 1997 master's, 1998 bachelor's, and 1998 master's recipients for important domains of interest. The median or average values of the DEFFs from these computations are given in table 5.
The following steps should be followed to approximate the standard error of an estimated percentage:
For example, suppose that the percentage of 1997 bachelor's degree recipients in engineering who were currently working in an S&E job was 67 percent (p = 67) and the number of engineering majors from the survey (sample size, n) was 1,100. The most appropriate domain from table 5 is engineering majors with bachelor's degrees from 1997 and the DEFF for this domain is 1.7. Approximate the standard error using equation (2) as:
In addition to sampling errors, the survey estimates are subject to nonsampling errors that can arise because of nonobservation (nonresponse or noncoverage), reporting errors, and errors made in the collection and processing of the data. These errors can sometimes bias the data. The 1999 NSRCG included procedures specifically designed to minimize nonsampling error. In addition, some special studies conducted during the previous cycles of the NSRCG provided some measures of nonsampling errors that are useful in understanding the data from the current survey as well.
Procedures to minimize nonsampling errors were followed throughout the survey. Extensive questionnaire design work was done by Mathematica Policy Research (MPR), NSF, and Westat. This work included focus groups, expert panel reviews, and mail and CATI pretests. This design work was done in conjunction with the other two SESTAT surveys.
Comprehensive training and monitoring of interviewers and data processing staff helped to ensure the consistency and accuracy of the data file. Data collection was done almost entirely by telephone to help reduce the amount of item nonresponse and item inconsistency. Mail questionnaires were used for cases difficult to complete by telephone. Nonresponse was handled in ways designed to minimize the impact on data quality (through weighting adjustments and imputation). In data preparation, a special effort was made in the area of occupational coding. Respondent-chosen codes were verified by data preparation staff using a variety of information collected on the survey and applying coding rules developed by NSF for the SESTAT system.
While general sampling theory can be used to estimate the sampling variability of a statistic, the measurement of nonsampling error is not easy and usually requires that an experiment be conducted as part of the data collection, or that data external to the study be used. In the 1995 NSRCG, two quality analysis studies were conducted: (1) an analysis of occupational coding; and (2) a CATI reinterview. As noted above, these special studies can also inform analysts about the 1999 survey data.
The occupational coding report included an analysis of the 1995 CATI autocoding of occupation and the best coding operation. During CATI interviewing, each respondent's verbatim occupation description was autocoded by computer into a standard SESTAT code whenever possible. Autocoding included both coding directly to a final category and coding to an intermediate code-selection screen. If the description could not be autocoded, the respondent was asked to select the appropriate occupation category during the interview. For the primary occupation, 22 percent of the responses were autocoded to a final category and 19 percent were autocoded to an intermediate screen. The results of the occupation autocoding were examined, and the process was found to be successful and efficient.
For the best coding operation, an occupational worksheet for each respondent was generated and reviewed by an experienced occupational coder. This review was based on the work-related information provided by the graduate. If the respondent's self-selected occupation code was inappropriate, a new, or "best," code was assigned. A total of 17,894 responses were received to the three occupation questions in the 1995 survey cycle. Of these, 25 percent received updated codes during the best coding process, with 16 percent being recoded from the "other" category and 9 percent recoded from the "nonother" categories. This analysis indicated that the best coding activity was necessary to ensure that the most appropriate occupation codes were included on the final data file. As a result of this 1995 NSRCG quality study, the best coding procedure was implemented in the 1997 and 1999 surveys as well.
The second quality analysis study conducted in the 1995 NSRCG involved a reinterview of a sample of 800 respondents. For this study, sampled respondents were interviewed a second time, and responses to the two interviews were compared. This analysis found that the questionnaire items in which respondents were asked to provide reasons for certain events or behaviors had relatively large index of inconsistency values. Examples include reasons for not working during the reference week and reasons for working part time. High response variability is typical for items that ask about reasons and beliefs rather than behaviors, and the results were not unusual for these types of items. Some of the other differences between the two interviews were attributed to the time lag between the original interview and reinterview.
For the 1993 NSRCG, two data quality studies were completed: (1) an analysis of interviewer variance and (2) a behavioral coding analysis of 100 recorded interviews. The interviewer variance study was designed to measure the impact of interviewer effects on the precision of the estimates. The results showed that interviewer effects for most items were minimal and thus had a very limited effect on the standard error of the estimates. Interviewer variance was highest for open-ended questions.
The behavioral coding study was done to observe the extent to which interviewers were following the structured interview and the extent to which it became necessary for them to give unstructured additional explanation or comments to respondents. As part of the study, 100 interviews were taped and then coded on a variety of behavioral dimensions. This analysis revealed that, on the whole, the interview proceeded in a very structured manner, with 85 percent of all question and answer "dyads" being "asked and answered only." Additional unstructured interaction/discussion took place most frequently for those questions in which there was some ambiguity in the topic. In most cases this interaction was judged to have facilitated obtaining the correct response.
For both survey cycles, results from the quality studies were used to identify those questionnaire items that might need additional revision for the next study cycle. Debriefing sessions concerning the survey were held with interviewers, and this information was also used in revising the survey for the next cycle.
A word of caution needs to be given concerning comparisons with previous NSRCG results. During the 1993 cycle, the SESTAT system underwent considerable revision in several areas, including survey eligibility, data collection procedures, questionnaire content and wording, and data coding and editing procedures. The changes made for the 1995 through 1999 cycles were less significant but might affect some data trend analysis. While the 1993 through 1999 survey data are fairly comparable, care must be taken when comparing results from the 1990s surveys to surveys from the 1980s, due to the significant changes made in 1993. For a detailed discussion of these changes, please see the 1993, 1995, 1997, and 1999 NSRCG methodology reports.
For the 1999 NSRCG, there were no significant procedural changes that would affect the comparison of results between the 1997 and 1999 survey cycles.
The National Center for Education Statistics (NCES) conducts a survey of the nation's postsecondary institutions, called the Integrated Postsecondary Education Data System (IPEDS). The IPEDS Completions Survey reports on the number of degrees awarded by all major fields of study, along with estimates by gender and race/ethnicity.
Although both the NSRCG and IPEDS are surveys of postsecondary education and both report on completions from those institutions, there are important differences in the target populations for the two surveys that directly affect the estimates of the number of graduates. The reason for the different target populations is that the goals of the surveys are not the same. The IPEDS estimates of degrees awarded are intended to measure the output of the educational system. The NSRCG estimates are intended to measure the supply and utilization of a portion of graduates in the years following their completion of degrees. These goals result in definitions of the target population that are not completely consistent for the two surveys. Other differences between the estimates can be explained to a very large extent by a few important aspects of the design or reporting procedures in the two surveys. The main differences between the two studies that affect comparisons of estimates overall and by race/ethnicity are listed below.
Despite these factors, the NSRCG and IPEDS estimates are consistent when appropriate adjustments for these differences are made. For example, the proportional distributions of graduates by field of study are nearly identical, and the numerical estimates are similar. Further information on the comparison of NSRCG and IPEDS estimates is available in the report, A Comparison of Estimates in the NSRCG and IPEDS, available in the SRS website, at http://www.nsf.gov/statistics/.
The following definitions are provided to facilitate the reader's use of the data in this report.
Major field of study: Derived from the survey major field category most closely related to the respondent's degree field. Exhibit 1 gives a listing of the detailed major field codes used in the survey. Exhibit 2 gives a listing of the summary major field codes developed by NSF and used in the tables. The appendix lists the eligible and ineligible major fields within each summary category.
Occupation: Occupation is derived from the survey job list category most closely related to the respondent's primary job. Exhibit 3 gives a listing of the detailed job codes used in the survey, and Exhibit 4 gives the summary occupation codes developed by NSF and used in the tables.
Labor force: The labor force includes individuals working full or part time as well as those not working but seeking work or on layoff. It is a sum of the employed and the unemployed.
Unemployed: The unemployed are those who were not working on April 15 and were seeking work or on layoff from a job.
Type of employer: Type of employer is the sector of employment in which the respondent was working on his or her primary job held during the week of April 15, 1999. The following are the definitions for each of these categories. Private industry and business includes all private for-profit and private not-for-profit companies, businesses, and organizations, except those reported as educational institutions. It also includes persons reporting that they were self-employed. Educational institutions include elementary and secondary schools, 2-year and 4-year colleges and universities, medical schools, university-affiliated research organizations, and all other educational institutions. Government includes local, state, and Federal Government; military; and commissioned corps.
Primary work activity: Primary work refers to the activity that occupied the most time on the respondent's job. In reporting the data, those who reported applied research, basic research, development, or design work were grouped together in "research and development (R&D)." Those who reported accounting, finance or contracts, employee relations, quality or productivity management, sales and marketing, or managing and supervising were grouped into "management, sales, administration." Those who reported production, operations, maintenance, professional services or other activities were given the code "other."
Full-time salary: Full-time salary is the annual salary for the full-time employed, defined as those who were not self-employed (either incorporated or not incorporated), whose principal job was not less than 35 hours per week, and who were not full-time students on the reference date (April 15, 1999). Graduates who did not receive salaries were asked to report earned income, excluding business expenses. To annualize salary, reported hourly salaries were multiplied by the reported number of hours paid per week, then multiplied by 52; reported weekly salaries were multiplied by 52; reported monthly salaries were multiplied by 12. Yearly and academic yearly salaries were left as reported.
Race/ethnicity: All graduates, both U.S. citizens and non-U.S. citizens, are included in the race/ethnicity data presented in this report. In tables with sufficient sample size, race/ethnicity data are presented by the specific categories of white, non-Hispanic; black, non-Hispanic; Hispanic; Asian or Pacific Islander; and American Indian or Alaskan Native. In tables where the sample size is not sufficient to present data by specific category, the groups of black, Hispanic, and American Indian or Alaskan Native are combined into the underrepresented minority category.
The tables in this report present information for two groups of recent graduates. The first of these groups consists of persons who earned bachelor's degrees in S&E fields from U.S. institutions during academic years 1997 and 1998. The second group includes those who earned S&E master's degrees during the same two years.
|Table||Text Table Name|
|1||Major fields and corresponding sampling rates, by cohort and degrees: April 1999||.xls|
|2||Number of graduates, unweighted graduate response rates, and weighted graduate response rates, by graduate characteristics: April 1999||.xls|
|3||Unweighted number, weighted estimate, and standard errors for 1997 and 1998 science and engineering bachelor's degree recipients, by graduate characteristics: April 1999||.xls|
|4||Unweighted number, weighted estimate, and standard errors for 1997 and 1998 science and engineering graduate degree recipients, by graduate characteristics: April 1999||.xls|
|5||Estimated parameters for computing generalized variances for estimates from the 1997 NSRCG||.xls|