Font Size:
|
||||||
MEPS HC-218: 2020 Jobs FileMarch 2022 Due to the COVID-19 pandemic, 2020 data collection moved primarily to phone rather than in-person. This posed a challenge in Panel 25 Round 1, which is difficult to start via phone, resulting in a low response rate. To balance this and increase the number of completes to be comparable to previous years, Panels 23 and 24 were extended to nine rounds of data collection. Phone data collection and the challenges of the pandemic present concerns about data quality. Please take this into consideration when comparing to or pooling with previous years.
Agency for Healthcare Research and Quality
A. Data Use Agreement A. Data Use AgreementIndividual identifiers have been removed from the micro-data contained in these files. Nevertheless, under Sections 308 (d) and 903 (c) of the Public Health Service Act (42 U.S.C. 242m and 42 U.S.C. 299 a-1), data collected by the Agency for Healthcare Research and Quality (AHRQ) and/or the National Center for Health Statistics (NCHS) may not be used for any purpose other than for the purpose for which they were supplied; any effort to determine the identity of any reported cases is prohibited by law. Therefore in accordance with the above referenced Federal Statute, it is understood that:
By using these data you signify your agreement to comply with the above stated statutorily based requirements with the knowledge that deliberately making a false statement in any matter within the jurisdiction of any department or agency of the Federal Government violates Title 18 part 1 Chapter 47 Section 1001 and is punishable by a fine of up to $10,000 or up to 5 years in prison. The Agency for Healthcare Research and Quality requests that users cite AHRQ and the Medical Expenditure Panel Survey as the data source in any publications or research based upon these data. B. Background1.0 Household ComponentThe Medical Expenditure Panel Survey (MEPS) provides nationally representative estimates of health care use, expenditures, sources of payment, and health insurance coverage for the U.S. civilian noninstitutionalized population. The MEPS Household Component (HC) also provides estimates of respondents’ health status, demographic and socio-economic characteristics, employment, access to care, and satisfaction with health care. Estimates can be produced for individuals, families, and selected population subgroups. The panel design of the survey, which typically includes five rounds of interviews covering two full calendar years. In 2020, in order to increase the number of completed interviews, the panel design has been extended to include seven rounds of interviews covering three full calendar years. The panel design of MEPS provides data for examining person level changes in selected variables such as expenditures, health insurance coverage, and health status. Using computer assisted personal interviewing (CAPI) technology, information about each household member is collected, and the survey builds on this information from interview to interview. All data for a sampled household are reported by a single household respondent. The MEPS HC was initiated in 1996. Each year a new panel of sample households is selected. Because the data collected are comparable to those from earlier medical expenditure surveys conducted in 1977 and 1987, it is possible to analyze long-term trends. Each annual MEPS HC sample size is about 15,000 households. Data can be analyzed at either the person or event level. Data must be weighted to produce national estimates. The set of households selected for each panel of the MEPS HC is a subsample of households participating in the previous year’s National Health Interview Survey (NHIS) conducted by the National Center for Health Statistics. The NHIS sampling frame provides a nationally representative sample of the U.S. civilian noninstitutionalized. In 2006, the NHIS implemented a new sample design, which included Asian persons in addition to households with Black and Hispanic persons in the oversampling of minority populations. NHIS introduced a new sample design in 2016 that discontinued oversampling of these minority groups. 2.0 Medical Provider ComponentUpon completion of the household CAPI interview and obtaining permission from the household survey respondents, a sample of medical providers are contacted by telephone to obtain information that household respondents cannot accurately provide. This part of the MEPS is called the Medical Provider Component (MPC) and information is collected on dates of visits, diagnosis and procedure codes, charges and payments. The Pharmacy Component (PC), a subcomponent of the MPC, does not collect charges or diagnosis and procedure codes but does collect drug detail information, including National Drug Code (NDC) and medicine name, as well as amounts of payment. The MPC is not designed to yield national estimates. It is primarily used as an imputation source to supplement/replace household reported expenditure information. 3.0 Survey Management and Data CollectionMEPS HC and MPC data are collected under the authority of the Public Health Service Act. Data are collected under contract with Westat, Inc. (MEPS HC) and Research Triangle Institute (MEPS MPC). Data sets and summary statistics are edited and published in accordance with the confidentiality provisions of the Public Health Service Act and the Privacy Act. The National Center for Health Statistics (NCHS) provides consultation and technical assistance. As soon as data collection and editing are completed, the MEPS survey data are released to the public in staged releases of micro data files and tables via the MEPS website. Additional information on MEPS is available from the MEPS project manager or the MEPS public use data manager at the Center for Financing, Access, and Cost Trends, Agency for Healthcare Research and Quality, 5600 Fishers Lane Rockville, MD 20857 (301-427-1406). C. Technical and Programming InformationSection C of this document offers a brief overview of the data provided in MEPS public use release HC-218, as well as the content and structure of the codebook, reserved code values, and variable naming conventions. It is followed by Section D containing the Variable-Source Crosswalk, Appendix 1 containing sample SAS program code, and Appendix 2 containing sample Stata program code. A copy of the survey instrument used to collect the information on this file is available on the MEPS website. 1.0 General InformationThis file is being released as a research file and has undergone the standard quality control procedures usually performed on MEPS data files. The file includes a total of 47,776 records, with each record representing a unique job for a person by round. This file presents information about jobs starting on or before 12/31/2020 only. The 2021 Jobs file release will present information on Panel 23, Panel 24, and Panel 25 jobs starting in 2021. In the Employment section, MEPS collects complete job-related information in the round in which a job is first reported. While they vary by job type (see Section 2.0), the data reported for a job in its first survey round may include earnings by type (gross salary, tips, etc.), start and stop dates, hours and weeks worked, establishment size and industry, occupation, presence of retirement and other benefits, self-employment versus other status, temporary or seasonal situations, and health insurance availability. Minimal data updates are available for later rounds in which the job continues. This Full-Year Jobs file contains job records from three MEPS panels. The 2020 Jobs file provided in this release, MEPS HC-218, contains job-level information collected in Rounds 5 through 7 for Panel 23, Rounds 3 through 5 for Panel 24, and Rounds 1 through 3 for Panel 25 of the Medical Expenditure Panel Survey. In order to obtain complete information for a job, users must note the round in which the job is first reported. This is because MEPS collects complete Jobs information in that round only, as noted above. For the first year panel, in this case Panel 25, data from Rounds 1, 2, and 3 are included in the 2020 Jobs file. Complete information for any Panel 25 job is available, whether that job was first reported in Round 1, 2, or 3. This is the case for any first year panel (the panel that began its first year of interviewing in the given year) in a Full-Year Jobs file. For the second year panel (the panel that continued with its second year of interviewing in the given year), in this case Panel 24, data from Rounds 3, 4, and 5 are included in this file. If the Round 3, 4, or 5 job continued from Round 1 or Round 2, users must look back to the Jobs file from the previous year (2019) to obtain complete information for the job. For the third year panel (the panel that continued with its third year of interviewing in the given year), in this case Panel 23, data from Rounds 5, 6, and 7 are included in this file. If the Round 5, 6, or 7 job continued from Round 3 or Round 4, users must look back to the Jobs file from the previous year (2019) to obtain complete information for the job. If the Round 5, 6, or 7 job continued from Round 1 or Round 2, users must look back to the Jobs file from two years prior (2018) to obtain complete information for the job. Note that in 2020, Panel 23 is a non-traditional cross-over year panel. Panel 23 Round 5 was collected as a traditional terminal panel, the Round 5 interview occurred in 2020 but questions are asked as of 12/31/2019. A traditional cross-over panel (like Panel 23 Round 3) would have the interview in 2020 and ask questions as of the interview date. To compensate, the new Panel 23 Round 6 interview asked questions as of 1/1/2020 (looking further back than the Round 5 interview date). As detailed below, in the Jobs file, adjustments were made so that the file could identify which jobs (and job characteristics) were relevant in the time period between 1/1/2020 and the Panel 23 Round 5 interview date. This allowed for the creation of Panel 23 Round 5 job records, and sometimes resulted in the deletion of Panel 23 Round 6 records (for example, if the Round 6 interview identified a job that ended between 1/1/2020 and Round 5 interview date). More details on this process will follow in Section 1.1. Appendix 1 includes sample SAS code and Appendix 2 contains sample Stata code to assist users in obtaining this information. Users should note that, because of differences in sample composition between the current year and the previous year files (i.e., a person was included in the previous year’s delivery but not the current year or vice versa), or because more accurate information was received in subsequent round comments following the delivery of the Jobs records in the previous year, there occasionally may not be a corresponding job in the previous year file. Users should note that due to the job adjustment process, Panel 23 Round 5 records contained in the 2020 Jobs file may differ from those contained in the 2019 Jobs file. Specifically, Panel 23 Round 5 job records in the 2019 Jobs file represent jobs (and job characteristics) that occurred between the Round 4 interview and 12/31/2019. In the 2020 Jobs file, Panel 23 Round 5 jobs represent jobs (and job characteristics) that occurred between 1/1/2020 and the Panel 23 Round 5 interview date. Job details for the 2020 portion of Round 5 were collected using a combination of data collected at the Panel 23 Round 5 and Round 6 interviews. 1.1 Third Panel in the 2020 Jobs FileDue to the impact of the COVID-19 pandemic on MEPS collection methods and lower response rates in the 2020 calendar year, AHRQ extended fielding for Panel 23 persons to include two additional rounds, Round 6 and Round 7, that collected information about 2020. As a result, the 2020 MEPS includes three panels of data: Panel 25 Rounds 1, 2, and 3, Panel 24 Rounds 3, 4, and 5, and Panel 23 Rounds 5, 6 and 7. The decision to extend Panel 23 to include additional rounds occurred after Panel 23 Round 5 was fielded. As a result, Round 5 was fielded as a traditional terminal round (where questions asked for information through 12/31/2019. Had the decision been made much earlier, Round 5 would have been fielded as a cross-year round (similar to Round 3), where questions would have covered the events through 2020 up to the Round 5 interview date. Given this impromptu design decision, in order to capture data after 12/31/2019, the Panel 23 Round 6 reference period began on January 1, 2020 (not the Round 5 interview date) and ended anywhere from July through October 2020. This duration (7 to 10 months) exceeds a standard round length, which averages 6 months. In order to create consistency in employment measures over time, job rosters and employment status were adjusted for FY 2020 public use files so that (a) Round 5 would represent employment for the time period between 1/1/2020 and the Round 5 interview date and (b) Round 6 would represent employment for the time period between the Round 5 interview date and the Round 6 interview date. Appendix 5 also discusses additional MEPS rounds and extended fielding for Panel 23. The adjustment lengthens Round 5 to resemble a standard Round 3 cross-year reference period and shortens Round 6 to resemble a standard Round 4 field period, as illustrated below. The adjustment process is further discussed after the description of the file contents and throughout this document. It is important to note that CAPI collected employment status for Panel 25 Round 3, Panel 24 Round 5, and Panel 23 Round 7 as of the interview date in 2021. As a result, adjustments discussed here for third year Panel 23, Round 5 will not be necessary in future data years.
Special 2020 Panel 23 Round 5 and Round 6 Job Roster Adjustment Persons Eligible for Job Roster Adjustment Any person who was part of a household in Round 6 and who had one or more Round 6 jobs that began or ended between January 1, 2020 and the Round 5 interview date was eligible for adjustment. In addition, a roster was adjusted if a person became 16 years of age between January 1, 2020 and the Round 5 interview date. Twenty-three persons had their age change from 15 to 16 and nine had a job requiring adjustment. Number of Jobs and Persons Involved in Special Processing Job rosters for 610 persons were evaluated for adjustment. These persons had 711 job records whose start and/or stop date was on or before the Round 5 interview date. Some persons had more than one job flagged for review. Of the 610 persons, 188 persons (with 289 job records) either had multiple jobs requiring adjustment or complex Round 5 and Round 6 rosters; these were evaluated and adjusted (described below). The remaining 422 job records were programmatically adjusted based on CAPI flow. Job Roster Adjustments The following describes rules for selecting a job for adjustment and the typical roster adjustment made:
Roster adjustment: the Round 5 current job became a Round 5 former job and the Round 6 job was deleted. In some cases, Round 7 job-level variables, such as CREATEQ, ORIGRND, and TYPECHGD, required adjustment. Adjusted Job Flag The constructed variable ADJR5 has been added to the 2020 Jobs file to flag those Panel 23 Round 5 and Round 6 jobs that were adjusted or created based on the interval between 1/1/2020 and the Round 5 interview data.
ADJR5 is set to a value of 1, 2 or 3 for all Panel 23 Round 5 jobs. ADJR5 is set to a value of 4 for Panel 23 Round 6 jobs that were part of the adjustment process. Otherwise, ADJR5 is set to “Inapplicable” (-1). 2.0 Data File Information2.1 File ContentsEach record in the 2020 Jobs file represents one job reported by a person in a round. All persons age 16 and older in the MEPS are asked to report on jobs held. Depending on an individual’s job history, these reported jobs may be held:
Only those persons reporting a job in a round will have a record in the 2020 Jobs file. Record Identifiers The unique record identifier is the variable JOBSIDX, which is composed of a person identifier (DUID + PID), a round identifier (RN), and a job number (JOBNUM). A panel indicator (PANEL) is included on the file to distinguish Round 3 jobs held by Panel 24 persons from Round 3 jobs held by those in Panel 25, as well as Round 5 jobs held by Panel 23 persons from Round 5 jobs held by those in Panel 24. The DUID identifier in this data release includes a 2-digit code to identify the panel and, as a result, JOBSIDX includes a panel identifier via DUID. The variable OrigRnd indicates the round a job was created. Therefore, it may or may not contain the same value as RN. Each identifier variable (JOBSIDX, DUID, DUPERSID) begins with the 2-digit panel number. This allows analysts to easily identify records delivered in a previous year Jobs file (when panel is used in conjunction with other variables, such as RN and OrigRnd). In addition, CAPI assigns a unique job number that may not be used in subsequent rounds on different jobs. This 3-byte number, JOBNUM, is unique to the reporting unit (RU) and is set to a value that corresponds with the RU in which a person’s job was first reported (e.g. A RU is ‘1’, B RU is ‘2’, C RU is ‘3’, etc). Initial Reporting Round Most persons held only one job at the first interview date – their “Current Main Job.” For persons who held more than one job at the round’s interview date (a current job), respondents were asked to identify the main job. This job was classified as the “Current Main Job” and any other simultaneously held job was classified as a “Current Miscellaneous Job.” The MEPS also obtained some information on any former job (Former Main Job or Former Miscellaneous Job) held in the reference period but not at the interview date. For those persons neither working at the interview date nor earlier in the reference period, limited information on the last job the person held was collected. Additionally, for those persons age 55 or older who were identified as having retired from a job, the MEPS obtained some job-level information (Retirement Job). The variable SUBTYPE indicates the type of job record - current main (1), current miscellaneous (2), former main (3), former miscellaneous (4), last job outside reference period (5), or retirement job (6). When a job is initially reported, MEPS asks for detailed information about any “Current Main Job” and basic information about other job types. Refer to the questionnaire to see which information was asked for each job type. The following variable list identifies when a variable could be set based on the job SUBTYPE. Self-employed and wage-earner status at a job also defines when a variable may be set. (Note: wage-earner is used to describe workers who are not self-employed.) The last column indicates if the variable is populated in the round in which the job is first reported (collection only), when the job is reviewed (review only), or both (collection and review).
For last jobs outside of reference period and retirement jobs that ended more than two years prior to the beginning of the reference period, certain questions (HHMEMBER_M18 and TOTLEMP_M18) are not asked. The precise calculation of the two-year cut-off date is not possible for some persons due to allowed negative values on stop year, stop month, and reference period start month. Therefore, HHMEMBER_M18 and TOTLEMP_M18 may be collected for some jobs that ended more than two years prior to the reference period. Skip Patterns Due to many skip patterns, it is recommended that users of the 2020 Jobs file become familiar with the Employment section in the MEPS questionnaire. To aid users, a crosswalk between variables and MEPS questionnaire numbers is provided in this release. The following examples of variables involved in skip patterns are presented to be illustrative; these examples do not represent the full range of variables affected by questionnaire skip patterns. In one example of a skip pattern, the MEPS does not obtain job-related benefits such as vacation, sick leave, and pension information for self-employed jobs, so those variables are coded as “Inapplicable” (-1) for those types of jobs. Nor does the MEPS attempt to obtain wage, salary, and information regarding whether the job was in the private sector, federal or local government (TYPEEMPL) for the self-employed. So again, due to the skip pattern, TYPEEMPL is coded as “Inapplicable” (-1) for self-employed jobs. Conversely, the questions relating to business organization type (BUSINC, PROPRIET) are asked only of the self-employed, so the skip pattern results in those variables being coded as “Inapplicable” (-1) for jobs performed by wage earners. Job Updates and “Inapplicable” (-1) Values The MEPS used dependent interviewing in Rounds 5, 6, and 7 for Panel 23, Rounds 3, 4, and 5 for Panel 24 and in Rounds 2 and 3 for Panel 25 (see Section RJ in the Employment section of the questionnaire). In these rounds, the MEPS asked about current main and current miscellaneous jobs held at the previous round interview date to determine whether the jobholder continued to work at these jobs. For other job types (former, last, or retirement) reported in the previous round, MEPS does not ask any follow-up questions. These jobs, by definition, are no longer held by the person and therefore are not included on the file except in the round they are first reported. With dependent interviewing, if a person still held a Current Main Job from the previous round, the MEPS asked whether the job was still the main job. For most jobholders, it was reported that they still worked at the same job and it was still their main job. If, in a subsequent interview, a job was no longer held, it was designated as a former job for that follow-up round. It is also possible, although unusual, for a job to change from main to miscellaneous (or vice versa) in a round subsequent to the initial report. If job status remained the same for a continuing job (either main or miscellaneous), the MEPS asked only a subset of the employment questions as a review. Because the MEPS asked only this subset of questions if job status for a person did not change in later rounds, many job-level variables on the subsequent round’s job records are coded as “Inapplicable” (-1); the complete information for a continued job is located on the record for the job in the first round in which it was reported. Thus, it is important to determine whether a job continues from the previous round when working with the job records. In rounds where this applies, the variables STILLAT (for jobs that were current main in the previous round) and STILLWRK (for jobs that were current miscellaneous in the previous round) indicate whether a person still holds the job at the subsequent round interview date. The variable SUBTYPE on the subsequent round record indicates whether the job is main or miscellaneous in that subsequent round. Note that if a Panel 24 job included in this 2020 file is continued from a Round 1 or 2 job or a Panel 23 job included in this 2020 file is continued from a Round 3 or 4 job (in the 2019 file), much of the information will be contained in the 2019 Jobs file (HC-211). Likewise, if a Panel 23 job included in this 2020 file is continued from a Round 1 or Round 2 job (in the 2018 file), much of the information will be contained in the 2018 Jobs file (HC-203).Use that file to obtain the desired job characteristics. Appendix 1 provides a sample SAS program showing how to do this, and Appendix 2 provides a sample Stata program showing how to do this. Both sample programs take into account the addition of the third panel. Any new job reported in a round following the initial interview is collected the same way as in the first interview round. Variables that relate only to the review of a job reported in a previous round (DIFFWAGE, ESTBTHRU, INSESTB, MAIN_JOB, NOWTAKEI, OFFTAKEI, STILLAT, StillWorkFTPT, STILLWRK, RvwTotNumEmp, WHY_LEFT_M18, WhyChngPTToFT, WhyChngFTToPT) were not asked in Round 1, and these variables are coded as “Inapplicable” (-1) on a Jobs record for the round in which the job is initially reported. Another type of job update pertains to situations where a reviewed current miscellaneous job becomes the current main job in the round. The flag variable TYPECHGD indicates if a job changed from a current miscellaneous job to a current main job. For these types of jobs, questions asked when the job was first reported as a current miscellaneous job are not re-asked, with three exceptions.
Exceptions to the “Inapplicable” (-1) Rule Unlike the situation explained above for most variables on the file, for certain variables a value other than “Inapplicable” (-1) does not necessarily mean that a job is newly reported. For a small subset of variables, previous round variables are carried forward to the next round, even if there have been no updates to the variables since they were originally reported. There are two distinct situations in which this special treatment is used, due to internal processing needs. The first type of exception occurs when questions related to the affected variables are skipped over as “Inapplicable” (-1) during the interview in rounds subsequent to the one in which the job was initially reported, but have their originally reported response carried forward from round to round. This group includes the following 14 variables: EMPLINS, HRSPRWK, HRS35WK, JOBTYPE, JSTRTY, JSTRTM, MORELOC, NUMEMPS, OFFRDINS, PROVDINS, TYPEEMPL, JOBHASHI, HRSALBAS, and RETIRJOB. Note that HRSALBAS and RETIRJOB may also be updated in subsequent rounds. The second type of exception occurs for certain questions that are asked during the review of a job in rounds following the round in which the job was initially reported. If there is no change based on the review, the value for the affected variable is copied forward from the previous round. If there is a change, the variable is updated to reflect the new information. These five variables are: JSTOPY, NOWTAKEI, OFFTAKEI, SUBTYPE, and TOTLEMP_M18. Variables related to earnings (such as HRLYWAGE, GROSSPAY, SALARIED) are treated similarly to the five variables just discussed. In the review section, the MEPS attempted to obtain information regarding changes in wages for the same job from round to round. If there were no wage changes (indicated by the DIFFWAGE variable), then the most recent round’s information was carried forward. If changes were recorded, then the relevant variables were updated. For every new job reported for a person, the MEPS attempted to obtain current wage information. Top-Coding, Bottom-Coding, Editing, and Confidentiality Outlier Wage Editing on Current Main Jobs In most years, wage information on current main job records is logically edited for consistency using established rules and guidance from AHRQ. However, in 2020 the typical wage outlier editing process was not performed. In typical years, outliers are checked for persons who report a wage change and the new reported wage (a) is substantially different from prior wage (change >=100%), (b) is no different than prior wage, (c) is low in value ($0 < wage < $1) or, (d) has a value higher than prior year’s top code value. There are numerous sources for these types of errors, including keystroke or respondent error. In a typical year, approximately 100 wages are reviewed per panel, resulting in approximately 50 wage edits (overall). Users should keep in mind that such edits were not performed in 2020 when using the wage variables, especially when comparing 2020 to other data years. To help users identify cases that would have been reviewed (but not necessarily edited) in this process, the 2020 Population Characteristics Public Use file data includes wage outlier flag variables, OUTFLAGrr. These round-specific wage outlier flag variables OUTFLAG31, OUTFLAG42, and OUTFLAG53 indicate that a person’s updated wage at the current main job would have been programmatically selected for review during the 2020 wage outlier editing process (but not necessarily edited). The flag is constructed prior to wage imputation, consistent with the timing of full year wage outlier review. A wage is selected for review if one (and only one) of the following is true:
Sometimes, a person’s wages can meet several conditions. In some cases, the OUTFLAG31, OUTFLAG42, and OUTFLAG53 will be set to a value that represents a combination, for example, values 5 or 6:
In other cases those cases are given a default value based on whether they are a high wage or low wage. Specifically, in rare circumstances, a person may have a mix of “updated” wage with no difference in calculated wage amount (category 3 – no wage change) with either of the following categories: low wage (category 1 - low) or high wage (category 4 - high). OUTFLAG is set to either “Low wage outlier” (1) or “High wage outlier” (4) respectively for these cases.
Users should also keep in mind that many of the high wage outlier amounts identified in OUTFLAG have wages that are subsequently top coded, as described below. Note that the wage outlier flag is included on the 2020 Population Characteristics Public Use file only. Users of the 2020 Jobs file can link to the 2020 Population Characteristics Public Use file through the variable DUPERSID, as described in Section 2.3 (Person-Level Estimates). Wage Top-Coding For reasons of confidentiality, earnings variables on the 2020 Jobs file were top-coded. The earnings variables include HRLYWAGE, BONSAMT, COMMAMT, TIPSAMT, DAYWAGE, WKLYAMT, GROSSPAY, and MAKEAMT. A value of “TOP CODED” (-10) for one of these variables on a record indicates that the variable had a positive value and that the hourly rate for that earnings variable for the record was greater than or equal to $103.36. The process by which the top-code value for the Jobs file is derived incorporates the wage top-code process used in the 2020 Population Characteristics Public Use file top-coding process. The purpose of top-coding is to ensure confidentiality for each person across files. In addition to using wages from the first report of a current main job, updated wages from that job reported in any subsequent round are also included in deriving the wage top-code value. On the 2020 Population Characteristics Public Use file, any person who has a wage for any job in any round that is greater than or equal to the top-code value will have all wages for all jobs top-coded, regardless of round. Any person whose wages are top-coded on the 2020 Population Characteristics Public Use file also has all wages on all jobs top-coded in the 2020 Jobs file. Moreover, because other jobs where wages are reported are included in the 2020 Jobs file but not summarized in the 2020 Population Characteristics Public Use file (i.e., newly reported former main jobs and current/former miscellaneous jobs), and these wages may exceed the current year top-code value, wages for these jobs and all jobs belonging to the same jobholder are top-coded on the 2020 Jobs file. In turn, the wages of these persons are top-coded in the 2020 Population Characteristics Public Use file as well. Note that there are also some jobs where respondents indicate that a supplemental wage, such as a commission, tip, or bonus, is greater than or equal to the wage top-code value but, at that same job, base wage such as the annual salary is not. For these cases, only the tips, commissions, or bonus amounts were top-coded on the job where they are greater than or equal to the wage top-code value (note, these supplemental wages only reside on the 2020 Jobs file). All other wage amounts for all jobs for these persons were left as reported. (This applies to wages and jobs on both the 2020 Population Characteristics Public Use and 2020 Jobs files.) Beginning in 2020, wages are also top-coded to -10 on the Jobs file for two situations where wages were formerly reset to “Cannot Be Computed” (-15) in prior year Jobs files. These situations are:
Additional Wage Information To improve the quality of wage reports, CAPI prompts the respondent to confirm wages reported in the Employment Wage section if a wage amount falls outside a specified wage range.
To calculate the hourly rate for earnings types not reported on an hourly basis, the number of hours per week worked and in some cases the number of weeks worked were used in conjunction with the various amounts. These hours and weeks are included on the file along with the reported earnings amounts, but not the calculated hourly rates. (Earnings variables were not reconciled with income data collected elsewhere in the MEPS.) Establishment Size Information The establishment size variable for the self-employed is TOTLEMP_M18. In addition, two variables are available containing the individual responses collected at RJ110 and EM740 (number of employees at a self-employed job). They are RvwTotNumEmp (establishment size at continuing self-employed job) and TotNumEmp (establishment size at newly reported self-employed job), respectively. The establishment size for wage-earners can be found in NUMEMPS (establishment size at non-self-employed job); this value is collected at EM430 (number of employees). Respondents who did not know the actual establishment size (NUMEMPS) are asked in question EM440 to choose approximate establishment size from a number of size ranges. These responses are used to create the variable ESTMATE1_M19. The categorical values available to respondents in EM440 are as follows.
The value “Cannot Be Computed” (-15) is not an allowed value for ESTMATE1_M19. In 2019, the constructed variable ESTMATE1_M18 contained a second coding schema for categorical values available to respondents in EM440 prior to Panel 24 Round 3 and Panel 23 Round 5. As of 2020, ESTMATE1_M18 is no longer delivered in this file. For confidentiality reasons, NUMEMPS, TOTLEMP_M18, RvwTotNumEmp and TotNumEmp were top coded to “-10 # OF EMP >= 13,000” for establishment sizes greater than or equal to 13,000 employees. Job Start/Stop Year In addition to top coding wages and establishment size, the start year of job (JSTRTY) and the stop year of job (JSTOPY) are bottom-coded. This is done because a person’s age may be calculated using the job start or stop year and that age may indicate that the jobholder is older than 85 years, the age top-code value. This value is calculated by taking the current delivery year (e.g. 2020), subtracting the age top-code value (i.e. 85 years of age), then adding back 15 (i.e. the age of a person in the year before entering the work force as defined in MEPS). For the 2020 Jobs file, the bottom code value for the job start and stop year on jobs reported in Panel 25 Round 1, Round 2, or Round 3, Panel 24 Round 4 or Round 5, and Panel 23 Round 6 or Round 7 is 1950. For Jobs that began in Panel 24 Round 1, Round 2 or Round 3, or Panel 23 Round 4 or Round 5, JOBSIDX were delivered in the 2019 Jobs file. These records may retain the 2019 bottom code value of 1949. For Jobs that began in Panel 23 Round 1, Round 2 or Round 3, JOBSIDX were delivered in the 2018 Jobs file. These records may retain the 2018 bottom code value of 1948. Temporary and Seasonal Jobs Two variables on the file pertain to the temporary and seasonal nature of a person’s main or miscellaneous job. The variable TEMPJOB indicates whether a main or miscellaneous job is temporary (i.e., is a current main job for a limited amount of time or until the completion of a project). The variable SESNLJOB indicates either that a main or miscellaneous job is available only during certain times of the year or that the individual is working throughout the entire year at that job. Teachers and other school personnel who work only during the school year are considered to work year round. These questions are asked of newly reported jobs only. These variables are set to “-1 INAPPLICABLE” for all subsequent rounds. These questions are not asked of newly reported former miscellaneous jobs, last jobs outside of reference period, and retirement jobs. Reason No Longer at Place of Employment In cases where a former job is newly reported, questions are asked regarding why the person is no longer at that place of work. For wage earners, this information is found in YLEFT_M18. For self-employed persons, this information is collected in YNOBUSN_M18. It is important to note that the retirement job classification in the variable SUBTYPE is independent of any retirement response in the following variables:
Health Insurance Data Questions about employment-related health insurance are asked both when any type of job is newly reported and when any continuing job is reviewed. For main jobs, either newly reported or changing from miscellaneous, the variable that indicates whether insurance is held through that establishment is EMPLINS. For all non-main jobs, the variable JOBHASHI indicates whether insurance is held through that establishment. For a newly reported job, depending on whether employment-related insurance is held or not, there may be follow-up information gathered which is contained in the following variables:
For a continuing job, when no health insurance was held through the job in the round in which the job was first reported but health insurance was offered through the job, the question RJ70 OFFTAKEI is asked in later rounds to determine whether the employee now holds the health insurance that is offered through the job. (Note: if health insurance through this job was reported as being held via RJ70 in the prior round, RJ70 is not asked in the current round.) Similarly, the insurance status question RJ80 (NOWTAKEI) is asked to determine whether health insurance is now held through the job in the following cases:
MEPS then includes several clarifying questions regarding health insurance availability at an employer. Where the person does not report, does not know, or refuses to indicate the insurance coverage status through the job at RJ70 or reports no insurance coverage through the job at RJ80, the respondent is asked if the person was offered insurance through the job at RJ90 (ESTBTHRU). Lastly, when a respondent indicates that the jobholder of a reviewed job neither holds insurance through the job nor was offered health insurance at the job, the respondent is asked if any other employees were offered health insurance through the job at RJ100 (INSESTB). In some cases, respondents will indicate in the Health Insurance section that health insurance reported in the Employment section was either wholly or partially reported in error. This is referred to as insurance being “disavowed.” If newly reported health insurance through the job is disavowed in the Health Insurance section, follow-up questions (HX21, HX22, HX23) regarding whether health insurance is offered at the job, whether more than one plan is available, and whether health insurance is offered to any employees are asked in the Health Insurance section. This information is used in an editing process whereby responses in the Health Insurance section are transferred into the Employment or Review of Jobs sections. As a result, the disavowal process may result in a change to values originally collected in the Employment or Review of Jobs section (wherever the health insurance was initially reported). The complete list of variables potentially impacted includes: EMPLINS, JOBHASHI, OFFRDINS, DIFFPLNS, ANYINS, and PROVDINS, collected in the Employment section, and NOWTAKEI, OFFTAKEI, ESTBTHRU, and INSESTB, collected in the Review of Jobs section. In some cases, a disavowal may result only in a change to the value of PROVDINS. Health insurance through an employer can be disavowed in MEPS based on a respondent’s answer to one of four questions (HX14, HX15, HX20, HP70). To help users understand the source of the disavowal, the variable HIDISAVW indicates which of the following questions resulted in the disavowal. HIDISAVW will include only one source among these options. Please note, however, that it is possible for a respondent to disavow one source of coverage at HX15 and then later disavow the second source of coverage at HP70. In these cases, HIDISAVW will be set to HP70.
For Panel 23 persons who underwent adjustment of Round 5 and/or Round 6 jobs variables, health insurance variables EMPLINS, JOBHASHI, OFFTAKEI, and NOWTAKEI reflect insurance status as of the adjusted end reference period date. Industry and Occupation Coding Industry and occupation codes were assigned by professional coders at the Census Bureau based on verbatim descriptions provided by respondents during the survey interview. The codes are determined at a detailed 4-digit level and then collapsed into broader groups on the file to ensure the confidentiality of the records. INDCODEX contains industry information and OCCCODEX contains occupation information. Appendices 3 and 4 contain crosswalks between the detailed and collapsed codes for industry and occupation. With the 2010 file, the Census Bureau began using 2007 Industry and 2010 Occupation codes, which were developed for the Bureau’s Current Population Survey and American Community Survey. These updated coding schemes incorporate minor changes from the 2003 industry and occupation codes used for the 2002-2009 files; therefore, INDCODEX and OCCCODEX for 2010 and later files will be comparable to those variables on the 2002-2009 files. (Industry and occupation variables for pre-2002 files are not comparable to those for later files.) 2.2 Other 2020 File ConsiderationsRound-to-Round Changes to Job Rosters COVID-19 greatly impacted response rates, increasing the likelihood that job characteristics of MEPS job holders vary more than typical MEPS round-to-round changes. Jobholder gender, race, educational attainment, industry, occupation, establishment size, and job place flexibility all impact employment stability more so since the advent of COVID-19. Users should also note that non-responding households in Panel 25 produced a higher drop-off in jobs reviewed in Round 2 than in prior first year panels. For instance, between Round 1 and Round 2, there was a 54% drop-off in job records in Panel 25, 14% more than the drop-off seen in Panel 24 between Round 1 and 2, and 15% more than the drop-off seen in Panel 23 between Round 1 and Round 2. For these reasons, it will be important for users to compare jobholder and job characteristics in their analysis. Moreover, since over 80% of all jobs in a panel are reported in the first round of the survey, given the low response rates in Round 1 and drop-off of reviewed jobs in Round 2, COVID-19 will have a sustained impact on Panel 25 employment data more so than in any other panel. Questions where Respondents Note COVID-19 Impacts Users of the 2020 Jobs file may find it helpful to know where in Employment sections field interviewers documented COVID-related impacts from respondent comments:
2.3 Person-Level EstimatesThis 2020 Jobs file does not include any weights necessary to extrapolate this data to the U.S. population. To make person-level estimates, link to any of the 2020 MEPS files and use the person-level weight for the appropriate panel. The link should be made through the variable DUPERSID. Note that not all persons in the MEPS have positive weights and job records; only those persons who have either a positive person-level or family-level weight in the 2020 Population Characteristics Public Use file are included in the 2020 Jobs file. 2.4 Codebook StructureFor each variable on the 2020 Jobs file, an unweighted frequency is provided in the accompanying codebook file. 2.5 Reserved Codes
The value -15 (CANNOT BE COMPUTED) assigned to MEPS constructed variables in cases where there is not enough information from the MEPS instrument to calculate the constructed variable. “Not enough information” is often the result of skip patterns in the data or from missing information resulting from MEPS responses of -7 (REFUSED) or -8 (DK). Note that reserved code -8 includes cases where the information from the question was “not ascertained” or where the respondent chose “don’t know”. 2.6 Codebook FormatThis codebook describes an ASCII dataset (with related SAS, SPSS, R, and Stata programming statements and data user information), although the data are also provided in a SAS data set, SAS transport file, Stata data set, and Excel file. The file contains 84 variables and has a logical record length of 269 with an additional 2-byte carriage return/line feed at the end of each record.
2.7 Variable Source and Naming ConventionsBeginning in 2018, as variable collection, universe, or categories are altered, the variable name will be appended with “_Myy” to indicate in which year the alterations took place. Details about these alterations can be found throughout this document. In general, variable names reflect the content of the variable. Due to system changes, variable names are no longer restricted to 8 characters. Variables contained on this file were derived from the questionnaire itself or from the CAPI. The source of each variable is identified in Section D. Variable-Source Crosswalk. Sources for each variable are indicated in one of two ways:
3.0 Discussion of Pandemic Effects on Quality of 2020 MEPS Data3.1 SummaryData collection for in-person sample surveys in 2020 presented real challenges after the onset of the COVID-19 pandemic at a national level in mid-March of that year. After major modifications to the standard MEPS study design, it was possible to collect data safely, but there were naturally concerns about the quality of the data after such modifications. Preliminary evaluations of survey estimates based on the sample weights established for the FY 2020 Population Characteristics PUF suggest that useful estimates can be obtained from the data provided on this PUF. However, such evaluations are continuing as part of the development of the FY 2020 Consolidated PUF to be released later in 2022, and there may be updates to the variables appearing on the FY 2020 Population Characteristics PUF as a result. 3.2 OverviewThe onset of the COVID-19 pandemic in 2020 had a major impact on the MEPS Household Component (MEPS-HC) as it did for most major federal surveys and, of course, American life generally. The following discussion describes 1) the general impact of the pandemic on three major federal surveys (the effects on two of which also affect MEPS); 2) modifications to the MEPS sample design and field operations in 2020 due to the pandemic; and 3) potential data quality issues in the FY 2020 MEPS data related to the COVID-19 pandemic. 3.3 The Impact of the Pandemic: Other Major Federal SurveysMany important federal surveys were collecting data when the pandemic became national in scope in March 2020. Among them were the Current Population Survey (CPS), the American Community Survey (ACS), and the National Health Interview Survey (NHIS). The ACS and the NHIS field fresh samples each year. The CPS includes rotating panels, meaning some of the sampled households fielded had participated in prior years while others were fresh. Two of these surveys have important roles in MEPS. CPS subgroup estimates serve as raking control totals in the MEPS weighting process while the Round 1 households fielded for MEPS are selected annually as a subsample from the set of NHIS responding households from the prior year. All three of these surveys have reported bias concerns resulting from the data collected in 2020 being of lower than usual quality. In fact, the ACS decided not to release a standard database for single year 2020 due to the uncertain quality of the 2020 data, while the CPS and the NHIS released data but included reports discussing concerns about bias. All three surveys have reported evidence of nonresponse bias, specifically, that households in higher socio-economic levels were relatively more likely to respond and the sample weighting was unable to fully compensate for this. As a result, analysts have been cautioned about the accuracy of survey estimates and the ability to compare resulting estimates to estimates obtained in the years prior to the pandemic. For instance, in the documentation for the Current Population Survey, 2021 Annual Social and Economic (ASEC) Supplement, in Appendix G (beginning on page G-6) the impact of the pandemic on survey estimates is discussed. The two following statements appear at the beginning and end of the section: Data users should exercise caution when comparing estimates for data years 2019 and 2020 from the reports or from the microdata files to those from previous years due to the effects that the coronavirus (COVID-19) had on G-9 interviewing and response rates. Interviewing for the March 2020 CPS began on March 15, 2020. - Using administrative data, Census Bureau researchers have documented that there are more (and larger) differences between respondents and nonrespondents in 2020 and 2021 than in earlier years. Of particular interest for the estimates in the ASEC reports are the differences in median income and educational attainment, indicating that respondents in 2020 and 2021 had relatively higher income and were more educated than nonrespondents. The full document, 2021 Annual Social and Economic Supplement (census.gov), is available on the Census Bureau web site. The quality of CPS data is of particular importance to MEPS as CPS estimates serve as the control totals for the raking component of the MEPS weighting process. These control totals are based on the following demographic variables: age, sex, race/ethnicity, region, MSA status, educational attainment, and poverty status. The CPS estimates used in the development of the FY 2020 MEPS PUF weights that were based on the variables of age, sex, race/ethnicity, region, and MSA status were evaluated by the Census Bureau and determined to be of high quality. However, similar evaluations of the corresponding CPS estimates associated with educational attainment and poverty status found that these estimates suffered from bias, as reported in the documentation quoted above. For the MEPS FY 2020 Population Characteristics PUF weights, an approach based on a regression model to incorporate educational attainment was developed that appears quite acceptable. The use of poverty status in some form for the FY 2020 Consolidated PUF weights is still under consideration. A set of references discussing the fielding of these three surveys during the pandemic and resulting bias concerns can be found in the References section. 3.4 The Impact of the Pandemic: MEPS-HCTo set the stage, for the MEPS-HC, face-to-face interviewing ceased due to the COVID-19 pandemic on March 17, 2020. At that time, there were two MEPS panels in the field for which 2020 data were being collected: Round 1 of Panel 25 and Round 3 of Panel 24. The sampled households for Panel 25 were being contacted and asked to participate in MEPS for the first time while those from Panel 24 had already participated in MEPS for two rounds. A third MEPS panel was also in the field in early 2020, Round 5 of Panel 23, collecting data for the last portion of 2019. In developing a plan for how best to resume MEPS data collection, the primary issues were how to do so safely for both sampled household members and interviewers and the potential impact on data quality. Telephone data collection, although not the preferred method of data collection in general for MEPS-HC, was the natural option because it did not require in-person contact with respondents and could be implemented relatively quickly. Telephone numbers for Panel 24 were readily available as they had been routinely collected during earlier rounds. Telephone numbers were also generally available for Panel 25 Round 1 households from the NHIS data collection effort. However, the impact of changing to telephone on both response rates and data quality was expected to be larger for Panel 25 Round 1. Round 1 response rates are always the lowest of the five rounds, telephone survey response rates are typically lower than those for in-person surveys, and Round 1 respondents are new to the task of keeping track of health care events and are not expected to report them as reliably as previous MEPS participants who have been encouraged to keep records of health care events for the next interview. Moreover, at the time in-person interviewing stopped in mid-March 2020 completion rates for Panels 23 and 24 were substantially higher than those for Panel 25, as the older panels typically begin data collection earlier in the calendar year and take less effort to complete than the new panel. Weighted analyses show, if each panel were to represent the full MEPS target population individually, Panel 23 in-person interviews would represent about 72 percent of the MEPS target population, with Panel 24, about 61 percent; and Panel 25 about 53 percent. With all this in mind, there were important data quality concerns related to Panel 25 resulting in a decision to change the MEPS-HC study design to mitigate the potential impact on FY 2020 data. It was decided to field Panel 23 for at least one more year, asking Panel 23 respondents if they would be open to further participation in MEPS in newly added Rounds 6 and 7. Extending Panel 23 was meant to both offset the decrease in the number of cases in the FY 2020 data related to lower expected sample yields for Panel 25 and to improve data quality by retaining a set of participants who were familiar with MEPS. These changes necessitated operational changes, including: interviewer training to support telephone work, training and distribution of personal protective equipment (PPE) for COVID-19 safety for limited in-person contact, developing a respondent website for show cards and other materials, and adding a fall Panel 23 Round 6 interview covering all 2020 events from January 1, 2020 to the date of the interview. 3.5 Data Quality Issues for MEPS for FY 2020An article by Zuvekas and Kashihara (2021) provides useful background information on how the MEPS study design was modified in 2020 in response to the pandemic and mentions analyses undertaken to examine potential impacts on data quality. Such analyses continue into 2022 to gain a more complete understanding of any data quality issues prior to the release of health care events and expenditure data on the Full Year 2020 Consolidated PUF scheduled for later in 2022. One set of analyses has focused on comparing preliminary (not finalized) data on health care events between panels. These comparisons have served as a tool to help assess survey estimates based on the Full Year 2020 Population Characteristics PUF data and weights. The strategy has been to compare event estimates between panels, based on the weights developed for each of the three individual panels. Following is an explanation of this strategy. For these analyses, data from Panel 24 are considered the least affected by MEPS study design changes (i.e., shifting from in-person to telephone interviewing, extending participation) in response to COVID-19. Thus, Panel 24 is regarded as the panel whose estimates would most accurately reflect the behavior of the MEPS target population both at the time of the onset of the pandemic and over the course of 2020. This assessment was based on a number of considerations. Panel 24 had reference periods that were of the usual length for MEPS Rounds 3 and 4, unlike Panel 23, which in Round 6 had to report on events going back to January 1, 2020. Panel 24 Round 3 response rates were only slightly lower than Round 3 response rates for previous panels, unlike the much lower-than-usual Panel 25 Round 1 response rates, and Panel 24 respondents were familiar with MEPS after having previously participated in the two prior rounds, unlike Panel 25 respondents who were new to MEPS. While using the Panel 24 data alone for this FY 2020 PUF was an option, including the data from the other two panels increases precision of the estimates and provides more power to detect differences. Typically there are not differences in health care utilization between panels in a data year, and so any differences in event estimates between Panel 24 and Panel 23 or between Panel 25 and Panel 24 may point to bias concerns arising from the challenges of collecting data in 2020. As discussed below, such bias concerns do arise. Comparisons of Panel 24 with Panel 23 health care events data indicate that the longer-than-usual reference period for Panel 23 Round 6 may have resulted in recall issues for respondents. Round 6 was initially fielded in the late summer and early fall of 2020, and because the Round 5 reference period ended December 31, 2019, the recall period for health care events and related information extended back to January 1, 2020, much longer than for typical MEPS rounds. There is evidence that events of a less memorable nature such as dental visits and office-based physician visits occurring in early 2020 were under-reported. This evidence was obtained through both the examination of data related to reported events early in 2020 as well as statistical comparisons of Panel 23 estimates of the mean number of events for 2020 to corresponding estimates for Panel 24. Comparisons of Panel 25 with Panel 24 health care events data show evidence that Panel 25 estimates overstate the mean number of events for several event types. As noted by Zuvekas and Kashihara, nonresponse bias for Panel 25 (the sample with no previous MEPS experience) mirrors that reported for the CPS, the ACS, and the NHIS, where participants with higher socio-economic status (as measured by higher education levels) contributed disproportionately to preliminary Panel 25 survey estimates of the mean number of events. The authors expected that nonresponse and raking adjustments associated with educational attainment would serve to reduce this source of bias. Nevertheless, while the educational attainment variable played a more prominent role than in prior years in the nonresponse adjustment component of the MEPS weighting process and was used in the raking component as well, the bias was not fully eliminated. Age group comparisons suggest that the disproportionate contributions are most evident for events associated with those in the age range 18-64. With evidence of bias leading to an overestimate of events for some event types in Panel 25 and an underestimate in Panel 23, the question remained what the combined effect would be on the estimates. A further analysis compared estimates based on the three panels pooled together using the Full Year 2020 Population Characteristics PUF weight with the estimates based only on Panel 24 data and weights and found them to be similar. Using the same approach, comparisons were made for health insurance estimates (some private, only public, no health insurance) with the same results. The results also held true when the age ranges 0-17, 18-64, and 65 or older were considered separately. Based on these analyses, the weights for the FY 2020 Population Characteristics PUF should serve a useful purpose for the development of preliminary estimates and analyses while further assessments of the MEPS weights and data quality are made. 4.0 Longitudinal AnalysisPanel-specific longitudinal files are available for downloading in the data section of the MEPS website. For all three panels (Panel 23, Panel 24, and Panel 25), the longitudinal file comprises MEPS survey data obtained in Rounds 1 through 5 of the panel and can be used to analyze changes over a two-year period. In addition, for Panel 23 a file representing a three-year period will also be established. Variables in the file pertaining to survey administration, demographics, employment, health status, disability days, quality of care, health insurance, and medical care use and expenditures were obtained from the MEPS full-year Consolidated files from the years covered by that panel. For more details or to download the data files, please see Longitudinal Weight Files. 4.1 Using MEPS Data for Trend AnalysisFirst, of course, we note that there are uncertainties associated with 2020 data quality for reasons discussed in Section 3.5. Preliminary evaluations of a set of MEPS estimates of particular importance suggest that they are of reasonable quality. Nevertheless, analysts are advised to exercise caution in interpreting these estimates, particularly in terms of trend analyses since access to health care was substantially affected by the COVID-19 pandemic as were related factors such as health insurance and employment status for many people. Further evaluations of data will be undertaken as part of the effort in developing the Full Year 2020 Consolidated PUF. In terms of other factors to be aware of, MEPS began in 1996, and the utility of the survey for analyzing health care trends expands with each additional year of data; however, it is important to consider a variety of factors when examining trends over time using MEPS. Tests of statistical significance should be conducted to assess the likelihood that observed trends are not attributable to sampling variation. The length of time being analyzed should also be considered. In particular, large shifts in survey estimates over short periods of time (e.g., from one year to the next) that are statistically significant should be interpreted with caution unless they are attributable to known factors such as changes in public policy, economic conditions, or MEPS survey methodology. With respect to methodological considerations, in 2013 MEPS introduced an effort focused on field procedure changes such as interviewer training to obtain more complete information about health care utilization from MEPS respondents with full implementation in 2014. This effort likely resulted in improved data quality and a reduction in underreporting starting in the 2014 full year files and have had some impact on analyses involving trends in utilization across years. The implementation of a new NHIS sample design in 2016 could also potentially affect trend analyses. The new NHIS sample design is based on more up-to-date information related to the distribution of housing units across the U.S. As a result, it can be expected to better cover the full U.S. civilian, noninstitutionalized population, the target population for MEPS, as well as many of its subpopulations. Better coverage of the target population helps to reduce the potential for bias in both NHIS and MEPS estimates. Another change with the potential to affect trend analysis involved modifications to the MEPS instrument design and data collection process, particularly in the events sections of the instrument. These were introduced in the Spring of 2018 and thus affected data beginning with Round 1 of Panel 23, Round 3 of Panel 22, and Round 5 of Panel 21. Since the Full Year 2017 PUFs were established from data collected in Rounds 1-3 of Panel 22 and Rounds 3-5 of Panel 21, they reflected two different instrument designs. In order to mitigate the effect of such differences within the same full year file, the Panel 22, Round 3 data and the Panel 21 Round 5 data were transformed to make them as consistent as possible with data collected under the previous design. The changes in the instrument were designed to make the data collection effort more efficient and easy to administer. In addition, expectations were that data on some items, such as those related to health care events, would be more complete with the potential of identifying more events. Increases in service use reported since the implementation of these changes are consistent with these expectations. Data users should be aware of possible impacts on the data and especially trend analyses for these data years due to the design transition. Process changes, such as data editing and imputation, may also affect trend analyses. For example, users should refer to Section 2.5.11 in the 2020 Consolidated Public Use file (HC-224) and, for more detail, the documentation for the prescription drug file (HC-220A) when analyzing prescription drug spending over time. As always, it is recommended that data users review relevant sections of the documentation for descriptions of these types of changes that might affect the interpretation of changes over time before undertaking trend analyses. Analysts may also wish to consider using techniques to smooth or stabilize analyses of trends using MEPS data such as comparing pooled time periods (e.g., 1996-97 versus 2011-2012), working with moving averages or using modeling techniques with several consecutive years of MEPS data to test the fit of specified patterns over time. Finally, statistical significance tests should be conducted to assess the likelihood that observed trends are not attributable to sampling variation. In addition, researchers should be aware of the impact of multiple comparisons on Type I error. Without making appropriate allowance for multiple comparisons, undertaking numerous statistical significance tests of trends increases the likelihood of concluding that a change has taken place when one has not. ReferencesBramlett, M.D., Dahlhamer, J.M., & Bose, J. (2021, September). Weighting Procedures and Bias Assessment for the 2020 National Health Interview Survey. Centers for Disease Control and Prevention. Current Population Survey: 2021 Annual Social and Economic (ASEC) Supplement. (2021). U.S. Census Bureau. Dahlhamer, J.M., Bramlett, M.D., Maitland, A., & Blumberg, S.J. (2021). Preliminary evaluation of nonresponse bias due to the COVID-19 pandemic on National Health Interview Survey estimates, April-June 2020. National Center for Health Statistics. Daily, D., Cantwell, P.J., Battle, K., & Waddington, D.G. (2021, October 27), An Assessment of the COVID-19 Pandemic’s Impact on the 2020 ACS 1-Year Data. U.S. Census Bureau. Lau, D.T., Sosa, P., Dasgupta, N., & He, H. (2021). Impact of the COVID-19 Pandemic on Public Health Surveillance and Survey Data Collections in the United States. American Journal of Public Health, 111 (12), pp. 2118-2121. Rothbaum, J. & Bee, A. (2021, May 3). Coronavirus Infects Surveys, Too: Survey Nonresponse Bias and the Coronavirus Pandemic. U.S. Census Bureau. Rothbaum, J., Eggleston, J., Bee, A., Klee, M., & Mendez-Smith, B. (2021). Addressing Nonresponse Bias in the American Community Survey During the Pandemic Using Administrative Data. U.S. Census Bureau. Villa Ross, C.A., Shin, H.B., & Marlay, M.C. (2021, October 27). Pandemic Impact on 2020 American Community Survey 1-Year Data. U.S. Census Bureau. Zuvekas, S.H. & Kashihara, D. (2021). The Impacts of the Covid-19 Pandemic on the Medical Expenditure Panel Survey. American Journal of Public Health, 111 (12), pp. 2157-2166. D. Variable-Source CrosswalkFOR MEPS PUBLIC USE RELEASE HC-218
Appendix 3. MEPS Industry Codes Condensing Rules
MEPS uses the 4-digit Census occupation and industry coding systems developed for the Current Population Survey and the American Community Survey. For industry coding, MEPS uses the 2007 4-digit Census industry codes. Descriptions of the 4-digit Census industry codes can be found at the U.S. Bureau of Labor Statistics website. For occupation coding, MEPS uses the 2010 4-digit Census occupation codes. Descriptions of the 4-digit Census occupation codes can be found at the U.S. Bureau of Labor Statistics website. See Census IO Index for more information on the Census coding systems used by MEPS. Appendix 4. MEPS Occupation Codes Condensing Rules
MEPS uses the 4-digit Census occupation and industry coding systems developed for the Current Population Survey and the American Community Survey. For industry coding, MEPS uses the 2007 4-digit Census industry codes. Descriptions of the 4- digit Census industry codes can be found at the U.S. Bureau of Labor Statistics website. For occupation coding, MEPS uses the 2010 4-digit Census occupation codes. Descriptions of the 4-digit Census occupation codes can be found at the U.S. Bureau of Labor Statistics website. See the Census IO Index for more information on the Census coding systems used by MEPS. Appendix 5 Modification of Employment Variables for Panel 23OverviewDue to the impact of the COVID-19 pandemic on MEPS collection methods and lower response rates in the 2020 calendar year, AHRQ extended fielding for Panel 23 persons to include two additional rounds, Round 6 and Round 7, that collected information about 2020. As a result, 2020 MEPS includes three panels of data, Panel 25 Rounds 1, 2, and 3, Panel 24 Rounds 3, 4, and 5, and Panel 23 Rounds 5, 6, and 7. The decision to extend Panel 23 to include additional rounds occurred after Panel 23 Round 5 was fielded. As a result, Round 5 was fielded as a traditional terminal round where questions are asked of respondents as of 12/31/2019. Had the decision been made much earlier, Round 5 would have been fielded as a cross-year round (similar to Round 3), where questions would have covered the events through 2020 up to the Round 5 interview date. Given this impromptu design decision, in order to capture data after 12/31/2019, the Panel 23 Round 6 reference period began on January 1, 2020 (not the Round 5 interview date) and ended anywhere from July through October 2020. This duration (7 to 10 months) exceeds a standard round length, which averages 6 months. In an effort to make the Panel 23 employment variables comparable to the data from prior years, a decision was made to use the data collected in Round 6, along with data in reported in prior rounds, to separately identify employment characteristics and jobs that occurred (a) between the Round 4 and the Round 5 interview date and (b) between the Round 5 and the Round 6 interview dates. If these adjustments had not been made, the collected data would have instead identified characteristics between (c) the Round 4 interview date and 12/31/2019 and (c) 1/1/2020 and the Round 6 interview date.
Special 2020 Panel 23 Round 6 Employment Data Processing Developing 2020 Employment Information In order to create consistency in employment measures over time, AHRQ requested that job rosters and employment status be adjusted for FY 2020 public use files so that (A) Round 5 would represent employment for the time period between 1/1/2020 and the Round 5 interview date and (B) Round 6 would represent employment for the time period between the Round 5 interview date and the Round 6 interview date. The adjustment lengthens Round 5 to resemble a standard Round 3 cross-year reference period and shortens Round 6 to resemble a standard Round 4 field period Persons Eligible for Job Roster Adjustment Any person who was part of a household in Round 6 and who had one or more Round 6 jobs that began or ended between January 1, 2020 and the Round 5 interview date was eligible for adjustment. Number of Jobs and Persons Involved in Special Processing Job rosters for 610 persons were evaluated for adjustment. These persons had 711 job records whose start and/or stop date was on or before the Round 5 interview date. Some persons had more than one job flagged for review. Of the 610 persons, 188 persons (with 289 job records) either had multiple jobs requiring adjustment or complex Round 5 and Round 6 rosters They include persons with multiple miscellaneous jobs or multiple former jobs in addition to jobs requiring adjustment. These rosters were evaluated and adjusted by experienced Employment section instrument analysts (described below). The remaining 422 job records were programmatically adjusted based on how jobs are allowed to be reported in CAPI flow. Required AdjustmentsJob Roster Adjustments The following describes rules for selecting a job for adjustment and the typical roster adjustment made:
In some cases, Round 7 job-level variables, such as CREATEQ, ORIGRND, and TYPECHGD, required adjustment. Person-Level Adjustments Employment variables rely on other person variables. Internal use versions of administrative variables, INSCOP31 and INSCOP42, and demographic variables, AGE31X and AGE42X, were specially recalculated for use in this process. Scope Scope is used in Employment variable construction to determine household status between Round 4 and Round 5 interview dates and Round 5 and Round 6 interview dates. For Employment purposes, situations of concern were those where (a) a person had a job in Round 5 but left the RU in Round 6 or (b) a person was not present in Round 5 but was part of the Round 6 interview. For the first group (a), a person maintained their ‘in scope’ status during adjustment so that Round 5 employment status (EMPST31) reflected employment activity in both calendar years consistent with a second panel Round 3 person whose reference period crosses over from one calendar year into the next. Maintaining an ‘in scope’ status enabled complete setting of EMPST31 for persons where the following is true:
Persons in the latter group (b) began participation with MEPS in Round 6, so there was no Round 5 data. Therefore, scope was adjusted for these persons during the adjustment process. These persons were set to either:
With either setting, per existing Employment variable logic, a person’s Round 5 employment status, EMPST31, can be derived and Employment variables can reflect employment activity.
R31 Employment variables were constructed for the following groups of persons who had no previous 2019 Round 5 Employment information:
Age An internal version of AGE31X was calculated based on the Round 5 interview date instead of 12/31/19. Twenty three persons had their age change from 15 to 16 and nine had a job requiring adjustment, two of which were current main jobs. Variable Construction Adjustments Once jobs were adjusted to reflect whether they occurred in round 5 and/or round 6, “31” and “42” variables identifying the job characteristics were constructed for the JOBS file. Not all person-round level data required for development of Employment variables could be adjusted. Consequently, variable logic was modified to use both Round 5 and Round 6 in variable development. These variables were EMPSTrr, HELDrrX, OFFERrrX, and DISVWrrX.
|
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||