Centers for Medicare & Medicaid Services (CMS), HHS.
Proposed rule.
This proposed rule would update Home Health Prospective Payment System (HH PPS) rates, including the national, standardized 60-day episode payment rates, the national per-visit rates, and the non-routine medical supply (NRS) conversion factor under the Medicare prospective payment system for home health agencies (HHAs), effective for episodes ending on or after January 1, 2016. As required by the Affordable Care Act, this proposed rule implements the third year of the four-year phase-in of the rebasing adjustments to the HH PPS payment rates. This proposed rule provides information on our efforts to monitor the potential impacts of the rebasing adjustments. This proposed rule also proposes: reductions to the national, standardized 60-day episode payment rate in CY 2016 and CY 2017 of 1.72 percent in each year to account for estimated case-mix growth unrelated to increases in patient acuity (nominal case-mix growth) between CY 2012 and CY 2014; a HH value-based purchasing (HHVBP) model to be implemented beginning January 1, 2016 in which all Medicare-certified HHAs in selected states will be required to participate; changes to the home health quality reporting program requirements; and minor technical regulations text changes. Finally, this proposed rule would update the HH PPS case-mix weights using the most current, complete data available at the time of rulemaking and provide an update on the Report to Congress regarding the home health (HH) study.
To be assured consideration, comments must be received at one of the addresses provided below, no later than 5 p.m. on September 4, 2015.
In commenting, please refer to file code CMS–1625–P. Because of staff and resource limitations, we cannot accept comments by facsimile (FAX) transmission.
You may submit comments in one of four ways (please choose only one of the ways listed):
1.
2.
Please allow sufficient time for mailed comments to be received before the close of the comment period.
3.
4.
a. For delivery in Washington, DC—Centers for Medicare & Medicaid Services, Department of Health and Human Services, Room 445–G, Hubert H. Humphrey Building, 200 Independence Avenue SW., Washington, DC 20201.
(Because access to the interior of the Hubert H. Humphrey Building is not readily available to persons without federal government identification, commenters are encouraged to leave their comments in the CMS drop slots located in the main lobby of the building. A stamp-in clock is available for persons wishing to retain a proof of filing by stamping in and retaining an extra copy of the comments being filed.)
b. For delivery in Baltimore, MD—Centers for Medicare & Medicaid Services, Department of Health and Human Services, 7500 Security Boulevard, Baltimore, MD 21244–1850.
If you intend to deliver your comments to the Baltimore address, please call (410) 786–7195 in advance to schedule your arrival with one of our staff members.
Comments mailed to the addresses indicated as appropriate for hand or courier delivery may be delayed and received after the comment period.
For information on viewing public comments, see the beginning of the
Hillary Loeffler, (410) 786–0456, for general information about the HH PPS.
Michelle Brazil, (410) 786–1648 for information about the HH quality reporting program.
Lori Teichman, (410) 786–6684, for information about HHCAHPS.
Robert Flemming, (844) 280–5628, for information about the HHVBP model.
Comments received timely will also be available for public inspection as they are received, generally beginning approximately 3 weeks after publication of a document, at the headquarters of the Centers for Medicare & Medicaid Services, 7500 Security Boulevard, Baltimore, Maryland 21244, Monday through Friday of each week from 8:30 a.m. to 4 p.m. EST.
To schedule an appointment to view public comments, phone 1–800–743–3951.
In addition, because of the many terms to which we refer by abbreviation in this proposed rule, we are listing these abbreviations and their corresponding terms in alphabetical order below:
This proposed rule would update the payment rates for HHAs for calendar year (CY) 2016, as required under section 1895(b) of the Social Security Act (the Act). This would reflect the third year of the four-year phase-in of the rebasing adjustments to the national, standardized 60-day episode payment rate, the national per-visit rates, and the NRS conversion factor finalized in the CY 2014 HH PPS final rule (78 FR 72256), as required under section 3131(a) of the Patient Protection and Affordable Care Act of 2010 (Pub. L. 111–148), as amended by the Health Care and Education Reconciliation Act of 2010 (Pub. L. 111–152) (collectively referred to as the “Affordable Care Act”).
This proposed rule also discusses our efforts to monitor the potential impacts of the rebasing adjustments mandated by section 3131(a) of the Affordable Care Act. This rule proposes: Reductions to the national, standardized 60-day episode payment rate in CY 2016 and CY 2017 of 1.72 percent in each year to account for case-mix growth unrelated to increases in patient acuity (nominal case-mix growth) between CY 2012 and CY 2014 under the authority of section 1895(b)(3)(B)(iv) of the Act; a HH Value-Based Purchasing (VBP) model, in which certain Medicare-certified HHAs would be required to participate beginning January 1, 2016, under the authority of section 1115(A) of the Act; changes to the home health quality reporting program requirements under section 1895(b)(3)(B)(v)(II) of the Act; and minor technical regulations text changes in 42 CFR parts 409, 424, and 484 to better align the payment requirements with recent statutory and regulatory changes for home health
As required by section 3131(a) of the Affordable Care Act, and finalized in the CY 2014 HH final rule, “Medicare and Medicaid Programs; Home Health Prospective Payment System Rate Update for 2014, Home Health Quality Reporting Requirements, and Cost Allocation of Home Health Survey Expenses” (78 FR 77256, December 2, 2013), we are implementing the third year of the four-year phase-in of the rebasing adjustments to the national, standardized 60-day episode payment amount, the national per-visit rates and the NRS conversion factor in section III.C.3. The rebasing adjustments for CY 2016 would reduce the national, standardized 60-day episode payment amount by $80.95, increase the national per-visit payment amounts by 3.5 percent of the national per-visit payment amounts in CY 2010 with the increases ranging from $1.79 for home health aide services to $6.34 for medical social services, and reduce the NRS conversion factor by 2.82 percent.
This proposed rule also discusses our efforts to monitor the potential impacts of the rebasing adjustments in section III.A. In the CY 2015 HH PPS final rule (79 FR 66072), we finalized our proposal to recalibrate the case-mix weights every year with more current data. In section III.B.1 of this rule, we are recalibrating the HH PPS case-mix weights, using the most current cost and utilization data available, in a budget neutral manner. In addition, in section III.B.2 of this rule, we propose to reduce to the national, standardized 60-day episode payment rate in CY 2016 and CY 2017 by 1.72 percent in each year to account for estimated case-mix growth unrelated to increases in patient acuity (nominal case-mix growth) between CY 2012 and CY 2014. In section III.C.1 of this rule, we propose to update the payment rates under the HH PPS by the home health payment update percentage of 2.3 percent (using the 2010-based Home Health Agency (HHA) market basket update of 2.9 percent, minus 0.6 percentage point for productivity as required by section 1895(b)(3)(B)(vi)(I) of the Act. In the CY 2015 final rule (79 FR 66083 through 66087), we incorporated new geographic area designations, set out in a February 28, 2013 office of Management and Budget (OMB) bulletin, into the home health wage index. For CY 2015, we implemented a wage index transition policy consisting of a 50/50 blend of the old geographic area delineations and the new geographic area delineations. In section III.C.2 of this proposed rule, we propose to update the CY 2016 home health wage index using solely the new geographic area designations. In section III.D of this proposed rule, we discuss payments for high cost outliers. In section III.E, we propose to make several technical corrections in § 409, 424, and § 484 to better align the payment requirements with recent statutory and regulatory changes for home health services. The sections include § 409.43(e), § 424.22(a), § 484.205(d), § 484.205(e), § 484.220, § 484.225, § 484.230, § 484.240(b), § 484.240(e), § 484.240(f), § 484.245. In section III.F, we discuss the Report to Congress on the home health study required by section 3131(d) of the Affordable Care Act and provide an update on subsequent research and analysis.
In section IV of this proposed rule, we propose a HHVBP model to be implemented beginning January 1, 2016. Medicare-certified HHAs selected for inclusion in the HHVBP model would be required to compete for payment adjustments to their current PPS reimbursements based on quality performance. A competing Medicare-certified HHA is defined as an agency having a current Medicare certification and which is being reimbursed by CMS for home health care delivered within any of the nine states randomly selected under CMS' proposed selection methodology.
This proposed rule also includes changes to the home health quality reporting program in section III.V, including the proposal of one new quality measure, the establishment of a minimum threshold for submission of Outcome and Assessment Information Set (OASIS) assessments for purposes of quality reporting compliance, and submission dates for Home Health Care Consumer Assessment of Healthcare Providers and Systems Survey (HHCAHPS) Survey through CY 2018.
The Balanced Budget Act of 1997 (BBA) (Pub. L. 105–33, enacted August 5, 1997), significantly changed the way Medicare pays for Medicare HH services. Section 4603 of the BBA mandated the development of the HH PPS. Until the implementation of the HH PPS on October 1, 2000, HHAs received payment under a retrospective reimbursement system.
Section 4603(a) of the BBA mandated the development of a HH PPS for all Medicare-covered HH services provided under a plan of care (POC) that were paid on a reasonable cost basis by adding section 1895 of the Social Security Act (the Act), entitled “Prospective Payment For Home Health Services.” Section 1895(b)(1) of the Act requires the Secretary to establish a HH PPS for all costs of HH services paid under Medicare.
Section 1895(b)(3)(A) of the Act requires the following: (1) The computation of a standard prospective payment amount include all costs for HH services covered and paid for on a reasonable cost basis and that such amounts be initially based on the most recent audited cost report data available to the Secretary; and (2) the standardized prospective payment amount be adjusted to account for the
Section 1895(b)(3)(B) of the Act addresses the annual update to the standard prospective payment amounts by the HH applicable percentage increase. Section 1895(b)(4) of the Act governs the payment computation. Sections 1895(b)(4)(A)(i) and (b)(4)(A)(ii) of the Act require the standard prospective payment amount to be adjusted for case-mix and geographic differences in wage levels. Section 1895(b)(4)(B) of the Act requires the establishment of an appropriate case-mix change adjustment factor for significant variation in costs among different units of services.
Similarly, section 1895(b)(4)(C) of the Act requires the establishment of wage adjustment factors that reflect the relative level of wages, and wage-related costs applicable to HH services furnished in a geographic area compared to the applicable national average level. Under section 1895(b)(4)(C) of the Act, the wage-adjustment factors used by the Secretary may be the factors used under section 1886(d)(3)(E) of the Act.
Section 1895(b)(5) of the Act gives the Secretary the option to make additions or adjustments to the payment amount otherwise paid in the case of outliers due to unusual variations in the type or amount of medically necessary care. Section 3131(b)(2) of the Patient Protection and Affordable Care Act of 2010 (the Affordable Care Act) (Pub. L. 111–148, enacted March 23, 2010) revised section 1895(b)(5) of the Act so that total outlier payments in a given year would not exceed 2.5 percent of total payments projected or estimated. The provision also made permanent a 10 percent agency-level outlier payment cap.
In accordance with the statute, as amended by the BBA, we published a final rule in the July 3, 2000
Section 5201(c) of the Deficit Reduction Act of 2005 (DRA) (Pub. L. 109–171, enacted February 8, 2006) added new section 1895(b)(3)(B)(v) to the Act, requiring HHAs to submit data for purposes of measuring health care quality, and links the quality data submission to the annual applicable percentage increase. This data submission requirement is applicable for CY 2007 and each subsequent year. If an HHA does not submit quality data, the HH market basket percentage increase is reduced by 2 percentage points. In the November 9, 2006
The Affordable Care Act made additional changes to the HH PPS. One of the changes in section 3131 of the Affordable Care Act is the amendment to section 421(a) of the Medicare Prescription Drug, Improvement, and Modernization Act of 2003 (MMA) (Pub. L. 108–173, enacted on December 8, 2003) as amended by section 5201(b) of the DRA. Section 421(a) of the MMA, as amended by section 3131 of the Affordable Care Act, requires that the Secretary increase, by 3 percent, the payment amount otherwise made under section 1895 of the Act, for HH services furnished in a rural area (as defined in section 1886(d)(2)(D) of the Act) with respect to episodes and visits ending on or after April 1, 2010, and before January 1, 2016. Section 210 of the Medicare Access and CHIP Reauthorization Act of 2015 (MACRA) (Pub. L. 114–10) amended section 421(a) of the MMA to extend the rural add-on for two more years. Section 421(a) of the MMA, as amended by section 210 of the MACRA, requires that the Secretary increase, by 3 percent, the payment amount otherwise made under section 1895 of the Act, for HH services provided in a rural area (as defined in section 1886(d)(2)(D) of the Act) with respect to episodes and visits ending on or after April 1, 2010, and before January 1, 2018.
Generally, Medicare makes payment under the HH PPS on the basis of a national standardized 60-day episode payment rate that is adjusted for the applicable case-mix and wage index. The national standardized 60-day episode rate includes the six HH disciplines (skilled nursing, HH aide, physical therapy, speech-language pathology, occupational therapy, and medical social services). Payment for non-routine supplies (NRS) is no longer part of the national standardized 60-day episode rate and is computed by multiplying the relative weight for a particular NRS severity level by the NRS conversion factor (See section II.D.4.e). Payment for durable medical equipment covered under the HH benefit is made outside the HH PPS payment system. To adjust for case-mix, the HH PPS uses a 153-category case-mix classification system to assign patients to a home health resource group (HHRG). The clinical severity level, functional severity level, and service utilization are computed from responses to selected data elements in the OASIS assessment instrument and are used to place the patient in a particular HHRG. Each HHRG has an associated case-mix weight which is used in calculating the payment for an episode.
For episodes with four or fewer visits, Medicare pays national per-visit rates based on the discipline(s) providing the services. An episode consisting of four or fewer visits within a 60-day period receives what is referred to as a low-utilization payment adjustment (LUPA). Medicare also adjusts the national standardized 60-day episode payment rate for certain intervening events that are subject to a partial episode payment adjustment (PEP adjustment). For certain cases that exceed a specific cost threshold, an outlier adjustment may also be available.
As required by section 1895(b)(3)(B) of the Act, we have historically updated the HH PPS rates annually in the
To account for the changes in case-mix that were not related to an underlying change in patient health status, we implemented a reduction, over 4 years, to the national, standardized 60-day episode payment rates. That reduction was to be 2.75 percent per year for 3 years beginning in CY 2008 and 2.71 percent for the fourth year in CY 2011. In the CY 2011 HH PPS final rule (76 FR 68532), we updated our analyses of case-mix change and finalized a reduction of 3.79 percent, instead of 2.71 percent, for CY 2011 and deferred finalizing a payment reduction for CY 2012 until further study of the case-mix change data and methodology was completed.
In the CY 2012 HH PPS final rule (76 FR 68526), we updated the 60-day national episode rates and the national per-visit rates. In addition, as discussed in the CY 2012 HH PPS final rule (76 FR 68528), our analysis indicated that there was a 22.59 percent increase in overall case-mix from 2000 to 2009 and that only 15.76 percent of that overall observed case-mix percentage increase was due to real case-mix change. As a result of our analysis, we identified a 19.03 percent nominal increase in case-mix. At that time, to fully account for the 19.03 percent nominal case-mix growth identified from 2000 to 2009, we finalized a 3.79 percent payment reduction in CY 2012 and a 1.32 percent payment reduction for CY 2013.
In the CY 2013 HH PPS final rule (77 FR 67078), we implemented a 1.32 percent reduction to the payment rates for CY 2013 to account for nominal case-mix growth from 2000 through 2010. When taking into account the total measure of case-mix change (23.90 percent) and the 15.97 percent of total case-mix change estimated as real from 2000 to 2010, we obtained a final nominal case-mix change measure of 20.08 percent from 2000 to 2010 (0.2390 * (1−0.1597) = 0.2008). To fully account for the remainder of the 20.08 percent increase in nominal case-mix beyond that which was accounted for in previous payment reductions, we estimated that the percentage reduction to the national, standardized 60-day episode rates for nominal case-mix change would be 2.18 percent. Although we considered proposing a 2.18 percent reduction to account for the remaining increase in measured nominal case-mix, we finalized the 1.32 percent payment reduction to the national, standardized 60-day episode rates in the CY 2012 HH PPS final rule (76 FR 68532).
Section 3131(a) of the Affordable Care Act requires that, beginning in CY 2014, we apply an adjustment to the national, standardized 60-day episode rate and other amounts that reflect factors such as changes in the number of visits in an episode, the mix of services in an episode, the level of intensity of services in an episode, the average cost of providing care per episode, and other relevant factors. Additionally, we must phase in any adjustment over a four-year period in equal increments, not to exceed 3.5 percent of the amount (or amounts) as of the date of enactment of the Affordable Care Act, and fully implement the rebasing adjustments by CY 2017. The statute specifies that the maximum rebasing adjustment is to be no more than 3.5 percent per year of the CY 2010 rates. Therefore, in the CY 2014 HH PPS final rule (78 FR 72256) for each year, CY 2014 through CY 2017, we finalized a fixed-dollar reduction to the national, standardized 60-day episode payment rate of $80.95 per year, increases to the national per-visit payment rates per year as reflected in Table 2, and a decrease to the NRS conversion factor of 2.82 percent per year. We also finalized three separate LUPA add-on factors for skilled nursing, physical therapy, and speech-language pathology and removed 170 diagnosis codes from assignment to diagnosis groups in the HH PPS Grouper. In the CY 2015 HH PPS final rule (79 FR 66032), we implemented the second year of the four-year phase-in of the rebasing adjustments to the HH PPS payment rates and made changes to the HH PPS case-mix weights. In addition, we simplified the face-to-face encounter regulatory requirements and the therapy reassessment timeframes.
HHS has a number of initiatives designed to encourage and support the adoption of health information technology and to promote nationwide health information exchange to improve health care. As discussed in the August 2013 Statement “Principles and Strategies for Accelerating Health Information Exchange” (available at
The Office of the National Coordinator for Health Information Technology (ONC) has released a document entitled “Connecting Health and Care for the Nation: A Shared Nationwide Interoperability Roadmap Draft Version 1.0 (draft Roadmap) (available at
In addition, ONC has released the draft version of the 2015 Interoperability Standards Advisory (available at
We encourage stakeholders to utilize health information exchange and certified health IT to effectively and efficiently help providers improve internal care delivery practices, engage patients in their care, support management of care across the continuum, enable the reporting of electronically specified clinical quality measures (eCQMs), and improve efficiencies and reduce unnecessary costs. As adoption of certified health IT increases and interoperability standards continue to mature, HHS will seek to reinforce standards through relevant policies and programs.
As part of our efforts in monitoring the potential impacts of the rebasing adjustments finalized in the CY 2014 HH PPS final rule (78 FR 72293), we continue to update our analysis of home health cost report and claims data. In the CY 2014 HH PPS final rule, using 2011 cost report and 2012 claims data, we estimated the 2013 60-day episode cost to be $2,565.51 (78 FR 72277). In that final rule, we stated that our analysis of 2011 cost report data and 2012 claims data indicated a need for a −3.45 percent rebasing adjustment to the national, standardized 60-day episode payment rate each year for four years. However, as specified by statute, the rebasing adjustment is limited to 3.5 percent of the CY 2010 national, standardized 60-day episode payment rate of $2,312.94 (74 FR 58106), or $80.95. We stated that given that a −3.45 percent adjustment for CY 2014 through CY 2017 would result in larger dollar amount reductions than the maximum dollar amount allowed under section 3131(a) of the Affordable Care Act of $80.95, we were limited to implementing a reduction of $80.95 (approximately 2.8 percent for CY 2014) to the national, standardized 60-day episode payment amount each year for CY 2014 through CY 2017.
In the CY 2015 HH PPS final rule, (79 FR 66032–66118) using 2012 cost report and 2013 claims data, we estimated the 2013 60-day episode cost to be $2,485.24 (79 FR 66037). Similar to our discussion in the CY 2014 HH PPS final rule, we stated that absent the Affordable Care Act's limit to rebasing, in order to align payments with costs, a −4.21 percent adjustment would have been applied to the national, standardized 60-day episode payment amount each year for CY 2014 through CY 2017.
For this proposed rule, we analyzed 2013 HHA cost report data and 2013 HHA claims data to determine whether the average cost per episode was higher using 2013 cost report data compared to the 2011 cost report and 2012 claims data used in calculating the rebasing adjustments. To determine the 2013 average cost per visit per discipline, we applied the same trimming methodology outlined in the CY 2014 HH PPS proposed rule (78 FR 40284) and weighted the costs per visit from the 2013 cost reports by size, facility type, and urban/rural location so the costs per visit were nationally representative according to 2013 claims data. The 2013 average number of visits was taken from 2013 claims data. We estimate the cost of a 60-day episode in CY 2013 to be $2,402.11 using 2013 cost report data (Table 3). Our latest analysis of 2013 cost report and 2013 claims data suggests that an even larger reduction (−5.02 percent) than the reduction described in the CY 2014 HH PPS final rule (−3.45 percent) or the reduction described in the CY 2015 HH PPS final rule (−4.21) would have been needed in order to align payments with costs.
Section 3131(a) of the Affordable Care Act required the Medicare Payment Advisory Commission (MedPAC) to assess, by January 1, 2015, the impact of the mandated rebasing adjustments on quality of and beneficiary access to home health care. As part of this assessment, the statute required MedPAC to consider the impact on care delivered by rural, urban, nonprofit, and for-profit home health agencies. MedPAC's Report to Congress noted that the rebasing adjustments are partially offset by the payment update each year and across all four years of the phase-in of the rebasing adjustments the cumulative net reduction would equal about 2 percent. MedPAC concluded that, as a result of the payment update offsets to the rebasing adjustments, HHA margins are likely to remain high under the current rebasing policy and quality of care and beneficiary access to care are unlikely to be negatively affected.
As we noted in the CY 2014 HH PPS final rule (78 FR 72291), MedPAC's past reviews of access to home health care found that access generally remained adequate during periods of substantial decline in the number of agencies. MedPAC stated that this is due in part to the low capital requirements for home health care services that allow the industry to react rapidly when the supply of agencies changes or contracts. As described in section III.A.3, the number of HHAs billing Medicare for home health services in CY 2013 is 80 percent higher than the number of HHAs billing Medicare for home health services in 2001. Even if some HHAs were to exit the program due to possible reimbursement concerns, the home health market would be expected to remain robust.
In the CY 2014 HH PPS final rule (78 FR 72256), some commenters expressed concern that the rebasing of the HH PPS payment rates would result in HHA closures and would therefore diminish access to home health services. In addition to examining more recent cost report data, for this proposed rule we examined home health claims data from the first year of the four-year phase-in of the rebasing adjustments (CY 2014), the first calendar year of the HH PPS (CY 2001), and claims data for the three years before implementation of the rebasing adjustments (CY 2011–2013). Preliminary analysis of CY 2014 home health claims data indicates that the number of episodes decreased by 3.8 percent between 2013 and 2014. In addition, the number of home health users decreased by approximately 3 percent between 2013 and 2014, while the number of FFS beneficiaries has remained the same. Between 2013 and 2014 there appears to be a net decrease in the number of HHAs billing Medicare for home health services of 1.6 percent, driven mostly by decreases TX and FL, two of the six states with the highest utilization of Medicare home health (
We would note that preliminary data on hospital and skilled nursing facility discharges and days indicates that there was a decrease in hospital discharges of approximately 3 percent and a decrease in SNF days of approximately 2 percent in CY 2014. Any decreases in hospital discharges and skilled nursing facility days could, in turn, impact home health utilization as those settings serve as important sources of home health referrals.
For the six states (TX, LA, OK, MS, FL, and IL) with the highest utilization of Medicare home health (as measured by the number of episodes per Part A and/or Part B FFS beneficiaries), the number of episodes decreased by 5.7 percent, the number of home health users decreased by 4.3 percent, and the number of HHAs billing Medicare decreased by 3.7 percent (5,280–5,085) between 2013 and 2014 (
In addition to examining home health claims data from the first year of the implementation of rebasing adjustments required by the Affordable Care Act and comparing utilization in that year (CY 2014) to the three years prior and to the first calendar year following the implementation of the HH PPS (CY 2001), we subsequently examined trends in home health utilization for all years starting in CY 2001 and up through CY 2014. Figure 1, displays the average number of visits per 60-day episode of care and the average payment per visit. While the average payment per visit has steadily increased from approximately $116 in CY 2001 to $162 for CY 2014, the average total number of visits per 60-day episode of care has declined, most notably between CY 2009 (21.7 visits per episode) and CY 2014 (18.0 visit per episode). As noted in section II.C, we implemented a series of reductions to the national, standardized 60-day episode payment rate to account for increases in nominal case-mix, starting in CY 2008. The reductions to the 60-day episode rate were: 2.75 percent each year for CY 2008, CY 2009, and CY 2010; 3.79 percent for CY 2011and CY 2012; and a 1.32 percent payment reduction for CY 2013. Figure 2 displays the average number of visits by discipline type for a 60-day episode of care and shows that while the number of therapy visits per 60-day episode of care has increased slightly, the number of skilled nursing and home health aide visits have decreased, between CY 2009 and CY 2014. Section III.F describes the results of the home health study required by section 3131(d) of the Affordable Care Act, which suggests that the current home health payment system may discourage HHAs from serving patients with clinically complex and/or poorly controlled chronic conditions who do not qualify for therapy but require a large number of skilled nursing visits. The home health study results seems to be consistent with the recent trend in the decreased number of visits per episode of care driven by decreases in skilled nursing and home health aide services evident in Figures 1 and 2.
We will continue to monitor for potential impacts due to rebasing adjustments required by section 3131(a) of the Affordable Care Act and other policy changes in the future. Independent effects of any one policy may be difficult to discern in years where multiple policy changes occur in any given year.
For CY 2014, as part of the rebasing effort mandated by the Affordable Care Act, we reset the HH PPS case-mix weights, lowering the average case-mix weight to 1.0000. To lower the HH PPS case-mix weights to 1.0000, each HH PPS case-mix weight was decreased by the same factor (1.3464), thereby maintaining the same relative values between the weights. This “resetting” of the HH PPS case-mix weights was done in a budget neutral manner by inflating the national, standardized 60-day episode rate by the same factor (1.3464) that was used to decrease the weights. For CY 2015, we finalized a policy to annually recalibrate the HH PPS case-mix weights—adjusting the weights relative to one another—using the most current, complete data available. To recalibrate the HH PPS case-mix weights for CY 2016, we propose to use the same methodology finalized in the CY 2008 HH PPS final rule (72 FR 49762), the CY 2012 HH PPS final rule (76 FR 68526), and the CY 2015 HH PPS final rule (79 FR 66032). Annual recalibration of the HH PPS case-mix weights ensures that the case-mix weights reflect, as accurately as possible, current home health resource use and changes in utilization patterns.
To generate the proposed CY 2016 HH PPS case-mix weights, we used CY 2014 home health claims data (as of December 31, 2014) with linked OASIS data. These data are the most current and complete data available at this time. We will use CY 2014 home health claims data (as of June 30, 2015) with linked OASIS data to generate the CY 2016 HH PPS case-mix weights in the CY 2016 HH PPS final rule. The process
In updating the four-equation model for CY 2016, using 2014 data (the last update to the four-equation model for CY 2015 used 2013 data), there were few changes to the point values for the variables in the four-equation model. These relatively minor changes reflect the change in the relationship between the grouper variables and resource use between 2013 and 2014. The CY 2016 four-equation model resulted in 130 point-giving variables being used in the model (as compared to the 124 variables for the 2015 recalibration). There were nine variables that were added to the model and three variables that were dropped from the model due to the absence of additional resources associated with the variable. The points for 18 variables increased in the CY 2016 four-equation model and the points for 43 variables decreased in the CY 2016 4-equation model. There were 58 variables with the same point values.
• Step 1: First and second episodes, 0–13 therapy visits.
• Step 2.1: First and second episodes, 14–19 therapy visits.
• Step 2.2: Third episodes and beyond, 14–19 therapy visits.
• Step 3: Third episodes and beyond, 0–13 therapy visits.
• Step 4: Episodes with 20+ therapy visits
We then divide the distribution of the clinical score for episodes within a step such that a third of episodes are classified as low clinical score, a third of episodes are classified as medium clinical score, and a third of episodes are classified as high clinical score. The same approach is then done looking at the functional score. It was not always possible to evenly divide the episodes within each step into thirds due to many episodes being clustered around one particular score.
For Step 2.1, 77.2% of episodes were in the low functional level (Most with score 2 and 4).
For Step 2.2, 67.1% of episodes were in the low functional level (All with score 0).
For Step 3, 60.9% of episodes were in the medium functional level (Most with score 10).
For Step 4, 49.8% of episodes were in the low functional level (Most with score 2).
To ensure the changes to the HH PPS case-mix weights are implemented in a budget neutral manner, we would apply a case-mix budget neutrality factor to the CY 2016 national, standardized 60-day episode payment rate (
Section 1895(b)(3)(B)(iv) of the Act gives the Secretary the authority to implement payment reductions for nominal case-mix growth (that is, case-mix growth unrelated to changes in patient acuity). Previously, we accounted for nominal case-mix growth through case-mix reductions implemented from 2008 through 2013 (76 FR 68528–68543). As stated in the 2013 final rule, the goal of the reductions for nominal case-mix growth is to better align payment with real changes in patient severity (77 FR 67077). Our analysis of data from CY 2000 through CY 2010 found that only 15.97 percent of the total case-mix change was real and 84.03 percent of total case-mix change was nominal (77 FR 41553). In the CY 2015 HH PPS final rule (79 FR 66032), we estimated that total case-mix increased by 2.76 percent between CY 2012 and CY 2013 and of that amount, we estimated that 2.32 percent was a result of nominal case-mix growth (2.76 − (2.76 × 0.1597)). However, for 2015, we did not implement a reduction to the 2015 national, standardized 60-day episode payment amount to account for nominal case-mix growth, but stated that we would continue to monitor case-mix growth and may consider proposing nominal case-mix reductions in the future. Since the publication of the CY 2015 HH PPS final rule (79 FR 66032),
We are proposing to implement this 3.41 percent reduction in equal increments over 2 years. Specifically, in addition to continuing our third year of implementation of the rebasing adjustments required under section 3131(a) of the Affordable Care Act, we are proposing to apply a 1.72 percent (1 − 1/(1.0232 × 1.0118)
We invite comments on the proposed reduction to the national, standardized 60-day episode payment amount of 1.72 percent in CY 2016 and 1.72 percent in CY 2017 to account for nominal case-mix growth from CY 2012 through CY 2014 and the associated changes in the regulations text at § 484.220.
Section 1895(b)(3)(B) of the Act requires that the standard prospective payment amounts for CY 2015 be increased by a factor equal to the applicable HH market basket update for those HHAs that submit quality data as required by the Secretary. The home health market basket was rebased and revised in CY 2013. A detailed description of how we derive the HHA market basket is available in the CY 2013 HH PPS final rule (77 FR 67080- 67090).
Section 3401(e) of the Affordable Care Act, adding new section 1895(b)(3)(B)(vi) to the Act, requires that, in CY 2015 (and in subsequent calendar years), the market basket percentage under the HHA prospective payment system as described in section 1895(b)(3)(B) of the Act be annually adjusted by changes in economy-wide productivity. The statute defines the productivity adjustment, described in section 1886(b)(3)(B)(xi)(II) of the Act, to be equal to the 10-year moving average of change in annual economy-wide private nonfarm business multifactor productivity (MFP) (as projected by the Secretary for the 10-year period ending with the applicable fiscal year, calendar year, cost reporting period, or other annual period) (the “MFP adjustment”). The Bureau of Labor Statistics (BLS) is the agency that publishes the official measure of private nonfarm business MFP. Please see
Multifactor productivity is derived by subtracting the contribution of labor and capital input growth from output growth. The projections of the components of MFP are currently produced by IGI, a nationally recognized economic forecasting firm with which CMS contracts to forecast the components of the market basket and MFP. As described in the CY 2015 HH PPS proposed rule (79 FR 38384 through 38386), in order to generate a forecast of MFP, IGI replicated the MFP measure calculated by the BLS using a series of proxy variables derived from IGI's U.S. macroeconomic models. In the CY 2015 HH PPS proposed rule, we identified each of the major MFP component series employed by the BLS to measure MFP as well as provided the corresponding concepts determined to be the best available proxies for the BLS series.
Beginning with the CY 2016 rulemaking cycle, the MFP adjustment is calculated using a revised series developed by IGI to proxy the aggregate capital inputs. Specifically, IGI has replaced the Real Effective Capital Stock used for Full Employment GDP with a forecast of BLS aggregate capital inputs recently developed by IGI using a regression model. This series provides a better fit to the BLS capital inputs as measured by the differences between the actual BLS capital input growth rates and the estimated model growth rates over the historical time period. Therefore, we are using IGI's most recent forecast of the BLS capital inputs series in the MFP calculations beginning with the CY 2016 rulemaking cycle. A complete description of the MFP projection methodology is available on our Web site at
Using IGI's first quarter 2015 forecast, the MFP adjustment for CY 2016 (the 10-year moving average of MFP for the period ending CY 2016) is projected to be 0.6 percent. Thus, in accordance with section 1895(b)(3)(B)(iii) of the Act, we propose to base the CY 2016 market basket update, which is used to determine the applicable percentage increase for the HH payments, on the most recent estimate of the proposed 2010-based HH market basket (currently estimated to be 2.9 percent based on IGI's first quarter 2015 forecast). We propose to then reduce this percentage increase by the current estimate of the MFP adjustment for CY 2016 of 0.6 percentage point (the 10-year moving average of MFP for the period ending CY 2016 based on IGI's first quarter 2015 forecast), in accordance with 1895(b)(3)(B)(vi). Therefore, the current estimate of the CY 2016 HH update is 2.3 percent (2.9 percent market basket update, less 0.6 percentage point MFP adjustment). Furthermore, we note that if more recent data are subsequently
Section 1895(b)(3)(B) of the Act requires that the home health update be decreased by 2 percentage points for those HHAs that do not submit quality data as required by the Secretary. For HHAs that do not submit the required quality data for CY 2016, the home health update would be 0.3 percent (2.3 percent minus 2 percentage points).
Sections 1895(b)(4)(A)(ii) and (b)(4)(C) of the Act require the Secretary to provide appropriate adjustments to the proportion of the payment amount under the HH PPS that account for area wage differences, using adjustment factors that reflect the relative level of wages and wage-related costs applicable to the furnishing of HH services. Since the inception of the HH PPS, we have used inpatient hospital wage data in developing a wage index to be applied to HH payments. We propose to continue this practice for CY 2016, as we continue to believe that, in the absence of HH-specific wage data, using inpatient hospital wage data is appropriate and reasonable for the HH PPS. Specifically, we propose to continue to use the pre-floor, pre-reclassified hospital wage index as the wage adjustment to the labor portion of the HH PPS rates. For CY 2016, the updated wage data are for hospital cost reporting periods beginning on or after October 1, 2011 and before October 1, 2012 (FY 2012 cost report data).
We would apply the appropriate wage index value to the labor portion of the HH PPS rates based on the site of service for the beneficiary (defined by section 1861(m) of the Act as the beneficiary's place of residence). Previously, we determined each HHA's labor market area based on definitions of metropolitan statistical areas (MSAs) issued by the Office of Management and Budget (OMB). In the CY 2006 HH PPS final rule (70 FR 68132), we adopted revised labor market area definitions as discussed in the OMB Bulletin No. 03–04 (June 6, 2003). This bulletin announced revised definitions for MSAs and the creation of micropolitan statistical areas and core-based statistical areas (CBSAs). The bulletin is available online at
In this proposed rule, we propose to continue to use the same methodology discussed in the CY 2007 HH PPS final rule (71 FR 65884) to address those geographic areas in which there are no inpatient hospitals, and thus, no hospital wage data on which to base the calculation of the CY 2015 HH PPS wage index. For rural areas that do not have inpatient hospitals, we would use the average wage index from all contiguous CBSAs as a reasonable proxy. For FY 2016, there are no rural geographic areas without hospitals for which we would apply this policy. For rural Puerto Rico, we would not apply this methodology due to the distinct economic circumstances that exist there (for example, due to the close proximity to one another of almost all of Puerto Rico's various urban and non-urban areas, this methodology would produce a wage index for rural Puerto Rico that is higher than that in half of its urban areas). Instead, we would continue to use the most recent wage index previously available for that area. For urban areas without inpatient hospitals, we would use the average wage index of all urban areas within the state as a reasonable proxy for the wage index for that CBSA. For CY 2016, the only urban area without inpatient hospital wage data is Hinesville, GA (CBSA 25980).
On February 28, 2013, OMB issued Bulletin No. 13–01, announcing revisions to the delineations of MSAs, Micropolitan Statistical Areas, and CBSAs, and guidance on uses of the delineation of these areas. This bulletin is available online at
While the revisions OMB published on February 28, 2013 are not as sweeping as the changes made when we adopted the CBSA geographic designations for CY 2006, the February 28, 2013 bulletin does contain a number of significant changes. For example, there are new CBSAs, urban counties that have become rural, rural counties that have become urban, and existing CBSAs that have been split apart.
In the CY 2015 HH PPS final rule (79 FR 66085 through 66087), we finalized changes to the HH PPS wage index based on the newest OMB delineations, as described in OMB Bulletin No. 13–01, beginning in CY 2015, including a one-year transition with a blended wage index for CY 2015. Because the one-year transition period expires at the end of CY 2015, the proposed HH PPS wage index for CY 2016 is fully based on the revised OMB delineations adopted in CY 2015. The proposed CY 2016 wage index is available on the CMS Web site at
The Medicare HH PPS has been in effect since October 1, 2000. As set forth in the July 3, 2000 final rule (65 FR 41128), the base unit of payment under the Medicare HH PPS is a national, standardized 60-day episode payment rate. As set forth in 42 CFR 484.220, we adjust the national, standardized 60-day episode payment rate by a case-mix relative weight and a wage index value based on the site of service for the beneficiary.
To provide appropriate adjustments to the proportion of the payment amount under the HH PPS to account for area wage differences, we apply the appropriate wage index value to the labor portion of the HH PPS rates. The labor-related share of the case-mix adjusted 60-day episode rate would continue to be 78.535 percent and the non-labor-related share would continue to be 21.465 percent as set out in the CY 2013 HH PPS final rule (77 FR 67068). The CY 2016 HH PPS rates would use the same case-mix methodology as set forth in the CY 2008 HH PPS final rule with comment period (72 FR 49762) and would be adjusted as described in section III.C. of this rule. The following are the steps we take to compute the case-mix and wage-adjusted 60-day episode rate:
1. Multiply the national 60-day episode rate by the patient's applicable case-mix weight.
2. Divide the case-mix adjusted amount into a labor (78.535 percent) and a non-labor portion (21.465 percent).
3. Multiply the labor portion by the applicable wage index based on the site of service of the beneficiary.
4. Add the wage-adjusted portion to the non-labor portion, yielding the case-mix and wage adjusted 60-day episode rate, subject to any additional applicable adjustments.
In accordance with section 1895(b)(3)(B) of the Act, this document constitutes the annual update of the HH PPS rates. Section 484.225 sets forth the specific annual percentage update methodology. In accordance with § 484.225(i), for a HHA that does not submit HH quality data, as specified by the Secretary, the unadjusted national prospective 60-day episode rate is equal to the rate for the previous calendar year increased by the applicable HH market basket index amount minus two percentage points. Any reduction of the percentage change would apply only to the calendar year involved and would not be considered in computing the prospective payment amount for a subsequent calendar year.
Medicare pays the national, standardized 60-day case-mix and wage-adjusted episode payment on a split percentage payment approach. The split percentage payment approach includes an initial percentage payment and a final percentage payment as set forth in § 484.205(b)(1) and (b)(2). We may base the initial percentage payment on the submission of a request for anticipated payment (RAP) and the final percentage payment on the submission of the claim for the episode, as discussed in § 409.43. The claim for the episode that the HHA submits for the final percentage payment determines the total payment amount for the episode and whether we make an applicable adjustment to the 60-day case-mix and wage-adjusted episode payment. The end date of the 60-day episode as reported on the claim determines which calendar year rates Medicare would use to pay the claim.
We may also adjust the 60-day case-mix and wage-adjusted episode payment based on the information submitted on the claim to reflect the following:
• A low-utilization payment adjustment (LUPA) is provided on a per-visit basis as set forth in § 484.205(c) and § 484.230.
• A partial episode payment (PEP) adjustment as set forth in § 484.205(d) and § 484.235.
• An outlier payment as set forth in § 484.205(e) and § 484.240.
Section 1895(3)(A)(i) of the Act required that the 60-day episode base rate and other applicable amounts be standardized in a manner that eliminates the effects of variations in relative case mix and area wage adjustments among different home health agencies in a budget neutral manner. To determine the CY 2016 national, standardized 60-day episode payment rate, we would apply a wage index standardization factor, a case-mix budget neutrality factor described in section III.B.1, a nominal case-mix growth adjustment described in section III.B.2, the rebasing adjustment described in section II.C, and the MFP-adjusted home health market basket update discussed in section III.C.1 of this proposed rule.
To calculate the wage index standardization factor, henceforth referred to as the wage index budget neutrality factor, we simulated total payments for non-LUPA episodes using the 2016 wage index and compared it to our simulation of total payments for non-LUPA episodes using the 2015 wage index. By dividing the total payments for non-LUPA episodes using the 2016 wage index by the total payments for non-LUPA episodes using the 2015 wage index, we obtain a wage index budget neutrality factor of 1.0006. We would apply the wage index budget neutrality factor of 1.0006 to the CY 2016 national, standardized 60-day episode rate.
As discussed in section III.B.1 of this proposed rule, to ensure the changes to the case-mix weights are implemented in a budget neutral manner, we would apply a case-mix weight budget neutrality factor to the CY 2016 national, standardized 60-day episode payment rate. The case-mix weight budget neutrality factor is calculated as the ratio of total payments when CY 2016 case-mix weights are applied to CY 2014 utilization (claims) data to total payments when CY 2015 case-mix weights are applied to CY 2014 utilization data. The case-mix budget neutrality factor for CY 2016 would be 1.0141 as described in section III.B.1 of this proposed rule.
Next, as discussed in section III.B.2 of this proposed rule, we would apply a reduction of 1.72 percent to the national, standardized 60-day episode payment rate in CY 2016 to account for nominal case-mix growth between CY 2012 and CY 2014. Then, we would apply the −$80.95 rebasing adjustment finalized in the CY 2014 HH PPS final rule (78 FR 72256) and discussed in section II.C. Lastly, we would update the payment rates by the CY 2016 HH payment update percentage of 2.3 percent (MFP-adjusted home health market basket update) as described in section III.C.1 of this proposed rule. The CY 2016 national, standardized 60-day episode payment rate is calculated in Table 10.
The CY 2016 national, standardized 60-day episode payment rate for an HHA that does not submit the required quality data is updated by the CY 2016 HH payment update (2.3 percent) minus 2 percentage points and is shown in Table 11.
The national per-visit rates are used to pay LUPAs (episodes with four or fewer visits) and are also used to compute imputed costs in outlier calculations. The per-visit rates are paid by type of visit or HH discipline. The six HH disciplines are as follows:
• Home health aide (HH aide);
• Medical Social Services (MSS);
• Occupational therapy (OT);
• Physical therapy (PT);
• Skilled nursing (SN); and
• Speech-language pathology (SLP).
To calculate the CY 2016 national per-visit rates, we start with the CY 2015 national per-visit rates. We then apply a wage index budget neutrality factor to ensure budget neutrality for LUPA per-visit payments and increase each of the six per-visit rates by the maximum rebasing adjustments described in section II.C. of this rule. We calculate the wage index budget neutrality factor by simulating total payments for LUPA episodes using the 2016 wage index and comparing it to simulated total payments for LUPA episodes using the 2015 wage index. By dividing the total payments for LUPA episodes using the 2016 wage index by the total payments for LUPA episodes using the 2015 wage index, we obtain a wage index budget neutrality factor of 1.0006. We would apply the wage index budget neutrality factor of 1.0006 to the CY 2016 national per-visit rates.
The LUPA per-visit rates are not calculated using case-mix weights. Therefore, there is no case-mix weights budget neutrality factor needed to ensure budget neutrality for LUPA payments. Finally, the per-visit rates for each discipline are updated by the CY 2016 HH payment update percentage of 2.3 percent. The national per-visit rates are adjusted by the wage index based on the site of service of the beneficiary. The per-visit payments for LUPAs are separate from the LUPA add-on payment amount, which is paid for episodes that occur as the only episode or initial episode in a sequence of adjacent episodes. The CY 2016 national per-visit rates are shown in Tables 12 and 13.
The CY 2016 per-visit payment rates for an HHA that does not submit the required quality data are updated by the CY 2016 HH payment update (2.3 percent) minus 2 percentage points and is shown in Table 13.
LUPA episodes that occur as the only episode or as an initial episode in a sequence of adjacent episodes are adjusted by applying an additional amount to the LUPA payment before adjusting for area wage differences. In the CY 2014 HH PPS final rule, we changed the methodology for calculating the LUPA add-on amount by finalizing the use of three LUPA add-on factors: 1.8451 for SN; 1.6700 for PT; and 1.6266 for SLP (78 FR 72306). We multiply the per-visit payment amount for the first SN, PT, or SLP visit in LUPA episodes that occur as the only episode or an initial episode in a sequence of adjacent episodes by the appropriate factor to determine the LUPA add-on payment amount. For example, for LUPA episodes that occur as the only episode or an initial episode in a sequence of adjacent episodes, if the first skilled visit is SN, the payment for that visit would be $248.90 (1.8451 multiplied by $134.90), subject to area wage adjustment.
Payments for NRS are computed by multiplying the relative weight for a particular severity level by the NRS conversion factor. To determine the CY 2016 NRS conversion factor, we start with the 2015 NRS conversion factor ($53.23) and apply the −2.82 percent rebasing adjustment described in section II.C. of this rule (1−0.0282 = 0.9718). We then update the conversion factor by the CY 2016 HH payment update percentage (2.3 percent). We do not apply a standardization factor as the NRS payment amount calculated from the conversion factor is not wage or case-mix adjusted when the final claim payment amount is computed. The NRS conversion factor for CY 2016 is shown in Table 14.
Using the CY 2015 NRS conversion factor, the payment amounts for the six severity levels are shown in Table 15.
For HHAs that do not submit the required quality data, we again begin with the CY 2015 NRS conversion factor ($53.23) and apply the −2.82 percent rebasing adjustment discussed in section II.C of this proposed rule (1−0.0282= 0.9718). We then update the NRS conversion factor by the CY 2016 HH payment update percentage (2.3 percent) minus 2 percentage points. The CY 2016 NRS conversion factor for HHAs that do not submit quality data is shown in Table 16.
The payment amounts for the various severity levels based on the updated conversion factor for HHAs that do not submit quality data are calculated in Table 17.
Section 421(a) of the MMA required, for HH services furnished in a rural areas (as defined in section 1886(d)(2)(D) of the Act), for episodes or visits ending on or after April 1, 2004, and before April 1, 2005, that the Secretary increase the payment amount that otherwise would have been made under section 1895 of the Act for the services by 5 percent.
Section 5201 of the DRA amended section 421(a) of the MMA. The amended section 421(a) of the MMA required, for HH services furnished in a rural area (as defined in section 1886(d)(2)(D) of the Act), on or after January 1, 2006 and before January 1, 2007, that the Secretary increase the payment amount otherwise made under section 1895 of the Act for those services by 5 percent.
Section 3131(c) of the Affordable Care Act amended section 421(a) of the MMA to provide an increase of 3 percent of the payment amount otherwise made under section 1895 of the Act for HH services furnished in a rural area (as defined in section 1886(d)(2)(D) of the Act), for episodes and visits ending on or after April 1, 2010, and before January 1, 2016.
Section 210 of the Medicare Access and CHIP Reauthorization Act of 2015 (MACRA) (Pub. L. 114–10) amended section 421(a) of the MMA to extend the rural add-on by providing an increase of 3 percent of the payment amount otherwise made under section 1895 of the Act for HH services provided in a rural area (as defined in section 1886(d)(2)(D) of the Act), for episodes and visits ending before January 1, 2018.
Section 421 of the MMA, as amended, waives budget neutrality related to this provision, as the statute specifically states that the Secretary shall not reduce the standard prospective payment amount (or amounts) under section 1895 of the Act applicable to HH services furnished during a period to offset the increase in payments resulting in the application of this section of the statute.
For CY 2016, home health payment rates for services provided to beneficiaries in areas that are defined as rural under the OMB delineations would be increased by 3 percent as mandated by section 210 of the MACRA. The 3 percent rural add-on is applied to the national, standardized 60-day episode payment rate, national per visit rates, and NRS conversion factor when HH services are provided in rural (non-CBSA) areas. Refer to Tables 18 through 21 for these payment rates.
Section 1895(b)(5) of the Act allows for the provision of an addition or adjustment to the national, standardized 60-day case-mix and wage-adjusted episode payment amounts in the case of episodes that incur unusually high costs due to patient care needs. Prior to the enactment of the Affordable Care Act, section 1895(b)(5) of the Act stipulated that projected total outlier payments could not exceed 5 percent of total projected or estimated HH payments in a given year. In the July 3, 2000 Medicare Program; Prospective Payment System for Home Health Agencies final rule (65 FR 41188 through 41190), we described the method for determining outlier payments. Under this system, outlier payments are made for episodes whose estimated costs exceed a threshold amount for each HH Resource Group (HHRG). The episode's estimated cost is the sum of the national wage-adjusted per-visit payment amounts for all visits delivered during the episode. The outlier threshold for each case-mix group or Partial Episode Payment (PEP) adjustment is defined as the 60-day episode payment or PEP adjustment for that group plus a fixed-dollar loss (FDL) amount. The outlier payment is defined to be a proportion of the wage-adjusted estimated cost beyond the wage-adjusted threshold. The threshold amount is the sum of the wage and case-mix adjusted PPS episode amount and wage-adjusted FDL amount. The proportion of additional costs over the outlier threshold amount paid as outlier payments is referred to as the loss-sharing ratio.
In the CY 2010 HH PPS final rule (74 FR 58080 through 58087), we discussed excessive growth in outlier payments, primarily the result of unusually high outlier payments in a few areas of the country. Despite program integrity efforts associated with excessive outlier payments in targeted areas of the country, we discovered that outlier expenditures still exceeded the 5 percent target and, in the absence of corrective measures, would continue do to so. Consequently, we assessed the appropriateness of taking action to curb outlier abuse. To mitigate possible billing vulnerabilities associated with excessive outlier payments and adhere to our statutory limit on outlier payments, we adopted an outlier policy that included a 10 percent agency-level cap on outlier payments. This cap was implemented in concert with a reduced FDL ratio of 0.67. These policies resulted in a projected target outlier pool of approximately 2.5 percent. (The previous outlier pool was 5 percent of total HH expenditure). For CY 2010, we first returned the 5 percent held for the previous target outlier pool to the national, standardized 60-day episode rates, the national per-visit rates, the LUPA add-on payment amount, and the NRS conversion factor. Then, we reduced the CY 2010 rates by 2.5 percent to account for the new outlier pool of 2.5 percent. This outlier policy was adopted for CY 2010 only.
As we noted in the CY 2011 HH PPS final rule (75 FR 70397 through 70399), section 3131(b)(1) of the Affordable Care Act amended section 1895(b)(3)(C) of the Act, and requires the Secretary to reduce the HH PPS payment rates such that aggregate HH PPS payments are reduced by 5 percent. In addition, section 3131(b)(2) of the Affordable Care Act amended section 1895(b)(5) of the Act by re-designating the existing language as section 1895(b)(5)(A) of the Act, and revising it to state that the Secretary may provide for an addition or adjustment to the payment amount for outlier episodes because of their unusual variation in the type or amount of medically necessary care. The total amount of the additional payments or payment adjustments for outlier episodes may not exceed 2.5 percent of the estimated total HH PPS payments for that year and outlier payments as a percent of total payments are capped for each HHA at 10 percent.
As such, beginning in CY 2011, our HH PPS outlier policy is that we reduce payment rates by 5 percent and target up to 2.5 percent of total estimated HH PPS payments to be paid as outliers. To do so, we first returned the 2.5 percent held for the target CY 2010 outlier pool to the national, standardized 60-day episode rates, the national per visit rates, the LUPA add-on payment amount, and the NRS conversion factor for CY 2010. We then reduced the rates by 5 percent as required by section 1895(b)(3)(C) of the Act, as amended by section 3131(b)(1) of the Affordable Care Act. For CY 2011 and subsequent calendar years we target up to 2.5 percent of estimated total payments to
For a given level of outlier payments, there is a trade-off between the values selected for the FDL ratio and the loss-sharing ratio. A high FDL ratio reduces the number of episodes that can receive outlier payments, but makes it possible to select a higher loss-sharing ratio, and therefore, increase outlier payments for qualifying outlier episodes. Alternatively, a lower FDL ratio means that more episodes can qualify for outlier payments, but outlier payments per episode must then be lower.
The FDL ratio and the loss-sharing ratio must be selected so that the estimated total outlier payments do not exceed the 2.5 percent aggregate level (as required by section 1895(b)(5)(A) of the Act). Historically, we have used a value of 0.80 for the loss-sharing ratio which, we believe, preserves incentives for agencies to attempt to provide care efficiently for outlier cases. With a loss-sharing ratio of 0.80, Medicare pays 80 percent of the additional estimated costs above the outlier threshold amount.
In the CY 2011 HH PPS final rule (75 FR 70398), in targeting total outlier payments as 2.5 percent of total HH PPS payments, we implemented an FDL ratio of 0.67, and we maintained that ratio in CY 2012. Simulations based on CY 2010 claims data completed for the CY 2013 HH PPS final rule showed that outlier payments were estimated to comprise approximately 2.18 percent of total HH PPS payments in CY 2013, and as such, we lowered the FDL ratio from 0.67 to 0.45. We stated that lowering the FDL ratio to 0.45, while maintaining a loss-sharing ratio of 0.80, struck an effective balance of compensating for high-cost episodes while allowing more episodes to qualify as outlier payments (77 FR 67080). The national, standardized 60-day episode payment amount is multiplied by the FDL ratio. That amount is wage-adjusted to derive the wage-adjusted FDL amount, which is added to the case-mix and wage-adjusted 60-day episode payment amount to determine the outlier threshold amount that costs have to exceed before Medicare would pay 80 percent of the additional estimated costs.
For this proposed rule, simulating payments using preliminary CY 2014 claims data (as of December 31, 2014) and the CY 2015 payment rates (79 FR 66088 through 66092), we estimate that outlier payments in CY 2015 would comprise 2.02 percent of total payments. Based on simulations using CY 2014 claims data and the CY 2016 payments rates in section III.C.3 of this proposed rule, we estimate that outlier payments would comprise approximately 2.34 percent of total HH PPS payments in CY 2016, a percent change of almost 16 percent. This increase is attributable to the increase in the national per-visit amounts through the rebasing adjustments and the decrease in the national, standardized 60-day episode payment amount as a result of the rebasing adjustment and the nominal case-mix growth reduction. Given similar rebasing adjustments and case-mix growth reduction would also occur for 2017, and hence a similar anticipated increase in the outlier payments, we estimate that for CY 2017 outlier payments as a percent of total HH PPS payments would exceed 2.5 percent.
At this time, we are not proposing a change to the FDL ratio or loss-sharing ratio for CY 2016 as we believe that maintaining an FDL of 0.45 and a loss-sharing ratio of 0.80 are appropriate given the percentage of outlier payments is estimated to increase as a result of the increase in the national per-visit amounts through the rebasing adjustments and the decrease in the national, standardized 60-day episode payment amount as a result of the rebasing adjustment and nominal case-mix growth reduction. In the final rule, we will update our estimate of outlier payments as a percent of total HH PPS payments using the most current and complete year of HH PPS data (CY 2014 claims data as of June 30, 2015). We would continue to monitor the percent of total HH PPS payments paid as outlier payments to determine if future adjustments to either the FDL ratio or loss-sharing ratio are warranted.
The current home health prospective payment system (HH PPS) pays a determined amount for a 60-day episode of care adjusted for case mix using 153 home health resource groups (HHRGs). The 153 HHRGs are determined based on the amount of therapy provided, the episode's timing in a sequence of episodes, and the patient's clinical and functional status determined from data reported on the Outcome and Assessment Information Set (OASIS). There has been criticism that home health providers have responded to Medicare's payment policy by altering the level of service provided to patients.
This raises the question whether there is a disparity in payment for those patients with clinically complex and/or poorly controlled chronic conditions who do not qualify for therapy but require a large number of skilled nursing visits.
Section 3131(d) of the Affordable Care Act directed the Secretary to conduct a study on HHA costs involved with providing ongoing access to care to low-income Medicare beneficiaries or beneficiaries in medically underserved areas, and in treating beneficiaries with high levels of severity of illness.
The results of the survey revealed that over 80 percent of HHAs and over 90 percent of physicians reported that access to home health care for Medicare fee-for-service beneficiaries in their local area was excellent or good. When survey respondents reported access issues, specifically their inability to place or admit Medicare fee-for-service patients into home health, the most common reason reported was that the patients did not qualify for the Medicare home health benefit. HHAs and physicians also cited family or caregiver issues as an important contributing factor in the inability to admit or place patients. About 17.2 percent of HHAs and 16.7 percent of physicians reported insufficient payment as an important contributing factor in the inability to admit or place patients. The survey results suggest that much of the variation in access to Medicare home health services is associated with social and personal conditions and therefore CMS' ability to improve access for certain vulnerable patient populations through payment policy may be limited.
Analysis of CY 2010 HHA payment and cost data suggests that margins may differ substantially across the HH PPS case-mix groups. In addition, particular beneficiary characteristics appear to be strongly associated with margin, and thus may create financial incentives to select certain patients over others. Margins were estimated to be lower in CY 2010 for patients who required parenteral nutrition, who had traumatic wounds or ulcers, or required substantial assistance in bathing. Given that these variables are already included in the HH PPS case-mix system, the results indicate that modifications to the case-mix system may be needed. Furthermore, in CY 2010, beneficiaries admitted after acute or post-acute stays or who had high Hierarchical Condition Category scores or certain poorly-controlled clinical conditions, such as poorly-controlled pulmonary disorders, were also associated with substantially lower home health margins. In addition, other characteristics, such as those describing assistance by informal caregivers for ADL needs and those describing socio- economic status, such as dual eligibility for Medicare and Medicaid, were strongly associated with lower margins. Exploration of potential payment methodology changes indicated that accounting for additional variables in HH PPS payment may decrease the difference in estimated margin between individuals in specific vulnerable subgroups and those not in the subgroups, thereby potentially decreasing financial incentives to select certain types of patients over others.
CMS awarded a follow-on contract to Abt Associates to further explore margin differences across patient characteristics and possible payment methodology changes suggested by the results of the home health study. Additionally, we have heard from various stakeholders that the current payment system methodology is overly complex and does not fully reflect the range of services provided under the home health benefit, and thus this follow-on study would look at these aspects of the current payment system as well.
Under the follow-on contract, Abt Associates convened a Clinical Workgroup meeting on June 25, 2014 to gain clinical insight from industry regarding the current HH PPS. Based upon the feedback provided during the Clinical Workgroup meeting, as well as CMS concerns about the current model given the findings from the Home Health Study, Abt Associates was tasked with developing model options for consideration and discussion. In September 2014, Abt Associates presented several payment model options for CMS consideration, which were also presented to a Technical Expert Panel meeting held on January 8, 2015.
• Diagnosis on Top Model:
The first model option, referred to as the “Diagnosis on Top” (DOT) model, combines diagnosis groups with a regression model to create separate weights for patients with different diagnoses. For its “Studies in Home Health Case Mix” project design report (January 7, 2002), Abt had explored the possibility of a DOT model for the home health payment system. At that time, there was a decision that the potential gains in payment accuracy which would result from implementing a DOT model were offset by the added complexity and burden to providers that a DOT model could introduce by requiring providers to classify their patients with a single diagnosis that would be used to determine payment. For present reform efforts, Abt revisited the DOT model with more current data and in the context of other potential changes to the payment system which a DOT model might be able to complement. In this analysis, we are removing the therapy variable, allowing us to explore new ideas and re-explore previously rejected ideas to see how we can increase the statistical power of the model without the therapy variable. In this most recent analysis, each episode is grouped into the following diagnosis groups based on the primary ICD–9–CM diagnosis code reported on the OASIS: (1) Orthopedic; (2) neurological; (3) diabetes; (4) cancer; (5) skin wounds & lesions; (6) cardiovascular; (7) pulmonary; (8) gastrointestinal; (9) genito-urinary; (10) mental/emotional disorders; (11) other diagnoses; (12) case-mix V-codes; and (13) non-case-mix V-codes. Unlike the current HH PPS case-mix system, the diagnosis on top model does not include any therapy thresholds. Under the diagnosis on top model, episodes are first divided into different diagnosis groups, prior to the determination of the clinical and functional levels, and payment model regressions would be run separately for each diagnosis group. This is intended to maximize the statistical performance of the payment system. The work conducted by Abt Associates also included OASIS and non-OASIS items (such as whether the patient was admitted from an acute or post-acute care setting and hierarchical condition categories) not used in the current payment system, but shown to correlate with resource use. In many ways, the regression component of the diagnosis on top model is very similar to the current 4-equation model except that, in later versions of Abt's work on the diagnosis on top model, the clinical and functional levels are replaced with an overall severity level. This change allows the diagnosis on top model to account for a richer set of variables than the clinical and functional levels in the current payment system.
The second model option is referred to as the “Predicted Therapy Model.” The basic structure of this model is similar to that of the current payment model. In this model option, actual therapy visits used in the current HH PPS model are replaced with predicted therapy visits to develop case mix weights and payment amounts based on
The third model is referred to as the “Home Health Groupings ” (HHG) model. The premise of this type of model is that it starts with a clinical foundation. This groupings model groups home health episodes by diagnoses and the expected types of home health interventions required. Using expert clinical judgment, each ICD–9 code is assigned to one of seven groups based on the intervention expected to be required. Those seven groups include: (1) Musculoskeletal Rehabilitation; (2) Neuro/Stroke Rehabilitation; (3) Skin/Non-Surgical Wound Care; (4) Post-Op Wound Aftercare; (5) Behavioral Health Care; (6) Complex Medical Care; and (7) Medication Management, Teaching, and Assessment. Unlike the current HH PPS case-mix system, the home health groupings model does not include any therapy thresholds. Abt Associates is currently in the process of further delineating the seven groups listed above using OASIS and non-OASIS items (such as whether the patient was admitted from an acute or post-acute care setting and hierarchical condition categories) not used in the current payment system, but shown to correlate with resource use. The HHG model groups home health episodes in a way that mirrors how clinicians would differentiate between different types of beneficiaries and would help explain why the beneficiary is receiving home health, something that the current HH PPS case-mix may be lacking. MedPAC noted that policy makers have faced challenges in defining the role of home health.
To inform the model options discussed above, Abt Associates also reviewed other Medicare prospective payment systems to identify alternative methods used in classifying patients and to better understand components of each system. In the future, we plan to issue a technical report under our contract with Abt Associates that would further describe and analyze the three model options. We also plan to reconvene the Clinical Workgroup and the Technical Experts Panel in the near future to help further inform CMS on the various model options developed and next steps.
First, we propose to make several technical corrections in part 484 to better align the payment requirements with recent statutory and regulatory changes for home health services. We propose to make changes to § 484. 205(e) to state that estimated total outlier payments for a given calendar year are limited to no more than 2.5 percent of total outlays under the HHA PPS, rather than 5 percent of total outlays, as required by section 1895(b)(5)(A) of the Act as amended by section 3131(b)(2)(B) of the Affordable Care Act. Similarly, we also propose to specify in § 484.240(e) that the fixed dollar loss and the loss sharing amounts are chosen so that the estimated total outlier payment is no more than 2.5 percent of total payments under the HH PPS, rather than 5 percent of total payments under the HH PPS as required by section 1895(b)(5)(A) of the Act as amended by section 3131(b)(2)(B) of the Affordable Care Act. We also propose to describe in § 484.240(f) that the estimated total amount of outlier payments to an HHA in a given year may not exceed 10 percent of the estimated total payments to the specific agency under the HH PPS in a given year. This update aligns the regulations text at § 484.240(f) with the statutory requirement in 1895(b)(5)(A) of the Act as amended by section 3131(b)(2)(B) of the Affordable Care Act. Finally, we propose a minor editorial change in § 484.240(b) to specify that the outlier threshold for each case-mix group is the episode payment amount for that group,
Second, in addition to the proposed changes to the regulations text pertaining to outlier payments under the HH PPS, we also propose to amend § 409.43(e)(iii) and to add language to § 484.205(d) to clarify the frequency of review of the plan of care and the provision of Partial Episode Payments (PEP) under the HH PPS as a result of a regulations text change in § 424.22(b) that was finalized in the CY 2015 HH PPS final rule (79 FR 66032). Specifically, we propose to change the definition of an intervening event to include transfers and instances where a patient is discharged and return to home health during a 60-day episode, rather than a discharge and return to the same HHA during a 60-day episode. In § 484.220, we propose to update the regulations text to reflect the downward adjustments to the 60-day episode payment rate due to changes in the coding or classification of different units of service that do not reflect real changes in case-mix (nominal case-mix growth) applied to calendar years 2012 and 2013, which were finalized in the CY 2012 HH PPS final rule (76 FR 68532). This also includes updating the CY 2011 adjustment to 3.79 percent as finalized in the CY 2011 HH PPS final rule (75 FR 70461). In § 484.225 we are proposing to eliminate references to outdated market basket index factors by removing paragraphs (b), (c), (d), (e), (f) and (g). In § 484.230 we propose to delete the last sentence as a result of a change from a separate LUPA add-on amount to a LUPA add-on factor finalized in the CY 2014 HH PPS final rule (78 FR 72256). Finally, we are deleting and reserving § 484.245 as we believe that this language is no longer applicable under the HH PPS, as it was meant to facilitate the transition to the original PPS established in CY 2000.
Lastly, we propose to make one technical correction in § 424.22 to re-designate paragraph (a)(1)(v)(B)(
We invite comments on these technical corrections and associated changes in the regulations at § 409, § 424, and § 484.
In the CY 2015 Home Health Prospective Payment System (HH PPS) final rule titled “Medicare and Medicaid Programs; CY 2015 Home Health
In the CY 2015 HH PPS final rule,
The basis for developing this proposed value-based purchasing (VBP) model, as described in the proposed regulations at § 484.300
Second, section 3006(b) of the Affordable Care Act directed the Secretary of the Department of Health and Human Services (the Secretary) to develop a plan to implement a VBP program for payments under the Medicare Program for HHAs and the Secretary issued an associated Report to Congress in March of 2012 (2012 Report).
Third, section 402(a)(1)(A) of the Social Security Amendments of 1967 (as amended) (42 U.S.C. 1395b–1(a)(1)(A)), provided authority for us to conduct the Home Health Pay-for-Performance (HHPFP) Demonstration that ran from 2008 to 2010. The results of that Demonstration found modest quality improvement in certain measures after comparing the quality of care furnished by Demonstration participants to the quality of care furnished by the control group. One important lesson learned from the HHPFP Demonstration was the need to link the HHA's quality improvement efforts and the incentives. HHAs in three of the four regions generated enough savings to have incentive payments in the first year of the Demonstration, but the size of payments were unknown until after the conclusion of the Demonstration. Also, the time lag between quality performance and payment incentives was too long to provide a sufficient motivation for HHAs to take necessary steps to improve quality. The results of the Demonstration published in a comprehensive evaluation report
Furthermore, the President's FY 2015 and 2016 Budgets proposed that VBP should be extended to additional providers including skilled nursing facilities, home health agencies, ambulatory surgical centers, and hospital outpatient departments. The FY 2015 Budget called for at least 2 percent of payments to be tied to quality and efficiency of care on a budget neutral basis. The FY 2016 Budget outlines a program which would tie at least 2 percent of Medicare payments to the quality and efficiency of care in the first 2 years of implementation beginning in 2017, and at least 5 percent beginning in 2019 without any impact to the budget. We propose in this HHVBP model to also follow a graduated payment adjustment strategy within certain selected states beginning January 1, 2016.
The Secretary has also set two overall delivery system reform goals for CMS. First, we seek to tie 30 percent of traditional, or fee-for-service, Medicare payments to quality or value-based payments through alternative payment models by the end of 2016, and to tie 50 percent of payments to these models by the end of 2018. Second, we seek to tie 85 percent of all traditional Medicare payments to quality or value by 2016 and 90 percent by 2018.
Finally, we have already successfully implemented the Hospital Value-Based Purchasing (HVBP) program, under which value-based incentive payments are made in a fiscal year to hospitals that meet performance standards established for a performance period with respect to measures for that fiscal year. The percentage of a participating hospital's base-operating DRG payment amount for FY 2015 discharges that is at risk, based on the hospital's performance under the program for that fiscal year, is 1.5 percent. That percentage will increase to 2.0 by FY 2017. We are proposing an HHVBP model that builds on the lessons learned and guidance from the HVBP program and other applicable demonstrations as discussed above, as well as from the evaluation report discussed earlier.
The proposed HHVBP model presents an opportunity to improve the quality of care furnished to Medicare beneficiaries and study what incentives are sufficiently significant to encourage HHAs to provide high quality care. The HHVBP model being proposed would offer both a greater potential reward for high performing HHAs as well as a greater potential downside risk for low performing HHAs. If implemented, the model would begin on January 1, 2016, and include an array of measures that would capture the multiple dimensions of care that HHAs furnish.
The proposed model would be tested by CMS's Center for Medicare and Medicaid Innovation (CMMI) under section 1115A of the Act. Under section 1115A(d)(1) of the Act, the Secretary may waive such requirements of Titles XI and XVIII and of sections 1902(a)(1), 1902(a)(13), and 1903(m)(2)(A)(iii) as may be necessary solely for purposes of carrying out section 1115A with respect to testing models described in section 1115A(b). The Secretary is not issuing any waivers of the fraud and abuse provisions in sections 1128A, 1128B, and 1877 of the SSA or any other Medicare or Medicaid fraud and abuse laws for this model. Thus, notwithstanding any other provisions of this proposed rule, all providers and suppliers participating in the HHVBP model must comply with all applicable fraud and abuse laws and regulations.
We are proposing to use the section 1115A(d)(1) waiver authority to apply a reduction or increase of up to 8 percent to current Medicare payments to Medicare-certified HHAs delivering care to beneficiaries within the boundaries of certain states, depending on the HHA's performance on specified quality measures relative to its peers. Specifically, the HHVBP model proposes to utilize the waiver authority to adjust Medicare payment rates under section 1895(b) of the Act.
The defined population would include all Medicare beneficiaries being provided care by any Medicare-certified HHA delivering care within the selected states. Medicare-certified HHAs that are delivering care within the boundaries of selected states are considered `Competing Medicare-certified Home Health Agencies' within the scope of this HHVBP Model. If care is delivered outside of boundaries of selected states, or inside the boundaries of a non-selected state that does not have a reciprocal agreement with a selected state, payments for those beneficiaries would not be considered within the scope of the model because we are basing participation in the model on state specific CMS Certification Numbers (CCNs). Payment adjustments for each year of the model would be calculated based on a comparison of how well each competing Medicare-certified HHA performed during the performance period for that year (proposed below to be one year in length, starting in CY 2016) with its performance on the same measures in 2015 (proposed below to be the baseline data year).
The first performance year would be CY 2016, the second would be CY 2017, the third would be CY 2018, the fourth would be 2019, and the fifth would be CY 2020. Greater details on performance periods are outlined in further detail in section D—Performance Assessment and Payment Periods. This model would test whether being subject to significant payment adjustments to the Medicare payment amounts that would otherwise be made to competing Medicare-certified HHAs would result in statistically significant improvements in the quality of care being delivered to this specific population of Medicare beneficiaries.
We propose to identify Medicare-certified HHAs for participation in this model using state borders as boundaries. We do so under the authority granted in section 1115A(a)(5) of the Act to elect to limit testing of a model to certain geographic areas. This decision is influenced by the 2012 Report to Congress mandated under section 3006(b) of the Affordable Care Act. This Report stated that HHAs which participated in previous value-based purchasing demonstrations “uniformly believed that all Medicare-certified HHAs should be required to participate in future VBP programs so all agencies experience the potential burdens and benefits of the program” and some HHAs expressed concern that absent mandatory participation, “low-performing agencies in areas with limited competition may not choose to pursue quality improvement.”
Section 1115A(b)(2)(A) of the Act requires that the Secretary select models to be tested where the Secretary determines that there is evidence that the model addresses a defined population for which there are deficits in care leading to poor clinical outcomes or potentially avoidable expenditures. The HHVBP model was developed to improve care for Medicare patients receiving care from HHAs based on evidence in the March 2014 MedPAC Report to Congress citing quality and cost concerns in the home health sector. According to MedPAC, “about 29 percent of post-hospital home health stays result in readmission, and there is tremendous variation in performance among providers within and across geographic regions.”
In § 484.305 we propose definitions for “applicable percent”, “applicable measure”, “benchmark”, “home health prospective payment system”, “larger-volume cohort”, “linear exchange function”, “Medicare-certified home health agency”, “New Measures”, “payment adjustment”, “performance period”, “smaller-volume cohort”, “selected states”, “starter set”, “Total Performance Score”, and “value-based purchasing” as they pertain to this subpart. The HHVBP model is being proposed to encompass five performance years and be implemented beginning January 1, 2016 and conclude on December 31, 2022. Payment and service delivery models are developed by CMMI in accordance with the requirements of section 1115A of the Act. During the development of new models, CMMI builds on the ideas received from internal and external stakeholders and consults with clinical and analytical experts.
In this proposed rule, we are outlining an HHVBP model for public notice and comment that has an overall purpose of improving the quality of home health care and delivering it to the Medicare population in a more efficient manner. The specific goals of the proposed model are to:
1. Incentivize HHAs to provide better quality care with greater efficiency;
2. Study new potential quality and efficiency measures for appropriateness in the home health setting; and,
3. Enhance current public reporting processes.
We are proposing that the HHVBP model would adjust Medicare HHA payments over the course of the model by up to 8 percent depending on the applicable performance year and the degree of quality performance demonstrated by each competing Medicare-certified HHA. The proposed model would reduce the HH PPS final claim payment amount to an HHA for each episode in a calendar year by an amount up to the applicable percentage defined in proposed § 484.305. The timeline of payment adjustments as they apply to each performance year is described in greater detail in the section entitled “Payment Adjustment Timeline.”
The model would apply to all Medicare-certified HHAs in each of the selected states, which means that all HHAs in the selected states would be required to compete. We propose to codify this policy at 42 CFR 484.310. Furthermore, a competing Medicare-certified HHA would only be measured on performance for care delivered to Medicare beneficiaries within selected states (with rare exceptions given for care delivered when a reciprocal agreement exists between states). The distribution of payment adjustments would be based on quality performance, as measured by both achievement and improvement, across a proposed set of quality measures rigorously constructed to minimize burden as much as possible and improve care. Competing Medicare-certified HHAs that demonstrate they can deliver higher quality of care in comparison to their peers (as defined by the volume of services delivered within the selected state), or their own past performance, could have their payment for each episode of care adjusted higher than the amount that otherwise would be paid under section 1895 of the Act. Competing Medicare-certified HHAs that do not perform as well as other competing Medicare-certified HHAs of the same size in the same state might have their payments reduced and those competing Medicare-certified HHAs that perform similarly to others of similar size in the same state might have no payment adjustment made. This operational concept is similar in practice to what is used in the HVBP program.
We expect that the risk of having payments adjusted in this manner would provide an incentive among all competing Medicare-certified HHAs delivering care within the boundaries of selected states to provide significantly better quality through improved planning, coordination, and management of care. The degree of the payment adjustment would be dependent on the level of quality achieved or improved from the baseline year, with the highest upward performance adjustments going to competing Medicare-certified HHAs with the highest overall level of performance based on either achievement or improvement in quality. The size of a Medicare-certified HHA's payment adjustment for each year under the model would be dependent upon that HHA's performance with respect to that calendar year relative to other competing Medicare-certified HHAs of similar size in the same state and relative to its own performance during the baseline year.
We are proposing that states would be selected randomly from nine regional groupings for model participation. A competing Medicare-certified HHA is only measured on performance for care delivered to Medicare beneficiaries within boundaries of selected states and only payments for HHA services provided to Medicare beneficiaries within boundaries of selected states would be subject to adjustment under the proposed model. Requiring all Medicare-certified HHAs within the boundaries of selected states to compete in the model would ensure that: (1) There is no self-selection bias, (2) competing HHAs are representative of HHAs nationally, and (3) there is sufficient participation to generate meaningful results. We believe it is necessary to require all HHAs delivering care within boundaries of selected states to be included in the model because, in our experience, Medicare-providers are generally reluctant to participate voluntarily in models in which their Medicare payments could be subject to possible reduction. This reluctance to participate in voluntary models has been shown to cause self-selection bias in statistical assessments and thus, may present challenges to our ability to evaluate the model. In addition, state boundaries represent a natural demarcation in how quality is currently being assessed through OASIS measures on Home Health Compare (HHC).
We are proposing to adopt a methodology that uses state borders as boundaries for demarcating which Medicare-certified HHAs will be required to compete in the model. We are proposing to select nine states from nine geographically-defined groupings of five or six states. Groupings were also defined in order to ensure that the successful implementation of the model would produce robust and generalizable results, as discussed later in this section.
We took into account five key factors when deciding to propose selection at the state-level for this model. First, if we required some, but not all, Medicare-certified HHAs that deliver care within the boundaries of a selected state to participate in the model, we believe the HHA market for the state could be disrupted because HHAs in the model would be competing against HHAs not in the model (herein referenced as either `non-model HHAs' or `non-competing HHAs'). Second, we wanted to ensure that the distribution of payment adjustments based on performance under the model could be extrapolated to the entire country. Statistically, the larger the sample to which payment adjustments are applied, the smaller the variance of the sampling distribution and the greater the likelihood that the distribution accurately predicts what would transpire if the methodology were applied to the full population of
We expect that when there is a risk for a downside payment adjustment based on quality performance measures, the use of a self-contained, mandatory cohort of HHA participants will create a stronger incentive to deliver greater quality among competing Medicare-certified HHAs. Specifically, it is possible the market would become distorted if non-model HHAs are delivering care within the same market as competing Medicare-certified HHAs because competition, on the whole, becomes unfair when payment is predicated on quality for one group and volume for the other group. In addition, we expect that evaluation efforts might be negatively impacted because some HHAs would be competing on quality and others on volume within the same market.
We are proposing the use of state boundaries after careful consideration of several alternative selection approaches, including randomly selecting HHAs from all HHAs across the country, and requiring participation from smaller geographic regions including the county; the Combined Statistical Area (CSA); the Core-Based Statistical Area (CBSA); rural provider level; and the Hospital Referral Region (HRR) level.
A methodology using a national sample of HHAs that are randomly selected from all HHAs across the country could be designed to include enough HHAs to ensure robust payment adjustment distribution and a sufficient sample size for the evaluation; however, this approach may present significant limitations when compared with the state boundaries selection methodology proposed in this model. Of primary concern with randomly selecting at the provider-level across the nation is the issue with market distortions created by having competing Medicare-certified HHAs operating in the same market as non-model HHAs.
Using smaller geographic areas than states, such as counties, CSAs, CBSAs, rural, and HRRs, could also present challenges for this model. These smaller geographic areas were considered as alternate selection options; however, their use could result in too small of a sample size of potential competing HHAs. As a result, we expect the distribution of payment adjustments could become highly divergent among fewer HHA competitors. In addition, the ability to evaluate the model could become more complex and may be less generalizable to the full population of Medicare-certified HHAs and the beneficiaries they serve across the nation. Further, the use of smaller geographic areas than states could increase the proportion of Medicare-certified HHAs that could fall into groupings with too few agencies to generate a stable distribution of payment adjustments. Thus, if we were to define geographic areas based on CSAs, CBSAs, counties, or HRRs, we would need to develop an approach for consolidating smaller regions into larger regions.
Home health care is a unique type of health care service when compared to other Medicare provider types. In general, the HHA's care delivery setting is in the beneficiaries' homes as opposed to other provider types that traditionally deliver care at a brick and mortar institution within beneficiaries' respective communities. As a result, the HHVBP model needs to be designed to account for the unique way that HHA care is provided in order to ensure that the results are generalizable to the population. HHAs are limited to providing care to beneficiaries in the state that they have a CCN however; HHAs are not restricted from providing service in a county, CSA, CBSA or HRR that they are not located in (as long as the other county/CBSA/HRR is in the same state in which the HHA is certified). As a result, using smaller geographic areas (than state boundaries) could result in similar market distortion and evaluation confounders as selecting providers from a randomized national sampling. The reason is that HHAs in adjacent counties/CSAs/CBSAs/HRRs may not be in the model but, would be directly competing for services in the same markets or geographic regions. Competing HHAs delivering care in the same market area as non-competing HHAs could generate a spillover effect where non-model HHAs would be vying for the same beneficiaries as competing HHAs. This spillover effect presents several issues for evaluation as the dependent variable (quality) becomes confounded by external influences created by these non-competing HHAs. These unintentional external influences on competing HHAs may be made apparent if non-competing HHAs become incentivized to generate greater volume at the expense of quality delivered to the beneficiaries they serve and at the expense of competing HHAs that are paid on quality instead of volume. Further, the ability to extrapolate these results to the full population of HHAs and the beneficiaries they serve becomes confounded by an artifact of the model and inferences would be limited from an inability to duplicate these results. While these concerns would decrease in some order of magnitude as larger regions are considered, the only way to eliminate these concerns entirely is to define participation among Medicare-certified HHAs at the state level.
In addition, home health quality data currently displayed on HHC allows users to compare HHA services furnished within a single state. Selecting HHAs using other geographic regions that are smaller and/or cross state lines could require the model to deviate from the established process for reporting quality. For these reasons, we believe a selection methodology based on the use of Medicare-certified HHAs delivering care within state boundaries would be the most appropriate for the successful implementation and evaluation of this model.
While, for the reasons described above, we are proposing that the geographic basis of selection remain at the state-level, we nevertheless seek comment on potential alternatives that might use smaller geographic areas. With consideration of alternatives, the public should reference the five aforementioned key factors used to consider selection at the state-level for this model as they relate to the evaluative framework and operational feasibility of this model. In particular, one potential alternative would be to split states into sub-state regions using a combination of CSAs and metropolitan statistical areas (MSA), a type of CBSA. For example, regions might be defined using the following process:
•
•
•
•
We seek comment on advantages and disadvantages of this approach relative to defining regions based on state boundaries. In particular, we note that because this approach would generate a larger number of regions, it could increase the statistical power of the model evaluation, and might improve our ability to determine what effects the model has on the quality of home health care, as well as other outcomes of interest. However, we note that because regions would no longer line up with full states in most cases, the regions selected to participate in the model would no longer align directly with those displayed on HHC and therefore, quality data would have to be recalculated and displayed differently from what is currently being reported on HHC. In addition, using sub-state regions could, as noted above, lead to undesirable spillover effects between participating and non-participating HHAs. These spillover concerns would be mitigated by the fact that none of the sub-state regions defined under this approach would cross state lines and the fact that the sub-state regions would be larger than under some approaches to defining sub-state regions (for example, at the county level). Nevertheless, it is unclear how severe these evaluation and operational concerns would be in practice and how the extent of these concerns would depend on the different characteristics of the selected regions. We welcome public comment on these proposed state selection methodologies.
We are requesting comments on the following proposed methodology for selecting states. The selection methodology employed will need to provide the strongest evidence of producing meaningful results representative of the national population of Medicare-certified HHAs and, in turn, meet the evaluation requirements of section 1115A(b)(4) of the Act.
The state selections listed in proposed § 484.310 are based on the described proposed randomized selection methodology and are subject to change in the CY 2016 HH PPS final rule as a result of any changes that may be made to the proposed randomized methodology in response to comments. However, if the final methodology differs from what we are proposing here, we will apply the final methodology and identify the states selected under the final methodology in the final rule. We propose to group states by each state's geographic proximity to one another and by accounting for key evaluation characteristics (that is, proportionality of service utilization, proportionality of organizations with similar tax-exempt status and HHA size, and proportionality of beneficiaries that are dually-eligible for Medicare and Medicaid).
Based on an analysis of OASIS quality data and Medicare claims data, we believe the use of nine geographic groupings is necessary to ensure that the model accounts for the diversity of beneficiary demographics, rural and urban status, cost and quality variations, among other criteria. To provide for comparable and equitable selection probabilities, these separate geographic groupings each include a comparable number of states. We are not proposing to adopt census-based geographic groupings or the CMS Medicare Administrative Contractor (MAC) jurisdictions because those groupings would not permit an equal opportunity of selection of Medicare-certified HHAs by state or an assurance that we would be able test the model among a diversity of agencies such as is found across the nation. Following this logic, under our proposed methodology, groupings are based on states' geographic proximity to one another, having a comparable number of states if randomized for an equal opportunity of selection, and similarities in key characteristics that would be considered in the evaluation study because the attributes represent different types of HHAs, regulatory oversight, and types of beneficiaries served. This is necessary to ensure that the evaluation study remains objective and unbiased and that the results of this study best represent the entire population of Medicare-certified HHAs across the nation.
Several of the key characteristics we used for grouping state boundaries into clusters for selection into the model are also used in the impact analysis of our annual HHA payment updates, a fact that reinforces their relevance for evaluation. The additional proposed standards for grouping (level of utilization and socioeconomic status of patients) are also important to consider when evaluating the program, because of their current policy relevance. Large variations in the level of utilization of the home health benefit has received attention from policymakers concerned with achieving high-value health care and curbing fraud and abuse.
Under the proposed methodology, in order to ensure that the Medicare-certified HHAs that would be required to participate in the model are not all in one region of the country, the states in each grouping are adjacent to each other whenever possible while creating logical groupings of states based on common characteristics as described above. Specifically, analysis based on quality data and claims data found that HHAs in these neighboring states tend to hold certain characteristics in common. These include having similar; patterns of utilization, proportionality of non-profit agencies, and types of beneficiaries served (for example, severity and number, type of co-
Under our proposed randomized selection methodology, each geographic region, or grouping, has a similar number of states. As a result, all states would have a 16.7 percent to 20 percent chance of being selected under our proposed methodology, and Medicare-certified HHAs would have a similar likelihood of being required to compete in the model by using this sampling design. We assert that this sampling design would ensure that no single entity is singled out for selection, since all states and Medicare-certified HHAs would have approximately the same chance of being selected. In addition, this sampling approach would mitigate the opportunity for HHAs to self-select into the model and thereby bias any results of the test.
Without sacrificing an equal opportunity for selection, the proposed state groupings are intended to ensure that important characteristics of Medicare-certified HHAs that deliver care within state boundaries can be used to evaluate the primary intervention with greater generalizability and representativeness of the entire population of Medicare-certified HHAs in the nation. Data analysis of these characteristics employed the full data set of Medicare claims and OASIS quality data. Although some characteristics, such as beneficiary age and case-mix, yield some variations from one state to another, other important characteristics do vary substantially and could influence how HHAs respond to the incentives of the model. Specifically, home health services utilization rates, tax-exemption status of the provider, the socioeconomic status of beneficiaries (as measured by the proportion of dually-eligible beneficiaries), and agency size (as measured by average number of episodes of care per HHA), are important characteristics that could influence outcomes of the model. Subsequently, we intend to study the impacts of these characteristics for purposes of designing future value-based purchasing models and programs. These characteristics and expected variations must be considered in the evaluation study to enable us to avoid erroneous inferences about how different types of HHAs will respond to HHVBP incentives.
Under this proposed state selection methodology, state groupings reflect regional variations that enhance the generalizability of the model. In line with this methodology, each grouping includes states that are similar in at least one important aforementioned characteristic while being geographically located in close proximity to one another. Using the criteria described above, the following geographic groupings were identified using Medicare claims-based data from calendar years 2013–2014. Each of the 50 states was assigned to one of the following geographic groups:
• Group #1:
States in this group tend to have larger HHAs and have average utilization relative to other states.
• Group #2:
States in this group tend to have larger HHAs, have lower utilization, and provide care to an average number of dually-eligible beneficiaries relative to other states.
• Group #3:
States in this group tend to have larger HHAs, have average utilization rates, and provide care to a high proportion of minorities relative to other states.
• Group #4:
States in this group have HHAs that tend to be for-profit, have very high utilization rates, and have a higher proportion of dually-eligible beneficiaries relative to other states.
• Group #5:
States in this group tend to have smaller HHAs, have average utilization rates, and are more rural relative to other states.
• Group #6:
States in this group tend to have smaller HHAs, have average utilization rates, and provide care to a high proportion of minorities relative to other states.
• Group #7:
States in this group tend to have smaller HHAs, have very low utilization rates, and are more rural relative to other states.
• Group #8:
States in this group tend to have HHAs that are of average size, have average utilization rates, and provide care to a higher proportion of dually-eligible beneficiaries relative to other states.
• Group #9:
States in this group tend to have HHAs with higher utilization rates relative to other states.
Upon the careful consideration of the aforementioned alternative selection methodologies, including selecting states on a non-random basis, we choose to propose the use of a selection methodology based on a randomized sampling of states within each of the nine regional groupings described above. We examined data on the evaluation elements listed in this section to determine if specific states could be identified in order to fulfill the needs of the evaluation. After careful review, we determined that each evaluation element could be measured by more than one state. As a result, we determined that it was necessary to apply a fair method of selection where each state would have a comparable opportunity of being selected and which would fulfill the need for a robust evaluation. The proposed nine groupings of states as described in this section permit the model to capture the essential elements of the evaluation including demographic, geographic, and market factors.
The randomized sampling of states is without bias to any characteristics of any single state within any specific regional grouping, where no states are excluded, and no state appears more than once across any of the groupings. The randomized selection of states was completed using a scientifically-accepted computer algorithm designed for randomized sampling. The randomized selection of states was run on each of the previously described regional groupings using exactly the same process and, therefore, reflects a commonly accepted method of randomized sampling. This computer algorithm employs the aforementioned sampling parameters necessary to define randomized sampling and omits any human interaction once it runs.
Based on this sampling methodology, SAS Enterprise Guide (SAS EG) 5.1 software was used to run a computer algorithm designed to randomly select states from each grouping. SAS EG 5.1 and the computer algorithm were employed to conduct the randomized selection of states. SAS EG 5.1 represents an industry-standard for generating advanced analytics and provided a rigorous, standardized tool by which to satisfy the requirements of randomized selection. The key SAS commands employed include a “PROC
In § 484.310, we propose to codify the names of the states selected utilizing this proposed methodology, where one state from each of the nine groupings was selected. For each of these groupings, we propose to use state borders to demarcate which Medicare certified HHAs would be required to compete in this model: Massachusetts was randomly selected from Group 1, Maryland was randomly selected from Group 2, North Carolina was randomly selected from Group 3, Florida was randomly selected from Group 4, Washington was randomly selected from Group 5, Arizona was randomly selected from Group 6, Iowa was randomly selected from Group 7, Nebraska was randomly selected from Group 8, and Tennessee was randomly selected from Group 9. Thus, if our methodology is finalized as proposed, all Medicare-certified HHAs that provide services in Massachusetts, Maryland, North Carolina, Florida, Washington, Arizona, Iowa, Nebraska, and Tennessee will be required to compete in this model.
However, should the methodology we propose in this rule change as a result of comments received during the rulemaking process, it could result in different states being selected for the model. In such an event, we would apply the final methodology and announce the selected states in the final rule. We therefore seek comment from all interested parties in every state on the randomized selection methodology proposed above and codified at § 484.310.
Based on the comments received from this proposed rule, the selection methodology for participation in the model may change from state boundaries to an approach based on sub-state regions built from CSAs/MSAs, CBSAs, rural provider level or HRRs. In that case, the goals of the model will remain the same, and therefore, we would expect to take a broadly similar approach to selecting participating regions to the approach that would be taken when regions are defined based on state boundaries. Specifically, as with the selection methodology outlined above, we would anticipate grouping sub-state regions together based on geographic proximity and other characteristics into groups of approximately equal size and then selecting some number of sub-state regions to participate from each group. The number of selected participants will be dependent on the selection methodology. We welcome public comment on these proposed state selection methodologies.
We are proposing that Total Performance Scores (TPS) and payment adjustments would be calculated based on an HHA's CCN
We are proposing the use of quarterly performance reports, annual payment adjustment reports, and annual publicly-available performance reports as a means of developing greater transparency of Medicare data on quality and aligning the competitive forces within the market to deliver care based on value over volume. The publicly-reported reports would inform home health industry stakeholders (consumers, physicians, hospitals) as well as all competing HHAs delivering care to Medicare beneficiaries within selected state boundaries on their level of quality relative to both their peers and their own past performance.
Competing HHAs would be scored for the quality of care delivered under the model based on their performance on measures compared to both the performance of their peers, defined by the same size cohort (either smaller- or larger-volume cohorts as defined in § 484.305), and their own past performance on the measures. We propose in § 484.305 to define larger-volume cohort to mean the group of Medicare-certified HHAs within the boundaries of a selected state that are participating in HHCAHPs in accordance with § 484.250 and to define smaller-volume cohort to mean the group of HHAs within the boundaries of a selected state that are exempt from participation in HHCAHPs in accordance with § 484.250. Where there are too few HHAs in the smaller-volume cohort in each state to compete in a fair manner (that is, when there is only one or two HHAs competing within a specific cohort), these specific HHAs would be included in the larger-volume cohort [for purposes of calculating the total performance score and payment adjustment] without being measured on HHCAHPS. We are requesting comments on this proposed methodology.
Quality performance scores and relative peer rankings would be determined through the use of a baseline year (calendar year 2015) and subsequent performance periods for each competing HHA. Further, these reports would provide competing HHAs with an opportunity to track their quality performance relative to their peers and their own past performance. Using these reports provides a convenient and timely means for competing HHAs to assess and track their own respective performance as capacity is developed to improve or sustain quality over time.
Beginning with the data collected during the first quarter of CY 2016 (that is, data for the period January 1, 2016 to March 31, 2016), and for every quarter of the model thereafter, we are proposing to provide each Medicare certified HHA with a quarterly report that contains information on their performance during the quarter. We expect to make the first quarterly report available in July 2016, and to make performance reports for subsequent quarters available in October, January and April. The final quarterly report would be made available in April 2021. The quarterly reports would include a competing HHA's model-specific performance results with a comparison to other competing HHAs within its cohort (larger- or smaller-volume) within the state boundary. These model-specific performance results would complement all quality data sources already being provided through the QIES system and any other quality tracking system possibly being employed by HHAs. We note that all performance measures that Medicare-certified HHAs will report through the QIES system are also already made available in the CASPER Reporting application. The primary difference between the two reports (CASPER reports and the model-specific performance report) is that the model-specific performance report we are proposing here consolidates the applicable performance measures used in the HHVBP model and provides a peer-ranking to other competing Medicare-certified HHAs within the same state and size-cohort. In addition, CASPER reports would provide quality data earlier than model-specific performance reports because CASPER reports are not limited by a quarterly run-out of data and a calculation of competing peer-rankings. For more information on the accessibility and functionality of the CASPER system, please reference the CASPER Provider Reporting Guide.
The model-specific quarterly performance report would be made available to each HHA through a dedicated CMMI model-specific platform for data dissemination and include each HHA's relative ranking amongst its peers along with measurement scores and overall performance rankings.
We are proposing that a separate payment adjustment report would be provided once a year to each of the competing HHAs. This report would focus primarily on the payment adjustment percentage and include an explanation of when the adjustment would be applied and how this adjustment was determined relative to performance scores. Each competing HHA would receive its own payment adjustment report viewable only to that HHA.
We are also proposing a separate, annual, publicly available quality report that would provide home health industry stakeholders, including providers and suppliers that refer their patients to HHAs, with an opportunity to ensure that the beneficiaries they are referring for home health services are being provided the best possible quality of care available. We seek public comment on the proposed reporting framework described above.
We propose at § 484.325 that Medicare-certified HHAs will be subject to upward or downward payment adjustments based on performance on quality measures. We propose this model would consist of 5 performance years, where each performance year would link performance to the opportunity and risk for payment adjustment up to an applicable percent as defined in proposed 42 CFR 484.305. The first performance year would transpire from January 1, 2016 through December 31, 2016, and subsequently, all other performance years would be assessed on an annual basis through 2020, unless modified through rulemaking. The first payment adjustment would begin January 1, 2018 applied to that calendar year based on 2016 performance data. Subsequently, all other payment adjustments would be made on an annual basis through the conclusion of the model, unless modified through rulemaking. We are proposing that payment adjustments will be increased incrementally over the course of the model with a maximum payment adjustment of (5 percent) upward or downward in 2018 and 2019, a maximum payment adjustment of 6 percent (upward or downward) in 2020, and a maximum payment adjustment of 8 percent (upward or downward) in 2021 and 2022. We propose to implement this model over a total of 7 years beginning on January 1, 2016, and ending on December 31, 2022.
The baseline year would run from January 1, 2015 through December 31, 2015 and provide a basis from which each respective HHA's performance would be measured in each of the performance years. Data related to performance on quality measures would continue to be provided from the baseline year through the model's tenure using a dedicated HHVBP web-based platform specifically designed to disseminate data in this model (this “portal” would present and archive the previously described quarterly and annual quality reports). Further, HHAs will provide performance data on the four new quality measures through this platform as well. Any new measures employed through the model's tenure, subject to rulemaking, would use data from the previous calendar year as the baseline.
New market entries (specifically, new Medicare-certified HHAs delivering care in the boundaries of selected states) would also be measured from their first full calendar year of services in the state, which would be treated as baseline data for subsequent performance years under this model. The delivery of services would be measured by the number of episodes of care for Medicare beneficiaries and used to determine whether an HHA falls into the smaller- or larger- volume cohort. Furthermore, these new market entries would be competing under the HHVBP model in the first full calendar year following the full calendar year baseline period.
HHAs would be notified in advance of their first performance level and payment adjustment being finalized, based on the 2016 performance period (January 1, 2016 to December 31, 2016), with their first payment adjustment to be applied January 1, 2018 through December 31, 2018. Each HHA would be notified of this first pending payment adjustment on August 1, 2017 and a preview period would run for 10 days through August 11, 2017. This preview period would provide each competing HHA an opportunity to reconcile any performance assessment issues relating to the calculation of scores prior to the payment adjustment taking effect, in accordance with the process proposed in section H—Preview and Period to Request Recalculation. Once the preview period ends, any changes would be reconciled and a report finalized no later than November 1, 2017 (or 60 days prior to the payment adjustment taking affect).
Subsequent payment adjustments would be calculated based on the applicable full calendar year of performance data from the quarterly reports, with HHAs notified and payments adjusted, respectively, every year thereafter. As a sequential example, the second payment adjustment would occur January 1, 2019 based on a full 12 months of the CY 2017 performance period. Notification of the adjustment
Beginning in CY 2019, we may consider revising this payment adjustment schedule and updating the payment adjustment more frequently than once each year if it is determined that a more timely application of the adjustment as it relates to performance improvement efforts that have transpired over the course of a calendar year would generate increased improvement in quality measures. Specifically, we would expect that having payment adjustments transpire closer together through more frequent performance periods would accelerate improvement in quality measures because HHAs would be able to justify earlier investments in quality efforts and be incentivized for improvements. In effect, this concept may be operationalized to create a smoothing effect where payment adjustments are based on overlapping 12-month performance periods that occur every 6 months rather than annually. As an example, the normal 12-month performance period occurring from January 1, 2020 to December 31, 2020 might have an overlapping 12-month performance period occurring from July 1, 2020 to June 30, 2021. Following the regularly scheduled January 1, 2022 payment adjustments, the next adjustments could be applied to payments beginning on July 1, 2022 through December 31, 2022. Depending on if and when more frequent payment adjustments would be applied, performance would be calculated based on the applicable 12-months of performance data, HHAs notified, and payments adjusted, respectively, every six months thereafter, until the conclusion of the model. As a result, separate performance periods would have a 6-month overlap through the conclusion of the model. HHAs would be notified through rulemaking and be given the opportunity to comment on any proposed changes to the frequency of payment adjustments. We seek public comment on the proposed payment adjustment schedule described above.
Initially, we propose the measures for the HHVBP model would be predominantly drawn from the current Outcome and Assessment Information Set (OASIS),
1. Use a broad measure set that captures the complexity of the HHA service provided;
2. Incorporate the flexibility to include Improving Medicare Post-Acute Care Transformation (IMPACT) Act of 2014 proposed measures that are cross-cutting amongst post-acute care settings;
3. Develop second-generation measures of patient outcomes, health and functional status, shared decision making, and patient activation;
4. Include a balance of process, outcome, and patient experience measures;
5. Advance the ability to measure cost and value;
6. Add measures for appropriateness or overuse; and,
7. Promote infrastructure investments.
A central driver of the proposed measure selection process was incorporating innovative thinking from the field while simultaneously drawing on the most current evidence-based literature and documented best practices. Broadly, we propose measures that have a high impact on care delivery and support the combined priorities of HHS and CMS to improve health outcomes, quality, safety, efficiency, and experience of care for patients. To frame the selection process, we utilized the domains described in the CMS Quality Strategy that maps to the six National Quality Strategy (NQS) priority areas (
We propose at § 484.315 that Medicare-certified HHAs would be evaluated using a starter set of quality measures (“starter set” refers to the proposed quality measures for the first year of this model) designed to encompass multiple NQS domains, and provide future flexibility to incorporate and study newly developed measures over time. New and evolving measures would be considered for inclusion in subsequent years of this model and proposed through future rulemaking.
To create the proposed starter set we began researching the current set of OASIS measures that are being used within the health home environment.
To ensure proposed measures for the HHVBP model take a more holistic view of the patient beyond a particular disease state or care setting, we are proposing measures, which include outcome measures as well as process measures, that have the potential to follow patients across multiple settings, reflect a multi-faceted approach, and foster the intersection of health care delivery and population health. A key consideration behind this approach is to use in performance year one (PY1) of the model proven measures that are readily available and meet a high impact need, and in subsequent model years augment this starter set with innovative measures that have the potential to be impactful and fill critical measure gap areas. All substantive changes or additions to the proposed starter set or new measures would be proposed for inclusion in future rulemaking. This approach to quality measure selection aims to balance the burden of collecting data with the inclusion of new and important measures. We carefully considered the potential burden on HHAs to report the measure data when developing the proposed starter set, and prioritized proposed measures that would draw both from claims data and data already collected in OASIS.
The majority of the proposed measures in this model would use OASIS data currently being reported to CMS and linked to state-specific CCNs for selected states in order to promote consistency and to reduce the data collection burden for providers. Utilizing primarily OASIS data would allow the model to leverage reporting structures already in place to evaluate performance and identify weaknesses in care delivery. This model would also afford the opportunity to study measures developed in other care settings and new to the home health industry (hereinafter referred to as “New Measures”). Many of the proposed New Measures have been used in other health care settings and are readily applicable to the home health environment (for example, influenza vaccination coverage for health care personnel). Proposed New Measures for PY1 are described in detail below. We
The initial set of measures proposed for PY1 of the model utilizes data collected via OASIS, Medicare claims, HHCAHPS survey data, and data reported directly from the HHAs to CMS. In total there are 10 process measures and 15 outcome measures (
We seek public comment on the methodology for constructing the proposed starter set of quality measures and on the proposed
Figure 5 provides details on the elements of the Home Health Care Consumer Assessment of Healthcare Providers and Systems Survey (HHCAHPS) we propose to include in the PY1 starter set. The HHVBP model would not alter the HHCAHPS current scoring methodology or the participation requirements in any way. Details on participation requirements for HHCAHPS can be found at 42 CFR 484.250
As discussed in the previous section, the New Measures we propose are not currently reported by Medicare-certified HHAs to CMS, but we believe fill gaps in the NQS Domains not completely covered by existing measures in the home health setting. All Medicare-certified HHAs in selected states, regardless of cohort size or number of episodes, will be required to submit data on the New Measures for all Medicare beneficiaries to whom they provide home health services within the state (unless an exception applies). We propose at § 484.315 that HHAs will be required to report data on these New Measures. Competing Medicare-certified HHAs would submit data through a dedicated HHVBP web-based platform. This web-based platform would function as a means to collect and distribute information from and to competing Medicare-certified HHAs. Also, for those HHAs with a sufficient number of episodes of care to be subject to a payment adjustment, New Measures scores included in the final TPS for PY1 are only based on whether the HHA has submitted data to the HHVBP web-based platform or not. We are proposing the following New Measures for competing Medicare-certified HHAs:
• Advance Care Planning;
• Adverse Event for Improper Medication Administration and/or Side Effects;
• Influenza Vaccination Coverage for Home Health Care Personnel; and,
• Herpes Zoster (Shingles) Vaccination received by HHA patients.
Advance Care Planning is an NQF-endorsed process measure in the NQS domain of Person- and Caregiver-centered experience and outcomes (see Figure 3). This measure is currently endorsed at the group practice/individual clinician level of analysis. We believe its adoption under the HHVBP model represents an opportunity to study this measure in the home health setting. This is an especially pertinent measure for home health care to ensure that the wishes of the patient regarding their medical, emotional, or social needs are met across care settings. The Advance Care Planning measure would focus on Medicare beneficiaries, including dually-eligible beneficiaries.
The measure would be numerically expressed by a ratio whose numerator and denominator are as follows:
Information on this numerator and denominator would be reported by HHAs through the HHVBP web-based platform, in addition to other information related to this measure as the Secretary deems appropriate.
Advance care planning ensures that the health care plan is consistent with the patient's wishes and preferences. Therefore, studying this measure within the HHA environment allows for further analysis of planning for the “what ifs” that may occur during the patient's lifetime. In addition, the use of this measure is expected to result in an increase in the number of patients with advance care plans. Increased advance care planning among the elderly is expected to result in enhanced patient autonomy and reduced hospitalizations and in-hospital deaths.
We welcome public comments on this measure's proposed adoption under the HHVBP model.
Adverse Event for Improper Medication Administration and/or Side Effects is a measure that aligns with the NQS domain of Safety (specifically “medication safety”—
An adverse drug event (ADE) is an injury related to medication use.
The annual incidence of ADEs in health care in the United States is high; authoritative estimates indicate that each year 400,000 preventable ADEs occur in hospitals, 800,000 in long term care settings and in excess of 500,000 among Medicare patients in outpatient settings.
The measure would be numerically expressed by a ratio whose numerator and denominator are as follows:
We welcome public comments on this measure's proposed adoption under the HHVBP model.
Staff Immunizations (Influenza Vaccination Coverage among Health Care Personnel) (NQF #0431) is an NQF-endorsed measure that addresses the NQS domain of Population Health (
The measure would be numerically expressed by a ratio whose numerator and denominator are as follows:
(1) Received an influenza vaccination administered at the health care agency, or reported in writing (paper or electronic) or provided documentation that influenza vaccination was received elsewhere; or
(2) Were determined to have a medical contraindication/condition of severe allergic reaction to eggs or to other component(s) of the vaccine, or history of Guillain-Barré Syndrome within 6 weeks after a previous influenza vaccination; or
(3) Declined influenza vaccination; or
(4) Persons with unknown vaccination status or who do not otherwise meet any of the definitions of the above-mentioned numerator categories.
Each of the above groups would be divided by the number of health care personnel who are working in the HHA for at least one working day between October 1 and March 31 of the following year, regardless of clinical responsibility or patient contact.
1. Employees: All persons who receive a direct paycheck from the reporting HHA (that is, on the agency's payroll);
2. Licensed independent practitioners: Include physicians (MD, DO), advanced practice nurses, and physician assistants only who are affiliated with the reporting agency who do not receive a direct paycheck from the reporting HHA; and
3. Adult students/trainees and volunteers: Include all adult students/trainees and volunteers who do not receive a direct paycheck from the reporting HHA.
This proposed measure for the HHVBP model is expected to result in increased influenza vaccination among home health professionals. Reporting health care personnel influenza vaccination status would allow HHAs to better identify and target unvaccinated personnel. Increased influenza vaccination coverage among HHA personnel would be expected to result in reduced morbidity and mortality related to influenza virus infection among patients, especially elderly and vulnerable populations.
Information on the above numerator and denominator would be reported by HHAs through the HHVBP web-based platform, in addition to other information related to this measure as the Secretary deems appropriate. We welcome public comments on this measure's proposed adoption under the HHVBP model.
We are proposing to adopt this measure for the HHVBP model because it aligns with the NQS Quality Strategy Goal to Promote Effective Prevention & Treatment of Chronic Disease. Currently this proposed measure is not endorsed by NQF or collected in OASIS. However, due to the severe physical consequences of symptoms associated with shingles,
The Food and Drug Administration (FDA) has approved the use of herpes zoster vaccine in adults age 50 and older. In addition, the Advisory Committee on Immunization Practices (ACIP) currently recommends that herpes zoster vaccine be routinely administered to adults, age 60 years and older.
The incidence of herpes zoster outbreak increases as people age, with a significant increase after age 50. Older people are more likely to experience the severe nerve pain known as post-herpetic neuralgia (PHN),
Studying this measure in the home health setting presents an ideal opportunity to address a population at risk which would benefit greatly from this vaccination strategy. For example, receiving the vaccine will often reduce the course and severity of the disease and reduce the risk of post herpetic neuralgia.
Information on the above numerator and denominator would be reported by HHAs through the HHVBP web-based platform, in addition to other information related to this measure as the Secretary deems appropriate. We welcome public comments on this measure's proposed adoption under the HHVBP model.
As previously stated, the quality measures that we are proposing to use in the performance years are aligned with the six NQS domains: Patient and Caregiver-centered experience and outcomes; Clinical quality of care; Care coordination; Population Health; Efficiency and cost reduction; and, Safety (see Figure 6).
We propose to filter these NQS domains and the proposed HHVBP quality measures into four classifications to align directly with the measure weighting utilized in calculating payment adjustments. The four HHVBP classifications we are proposing are: Clinical Quality of Care, Outcome and Efficiency, Person- and Caregiver-Centered Experience, and New Measures reported by the HHAs.
These four classifications capture the multi-dimensional nature of health care provided by the HHA. These classifications are further defined as:
• Classification I—Clinical Quality of Care: Measures the quality of health care services provided by eligible professionals and paraprofessionals within the home health environment.
• Classification II—Outcome and Efficiency: Outcomes measure the end result of care provided to the beneficiary. Efficiencies measure maximizing quality and minimizing use of resources.
• Classification III—Person- and Caregiver-Centered Experience: Measures the beneficiary and their caregivers' experience of care.
• Classification IV—New Measures: Measures not currently reported by Medicare-certified HHAs to CMS, but that may fill gaps in the NQS Domains not completely covered by existing measures in the home health setting.
We seek public comment on our proposed measure classifications for the HHVBP model.
We propose that measures within each classification will be weighted the same for the purposes of payment adjustment. We are weighting at the individual measure level and not the classification level. Classifications are for organizational purposes only. We selected this approach since we did not want any one measure within a classification to be more important than another measure. This approach ensures that a measure's weight will remain the same even if some of the measures within a classification group have no available data. Weighting will be re-examined in subsequent years of the model and be subject to the rulemaking process.
We welcome public comments on this proposed weighting methodology under the HHVBP model.
The methodology we are proposing for assessing each HHA's total annual performance is based on a score calculated using the proposed starter set of quality measures that apply to the HHA (based on a minimum number of cases, as discussed herein). The methodology we propose would provide an assessment on a quarterly basis for each HHA and would result in an annual distribution of value-based payment adjustments among HHAs so that HHAs achieving the highest performance scores would receive the largest upward payment adjustment. The methodology we are proposing includes three primary features:
• The HHA's Total Performance Score (TPS) would be determined using the higher of an HHA's achievement or improvement score for each measure;
• All measures in the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications will have equal weight and will account for 90 percent of the TPS (see section 2 below) regardless of the number of measures in the three classifications. Points for New Measures are awarded for submission of data on the New Measures via the HHVBP web-based platform, and withheld if data is not submitted. Data reporting for each New Measure will have equal weight and will account for 10 percent of the TPS for the first performance year; and,
• The HHA performance score would reflect all of the measures that apply to the HHA based on a minimum number of cases defined below.
In § 484.320 we propose to calculate the TPS by adding together points awarded to Medicare-certified HHAs on the starter set of measures, including the New Measures. We considered several factors when developing the proposed performance scoring methodology for the HHVBP model. First, we believe it is important that the performance scoring methodology be straightforward and transparent to HHAs, patients, and other stakeholders. HHAs must be able to clearly understand performance scoring methods and performance expectations to maximize quality improvement efforts. The public must understand performance score methods to utilize publicly-reported information when choosing HHAs.
Second, we believe the proposed performance scoring methodology for the HHVBP model should be aligned appropriately with the quality measurements adopted for other Medicare value-based purchasing programs including those introduced in the hospital and skilled nursing home settings. This alignment would facilitate the public's understanding of quality measurement information disseminated in these programs and foster more informed consumer decision-making about their health care choices.
Third, we believe that differences in performance scores must reflect true differences in quality performance. To ensure that this point is addressed in the proposed performance scoring methodology for the HHVBP model, we assessed quantitative characteristics of the measures, including the current
Fourth, we believe that both quality achievement and improvement must be measured appropriately in the performance scoring methodology for the HHVBP model. The proposed methodology specifies that performance scores under the HHVBP model are calculated utilizing the higher of achievement or improvement scores for each measure. The impact of performance scores utilizing achievement and improvement on HHAs' behavior and the resulting payment implications was also considered. Using the higher of achievement or improvement scores allows the model to recognize HHAs that have made great improvements, though their measured performance score may still be relatively lower in comparison to other HHAs.
Fifth, through careful measure selection we intend to eliminate, or at least control for, unintended consequences such as undermining better outcomes to patients or rewarding inappropriate care. As discussed above, when available, NQF endorsed measures would be used. In addition we propose to adopt measures that we believe are closely associated with better outcomes in the HHA setting in order to incentivize genuine improvements and sustain positive achievement while retaining the integrity of the model.
Sixth, we intend to ensure the model utilizes the most currently available data to assess HHA performance. We recognize that these data would not be available instantaneously due to the time required to process quality measurement information accurately; however, we intend to make every effort to process data in the timeliest fashion. Using more current data would result in a more accurate performance score while recognizing that HHAs need time to report measure data.
Many of the key elements of the proposed HHVBP model performance scoring methodology would be aligned with the scoring methodology of the Hospital Value-Based Purchasing Program (HVBP) in order to leverage the rigorous analysis and review underpinning that Program's approach to value-based purchasing in the hospital sector. The HVBP Program includes as one of its core elements the scoring methodology included in the 2007 Report to Congress “Plan to Implement a Medicare Hospital Value-Based Purchasing Program” (hereinafter referred to as “The 2007 HVBP Report”).
In the HVBP Program, the Performance Assessment Model aggregates points on the individual quality measures across different quality measurement domains to calculate a hospital's TPS. Similarly, the proposed HHVBP model would aggregate points on individual measures across four measure classifications derived from the 6 CMS/NQS domains as described above (see Figure 3) to calculate the HHA's TPS. In addition, the proposed HHVBP payment methodology is also aligned with the HVBP Program with respect to evaluating an HHA's performance on each quality measure based on the higher of an achievement or improvement score in the performance period. The proposed model is not only designed to provide incentives for HHAs to provide the highest level of quality, but also to provide incentives for HHAs to improve the care they provide to Medicare beneficiaries. By rewarding HHAs that provide high quality and/or high improvement, we believe the proposed HHVBP model would ensure that all HHAs would be incentivized to commit the resources necessary to make the organizational changes that would result in better quality.
Under the proposed model an HHA would be awarded points only for “applicable measures.” An “applicable measure” is one for which the HHA has provided 20 home health episodes of care per year. Points awarded for each applicable measure would be aggregated to generate a TPS. As described in the benchmark section below, HHAs would have the opportunity to receive 0 to 10 points for each measure in the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications. Each measure would have equal weight regardless of the total number of measures in each of the first three classifications. In contrast, we propose to score the New Measures in a different way. For each New Measure, HHAs would receive 10 points if they report the New Measure or 0 points if they do not report the measure during the performance year. In total, the New Measures would account for 10 percent of the TPS regardless of the number of measures applied to an HHA in the other three classifications.
We propose to calculate the TPS for the HHVBP methodology similarly to the TPS calculation that has been finalized under the HVBP program. The performance scoring methodology for the HHVBP model would include determining performance standards (benchmarks and thresholds) using the 2015 baseline period performance year's quality measure data, scoring HHAs based on their achievement and/or improvement with respect to those performance standards, and weighting each of the classifications by the number of measures employed, as presented in further detail in Section G below.
For scoring HHAs' performance on measures in the proposed Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications, we propose that the HHVBP model would adopt an approach using several key elements from the scoring methodology set forth in the 2007 HVBP Report and the successfully implemented HVBP Program
In determining the achievement points for each measure, HHAs would receive points along an achievement range, which is a scale between the achievement threshold and a benchmark. We propose to calculate the achievement threshold as the median of all HHAs' performance on the specified quality measure during the baseline period and to calculate the benchmark as the mean of the top decile of all HHAs' performance on the specified quality measure during the baseline period. Unlike the HVBP Program that uses a national sample, this model would calculate both the achievement threshold and the benchmark separately for each selected state and for HHA cohort size. Under this proposed methodology, we would have benchmarks and achievement
We are proposing that achievement scoring under the HHVBP model would be based on the Performance Assessment Model set forth in the 2007 HVBP Report and as implemented under the HVBP Program. An HHA would earn 0–10 points for achievement for each measure in the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications based on where its performance during the performance period falls relative to the achievement threshold and the benchmark, according to the following formula:
All achievement points would be rounded up or down to the nearest point (for example, an achievement score of 4.555 would be rounded to 5). HHAs would receive an achievement score as follows:
• An HHA with performance equal to or higher than the benchmark would receive the maximum of 10 points for achievement.
• An HHA with performance equal to or greater than the achievement threshold (but below the benchmark) would receive 1–9 points for achievement, by applying the formula above.
• An HHA with performance less than the achievement threshold would receive 0 points for achievement.
We welcome public comment on this proposed methodology for scoring HHAs on achievement under the proposed HHVBP model.
In keeping with the approach used by the HVBP program, we propose that an HHA would earn 0–10 points based on how much its performance during the performance period improved from its performance on each measure in the proposed Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications during the baseline period. A unique improvement range for each measure would be established for each HHA that defines the difference between the HHA's baseline period score and the same state and size level benchmark for the measure used in the achievement scoring calculation described previously, according to the following formula:
• Equal to or higher than the benchmark score, the HHA would receive an improvement score of 10 points;
• Greater than its baseline period score but below the benchmark (within the improvement range), the HHA would receive an improvement score of 0–10, based on the formula above; or
• Equal to or lower than its baseline period score on the measure, the HHA would receive 0 points for improvement.
We welcome public comments on this proposed methodology for scoring HHAs on improvement under the proposed HHVBP model.
For illustrative purposes we present the following examples of how the proposed performance scoring methodology would be applied in the context of the proposed measures in the proposed Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications. These HHA examples were selected from an empirical database created from 2013/2014 data from the Home Health Compare archived data, claims data and enrollment data to support the development of the HHVBP permutation of the Performance Assessment Model, and all performance scores are calculated for the pneumonia measure, with respect to the number of individuals assessed and administered the pneumococcal vaccine.
Figure 7 shows the scoring for HHA `A', as an example. The benchmark calculated for the pneumonia measure in this case was 0.87 (the mean value of the top decile in 2013), and the achievement threshold was 0.47 (the performance of the median or the 50th percentile among HHAs in 2013). HHA A's 2014 performance rate of 0.91 during the performance period for this measure exceeds the benchmark, so HHA A would earn 10 (the maximum)
Figure 7 also shows the scoring for HHA `B'. As referenced below, HHA B's performance on this measure went from 0.21 (which was below the achievement threshold) in the baseline period to 0.70 (which is above the achievement threshold) in the performance period. Applying the achievement scale, HHA B would earn 6 points for achievement, calculated as follows: [9 * ((0.70 − 0.47)/(0.87 − 0.47))] + 0.5 = 5.675, and then rounded to 6 points.
Checking HHA B's improvement score yields the following result: Based on HHA B's period-to-period improvement, from 0.21 in the baseline year to 0.70 in the performance year, HHA B would earn 7 points, calculated as follows: [10 * ((0.70 − 0.21)/(0.87 − 0.21))] − 0.5 = 6.92, rounded to 7 points. Because the higher of the achievement and improvement scores is used, HHA B would receive 7 points for this measure.
In Figure 8, HHA `C' yielded a decline in performance on the pneumonia measure, falling from 0.57 to 0.46 (a decline of 0.11 points). HHA C's performance during the performance period is lower than the achievement threshold of 0.47 and, as a result, receives 0 points based on achievement. It also receives 0 points for improvement, because its performance during the performance period is lower than its performance during the baseline period.
The HHVBP model provides us with the opportunity to study new quality measures. The four New Measures that we have proposed to adopt for the model for PY1 would be reported directly by the HHA and would account for 10 percent of the TPS regardless of the number of measures in the other three classifications. We are proposing that HHAs that report on these measures would receive 10 points out of a maximum of 10 points for each of the 4 measures in the New Measure classification. Hence a HHA that reports on all four measures would receive 40 points out of a maximum of 40. An HHA would receive 0 points for each measure that it fails to report on. If an HHA reports on all four measures, it would receive 40 points for the classification and 10 points (40/40 * 10 points) would be added to its TPS because the New Measure classification has a maximum weight of 10 percent. If an HHA reports on 3 of 4 measures, it would receive 30 points of 40 points available for the classification and 7.5 points (30/40 * 10 points) added to its TPS. If an HHA reports on 2 of 4 measures, they would receive 20 points of 40 points available for the classification and 5.0 points (20/40 * 10 points) added to their TPS. If an HHA reports on 0 of 4 measures, they would receive 0 points and have no points added to their TPS. We intend to update these measures through future rulemaking to allow us to study newer, leading-edge measures as well as retire measures that no longer require such analysis. We request comment on this proposed scoring methodology for new measures.
While no HHA in a selected state would be exempt from the HHVBP model, there may be periods when an HHA does not receive a payment adjustment because there are not an adequate number of episodes of care to generate sufficient quality measure data. The minimum threshold for an HHA to receive a score on a given measure is 20 home health episodes of care per year for HHAs that have been certified for at least 6-months. If an HHA does not meet this threshold to generate scores on five or more of the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience measures, no payment adjustment will be made, and the Medicare-certified HHA would be paid for HHA services in an amount equivalent to the amount it would have been paid under section 1895 of the Act.
HHAs with very low volumes will either increase their volume in later performance years and be subject to future payment adjustment, or the HHAs' volume will remain very low and the HHAs would continue to not have their payment adjusted in future years. Based on the most recent data available at this time, a very small number of HHAs are reporting on less than five of the total number of measures included in the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications and account for less than 0.5 percent of the claims made over 1,900 HHAs delivering care within the nine proposed selected states. We expect very little impact of very low service volume HHAs on the model due to the low number of low volume HHAs and because it is unlikely that a HHA will reduce the amount of service to such a low level to avoid a payment adjustment. Although these HHAs would not be subject to payment adjustments, they would remain in the model and have access to the same technical assistance as all other HHAs in the model, and would receive quality reports on any measures for which they do have 20 episodes of care, and a future opportunity to compete for payment adjustments.
We propose the HHA's TPS would be based on all the Clinical Quality of Care, Outcome and Efficiency, Person and Caregiver-Centered Experience measures and the New Measures that apply to the HHA. As described above, each measure in the Clinical Quality of Care, Outcome and Efficiency and Person and Caregiver-Centered
As an example, HHA “A” has at least 20 episodes of care in a 12-month period for only 9 quality measures out of a possible 25 measures from three of the four classifications (except the New Measures). Under the proposed scoring methodology outlined above, HHA A would be awarded 0, 0, 3, 4, 5, 7, 7, 9, and 10 points, respectively, for these measures. HHA A's total earned points for the three classifications would be calculated by adding together all the points awarded to HHA A, resulting in a total of 45 points. HHA A's total possible points would be calculated by multiplying the total number of measures for which the HHA reported on least 20 episodes (nine) by the maximum number of points for those measures (10), yielding a total of 90 possible points. HHA A's score for the three classifications would be the total earned points (45) divided by the total possible points (90) multiplied by 90 because as mentioned in section E7, the Clinical Quality of Care, Outcome and Efficiency, and Person and Caregiver-Centered Experience classifications account for 90 percent of the TPS and the New Measures classification accounts for 10 percent of the TPS, which yields a result of 45. In this example, HHAs also reported all four numbers and would receive the full 10 points for the new measure. As a result, the TPS for HHA A would be 55 (45 plus 10). In addition, as specified in Section E:7—Weighting, all measures have equal weights regardless of their classification (except for New Measures) and the total earned points for the three classifications can be calculated by adding the points awarded for each such measure together. We seek public comment on our proposal of the minimum number of cases for outcome and clinical quality measures.
We propose to codify at 42 CFR 484.330 a methodology for applying value-based payment adjustments to home health services under the HHVBP model. Payment adjustments would be made to the HH PPS final claim payment amount as calculated in accordance with § 484.205 using a linear exchange function (LEF) similar to the methodology utilized by the HVBP Program. The LEF is used to translate an HHA's TPS into a percentage of the value-based payment adjustment earned by each HHA under the HHVBP model. The LEF was identified by the HVBP Program as the simplest and most straightforward option to provide the same marginal incentives to all hospitals, and we believe the same to be true for HHAs. We propose the function's intercept at zero percent, meaning those HHAs that have a TPS that is average in relationship to other HHAs in their cohort (a zero percent), would not receive any payment adjustment. Payment adjustments for each HHA with a score above zero percent would be determined by the slope of the LEF. In addition we propose to set the slope of the LEF for the first performance year, CY 2016, so that the estimated aggregate value-based payment adjustments for CY 2016 are equal to 5 percent of the estimated aggregate base operating episode payment amount for CY 2018. The estimated aggregate base operating episode payment amount is the total amount of episode payments made to all the HHAs by Medicare in each individual state in the larger- and smaller-volume cohorts respectively (we are proposing nine states, which would create 18 separate aggregate base operating episode payment amounts).
Figure 9 provides an example of how the LEF is calculated and how it is applied to calculate the percentage payment adjustment to a HHA's TPS. For this example, we applied the 8 percent payment adjustment level that is proposed for the final two years of the HHVBP model. The proposed rate for the payment adjustments for other years would be proportionally less.
Step #1 involves the calculation of the `Prior Year Aggregate HHA Payment Amount' (See C2 in Figure 9) that each HHA was paid in the prior year. From claims data, all payments are summed together for each HHA for CY 2015, the year prior to the HHVBP Model.
Step #2 involves the calculation of the `8 percent Payment Reduction Amount' (C3 of Figure 9) for each HHA. The `Prior Year Aggregate HHA Payment Amount' is multiplied by the `8 percent Payment Reduction Rate'. The aggregate of the `8-percent Payment Reduction Amount' is the numerator of the LEF.
Step #3 involves the calculation of the `Final TPS Adjusted Reduction Amount' (C4 of Figure 9) by multiplying the `8-percent Payment Reduction Amount' from Step #2 by the TPS (C1) divided by 100. The aggregate of the `TPS Adjusted Reduction Amount' is the denominator of the LEF.
Step #4 involves calculating the LEF (C5 of Figure 9) by dividing the aggregate `8 percent Payment Reduction Amount' by the aggregate `TPS Adjusted Reduction Amount'.
Step #5 involves the calculation of the `Final TPS Adjusted Payment Amount' (C6 of Figure 9) by multiplying the `TPS Adjusted Reduction Amount' (C4) by the LEF (C5). This is an intermediary value used to calculate `Quality Adjusted Payment Rate'.
Step #6 involves the calculation of the `Quality Adjusted Payment Rate' (C7 of Figure 9) that the HHA would receive instead of the 8 percent reduction in payment. This is an intermediary step to determining the payment adjustment rate. For CYs 2021 and 2022, the payment adjustment in this column would range from 0 percent to 16 percent depending on the quality of care provided.
Step #7 involves the calculation of the `Final Percent Payment Adjustment' (C8 of Figure 9) that would be applied to the HHA payments after the performance period. It simply involves the CY payment adjustment percent (in 2018, 5 percent; in 2019, 5 percent; in 2020, 6 percent; in 2021, 8 percent; and in 2022, 8 percent). In this example, we use the maximum eight-percent (8 percent) subtraction to the `Quality Adjusted Payment Rate'. Note that the payment adjustment percentage is capped at no more than plus or minus 8 percent for each respective performance period and the payment adjustment would occur on the final claim payment amount.
We invite public comments on this proposed payment adjustment methodology.
We are proposing to provide HHAs two separate opportunities to review scoring information under the HHVBP model. First, HHAs will have the opportunity to review their quarterly quality reports following each quarterly posting; second, Medicare-certified HHAs will have the opportunity to review their TPS and payment adjustment calculations, and request a recalculation if a discrepancy is identified due to a CMS error as described in this section. These processes would also help educate and inform each competing Medicare-certified HHA on the direct relation between the payment adjustment and performance measure scores.
The proposed model design calls for us to inform HHA quarterly of their performance on each of the individual quality measures used to calculate the TPS. We propose that HHAs will have 10 days after the quarterly reports are provided to request a recalculation of a measure scores if it believes there is evidence of a discrepancy. We would adjust the score if it is determined that the discrepancy in the calculated measure scores was the result of our failure to follow measurement calculation protocols.
In addition, the proposed model design also calls for us to inform each Medicare-certified HHA of the TPS and payment adjustment amount in an annual report. We propose that these annual reports be provided to Medicare-certified HHAs each August prior to the calendar year for which the payment adjustment would be applied. Similar to quarterly reports, HHAs will have 10 days to request a recalculation of their TPS and payment adjustment amount from the date information is made available. For both the quarterly reports and the annual report containing the TPS and payment adjustments, Medicare-certified HHAs will only be permitted to request scoring recalculations, and must include a specific basis for the requested recalculation. We will not be responsible for providing HHAs with the underlying source data utilized to generate performance measure scores. Each HHA has access to this data via the QIES system. The final TPS and payment adjustment would then be provided to competing Medicare-certified HHAs in a final report no later than 60 days in advance of the payment adjustment taking effect.
The TPS from the annual performance report would be calculated based on the calculation of performance measures contained in the quarterly reports that have already been provided and reviewed by the HHAs. As a result, we believe that quarterly reviews would provide substantial opportunity to identify and correct errors and resolve discrepancies, thereby minimizing the challenges to the annual performance scores linked to payment adjustment.
As described above, a quarterly performance report would be provided to all Medicare-certified HHAs within the selected states beginning with the first quarter of CY 2016 being reported in July 2016. We propose that HHAs would submit recalculation requests for both quarterly quality performance measure reports and for the TPS and payment adjustment reports via an email link provided on the model-specific Web page. The request form would be entered by a person who has authority to sign on behalf of the HHA and be submitted within 10 days of receiving the quarterly data report or the annual TPS and payment adjustment report.
Requests for both quarterly report measure score recalculations or TPS and payment adjustment recalculations would contain the following information:
• The provider's name, address associated with the services delivered, and CMS Certification Number (CCN);
• The basis for requesting recalculation to include the specific quality measure data that the HHA believes is inaccurate or the calculation the HHA believes is incorrect;
• Contact information for a person at the HHA with whom CMS or its agent can communicate about this request, including name, email address, telephone number, and mailing address (must include physical address, not just a post office box); and,
• A copy of any supporting documentation the HHA wishes to submit in electronic form via the model-specific Web page.
Following receipt of a request for quarterly report measure score recalculations or a request for TPS and payment adjustment recalculation, CMS or its agent would:
+ Provide an email acknowledgement, using the contact information provided in the recalculation request, to the HHA contact notifying the HHA that the request has been received;
+ Review the request to determine validity, and determine whether the requested recalculation would result in a score change altering performance measure scores or the HHA's TPS;
+ If recalculation would result in a performance measure score or TPS change, conduct a review of quality data and if an error is found, recalculate the TPS using the corrected performance data; and,
+ Provide a formal response to the HHA contact, using the contact information provided in the recalculation request, notifying the HHA of the outcome of the review and recalculation process.
Recalculation and subsequent communication of the results of these determinations would occur as soon as administratively feasible following the submission of requests. We request comment on our proposed quarterly quality report measure review, TPS preview period, and our proposed process for requesting recalculation of the quarterly performance measure scores, and the TPS and payment adjustment. We intend to codify these processes in regulation text in future rulemaking.
Additionally, we will develop and adopt an appeals mechanism under the model through future rulemaking in advance of the application of any payment adjustments.
We propose to codify at 484.315(c) that HHAs in selected states would be required to collect and report information to CMS necessary for the purposes of monitoring and evaluating this model as required by statute.
We intend to use a multilevel approach to evaluation. Here, we intend to conduct analyses at the state, HHA, and patient levels. Based on the state groupings discussed in the section on selection of Medicare certified HHAs, we believe there are several ways in which we can draw comparison groups and remain open to scientifically-sound, rigorous methods for evaluating the effect of the model intervention.
The evaluation effort may require of HHAs participating in the Model additional data specifically for evaluation purposes. Such requirements for additional data to carry out model evaluation would be in compliance with 42 CFR 403.1105 which, as of January 1, 2015, requires entities participating in the testing of a model under section 1115A to collect and report such information, including protected health information (as defined at 45 CFR 160.103), as the Secretary determines is necessary to monitor and evaluate the model. We would consider all Medicare-certified HHAs providing services within a state selected for the Model to be participating in the testing of this model because the competing HHAs would be receiving payment from CMS under the model.
We invite public comments on this proposed evaluation plan.
Section 1895(b)(3)(B)(v)(II) of the Act requires that for 2007 and subsequent years, each HHA submit to the Secretary in a form and manner, and at a time, specified by the Secretary, such data that the Secretary determines are appropriate for the measurement of health care quality. To the extent that an HHA does not submit data in accordance with this clause, the Secretary is directed to reduce the home health market basket percentage increase applicable to the HHA for such year by 2 percentage points. As provided at section 1895(b)(3)(B)(vi) of the Act, depending on the market basket percentage for a particular year, the 2 percentage point reduction under section 1895(b)(3)(B)(v)(I) of the Act may result in this percentage increase, after application of the productivity adjustment under section 1895(b)(3)(B)(vi)(I) of the Act, being less than 0.0 percent for a year, and may result in payment rates under the Home Health PPS for a year being less than payment rates for the preceding year.
Section 2(a) of the Improving Medicare Post-Acute Care Transformation Act of 2014 (the IMPACT Act) (Pub. L. 113–185, enacted on Oct. 6, 2014) amended Title XVIII of the Act, in part, by adding a new section 1899B, which imposes new data reporting requirements for certain post-acute care (PAC) providers, including HHAs. New section 1899B of the Act is titled, “Standardized Post-Acute Care (PAC) Assessment Data for Quality, Payment, and Discharge Planning”. Under section 1899B(a)(1) of the Act, certain post-acute care (PAC) providers (defined in section 1899B(a)(2)(A) of the Act to include HHAs, SNFs, IRFs, and LTCHs) must submit standardized patient assessment data in accordance with section 1899B(b) of the Act, data on quality measures required under section 1899B(c)(1) of the Act, and data on resource use, and other measures required under section 1899B(d)(1) of the Act. The Act also sets out specified application dates for each of the measures. The Secretary must specify the quality, resource use, and other measures no later than the applicable specified application date defined in section 1899B(a)(2)(E) of the Act.
Section 1899B(b) of the Act describes the standardized patient assessment data that PAC providers are required to submit in accordance with section 1899B(b)(1) of the Act; requires the Secretary, to the extent practicable, to match claims data with standardized patient assessment data in accordance with section 1899B(b)(2) of the Act; and requires the Secretary, as soon as practicable, to revise or replace existing patient assessment data to the extent that such data duplicate or overlap with standardized patient assessment data, in accordance with section 1899B(b)(3) of the Act.
Sections 1899B(c)(1) and (d)(1) of the Act direct the Secretary to specify measures that relate to at least five stated quality domains and three stated resource use and other measure domains. Section 1899B(c)(1) of the Act provides that the quality measures on which PAC providers, including HHAs, are required to submit standardized patient assessment data and other necessary data specified by the Secretary must be in accordance with, at least, the following domains:
• Functional status, cognitive function, and changes in function and cognitive function;
• Skin integrity and changes in skin integrity;
• Medication reconciliation;
• Incidence of major falls; and
• Accurately communicating the existence of and providing for the transfer of health information and care preferences of an individual to the individual, family caregiver of the individual, and providers of services furnishing items and services to the individual when the individual transitions (1) from a hospital or Critical Access Hospital (CAH) to another applicable setting, including a PAC provider or the home of the individual, or (2) from a PAC provider to another applicable setting, including a different PAC provider, hospital, CAH, or the home of the individual.
Section 1899B(c)(2)(A) provides that, to the extent possible, the Secretary must require such reporting through the use of a PAC assessment instrument and modify the instrument as necessary to enable such use.
Section 1899B(d)(1) of the Act provides that the resource use and other measures on which PAC providers, including HHAs, are required to submit any necessary data specified by the Secretary, which may include standardized assessment data in addition to claims data, must be in accordance with, at least, the following domains:
• Resource use measures, including total estimated Medicare spending per beneficiary;
• Discharge to community; and
• Measures to reflect all-condition risk-adjusted potentially preventable hospital readmission rates.
Sections 1899B(c) and (d) of the Act indicate that data satisfying the eight measure domains in the IMPACT Act is the minimum data reporting requirement. Therefore, the Secretary may specify additional measures and additional domains.
Section 1899B(e)(1) of the Act requires that the Secretary implement the quality, resource use, and other measures required under sections 1899B(c)(1) and (d)(1) of the Act in phases consisting of measure specification, data collection, and data analysis; the provision of feedback reports to PAC providers in accordance with section 1899B(f) of the Act; and public reporting of PAC providers' performance on such measures in accordance with section 1899B(g) of the Act. Section 1899B(e)(2) of the Act generally requires that each measure specified by the Secretary under section 1899B of the Act be NQF-endorsed, but authorizes an exception under which the Secretary may select non-NQF-endorsed quality measures in the case of specified areas or medical topics determined appropriate by the Secretary for which a feasible or practical measure has not been endorsed by the NQF, as long as due consideration is given to measures that have been endorsed or adopted by a consensus organization identified by the Secretary. Section 1899B(e)(3) of the Act provides that the pre-rulemaking process required by section 1890A of the Act applies to quality, resource use, and other measures specified under sections 1899B(c)(1) and (d)(1) of the Act, but authorizes exceptions under which the Secretary may (1) use expedited procedures, such as ad hoc reviews, as necessary in the case of a measure required with respect to data submissions during the 1-year period before the applicable specified application date, or (2) alternatively, waive section 1890A of the Act in the case of such a measure if applying section 1890A of the Act (including through the use of expedited procedures) would result in the inability of the Secretary to satisfy any deadline specified under section 1899B of the Act with respect to the measure.
Section 1899B(f)(1) of the Act requires the Secretary to provide confidential feedback reports to PAC providers on the performance of such PAC providers with respect to quality, resource use, and other measures required under sections 1899B(c)(1) and (d)(1) of the Act beginning 1 year after the applicable specified application date.
Section 1899B(g) of the Act requires the Secretary to establish procedures for making available to the public information regarding the performance of individual PAC providers with respect to quality, resource use, and other measures required under sections 1899B(c)(1) and (d)(1) beginning not later than 2 years after the applicable specified application date. The procedures must ensure, including through a process consistent with the process applied under section 1886(b)(3)(B)(viii)(VII) for similar purposes, that each PAC provider has the opportunity to review and submit corrections to the data and information that are to be made public with respect to the PAC provider prior to such data being made public.
Section 1899B(h) of the Act sets out requirements for removing, suspending, or adding quality, resource use, and other measures required under sections 1899B(c)(1) and (d)(1) of the Act. In addition, section 1899B(j) of the Act requires the Secretary to allow for stakeholder input, such as through town halls, open door forums, and mailbox submissions, before the initial rulemaking process to implement section 1899B of the Act.
Section 2(c)(1) of the IMPACT Act amended section 1895 of the Act to address the payment consequences for HHAs with respect to the additional data which HHAs are required to submit under section 1899B of the Act. These changes include the addition of a new section 1895(3)(B)(v)(IV), which requires HHAs to submit the following additional data: (1) For the year beginning on the applicable specified application date and subsequent years, data on the quality, resource use, and other measures required under sections 1899B(c)(1) and (d)(1) of the Act; and (2) for 2019 and subsequent years, the standardized patient assessment data required under section 1899B(b)(1) of the Act. Such data must be submitted in the form and manner, and at the time, specified by the Secretary.
As stated above, the IMPACT Act adds a new section 1899B that imposes new data reporting requirements for certain post-acute care (PAC) providers, including HHAs. Sections 1899B(c)(1) and 1899B(d)(1) collectively require that the Secretary specify quality measures and resource use and other measures with respect to certain domains not later than the specified application date that applies to each measure domain and PAC provider setting. Section 1899B(a)(2)(E) delineates the specified application dates for each measure domain and PAC provider. The IMPACT Act also amends other sections of the Act, including section 1895(b)(3)(B)(v), to require the Secretary to reduce the otherwise applicable PPS payment to a PAC provider that does not report the new data in a form and manner, and at a time, specified by the Secretary. For HHAs, amended section 1895(b)(3)(B)(v) would require the Secretary to reduce the payment update for any HHA that does not satisfactorily submit the new required data.
Under the current HH QRP, the general timeline and sequencing of measure implementation occurs as follows: Specification of measures; proposal and finalization of measures through notice-and-comment rulemaking; HHA submission of data on the adopted measures; analysis and processing of the submitted data; notification to HHAs regarding their quality reporting compliance with respect to a particular year; consideration of any reconsideration requests; and imposition of a payment reduction in a particular year for failure to satisfactorily submit data with respect to that year. Any payment reductions that are taken with respect to a year begin approximately 1 year after the end of the data submission period for that
To the extent that the IMPACT Act could be interpreted to shorten this timeline, so as to require us to reduce HH PPS payment for failure to satisfactorily submit data on a measure specified under section 1899B(c)(1) or (d)(1) of the IMPACT Act beginning with the same year as the specified application date for that measure, such a timeline would not be feasible. The current timeline discussed above reflects operational and other practical constraints, including the time needed to specify and adopt valid and reliable measures, collect the data, and determine whether a HHA has complied with our quality reporting requirements. It also takes into consideration our desire to give HHAs enough notice of new data reporting obligations so that they are prepared to timely start reporting data. Therefore, we intend to follow the same timing and sequence of events for measures specified under sections 1899B(c)(1) and (d)(1) of the Act that we currently follow for other measures specified under the HH QRP. We intend to specify each of these measures no later than the specified application dates set forth in section 1899B(a)(2)(E) of the Act and propose to adopt them consistent with the requirements in the Act and Administrative Procedure Act. To the extent that we finalize a proposal to adopt a measure for the HH QRP that satisfies an IMPACT Act measure domain, we intend to require HHAs to report data on the measure for the year that begins 2 years after the specified application date for that measure. Likewise, we intend to require HHAs to begin reporting any other data specifically required under the IMPACT Act for the year that begins 2 years after we adopt requirements that would govern the submission of that data.
Lastly, on April 1, 2014, the Congress passed the Protecting Access to Medicare Act of 2014 (PAMA) (Pub. L. 113–93), which stated the Secretary may not adopt ICD–10 prior to October 1, 2015. On August 4, 2014, HHS published a final rule titled “Administrative Simplification: Change to the Compliance Date for the International Classification of Diseases, 10th Revision (ICD–10–CM and ICD–10–PCS Medical Data Code Sets” (79 FR 45128), which announced October 1, 2015 as the new compliance date. The OASIS–C1 data item set had been previously approved by the Office of Management and Budget (OMB) on February 6, 2014 and scheduled for implementation on October 1, 2014. We intended to use the OASIS–C1 to coincide with the original implementation date of the ICD–10. The approved OASIS–C1 included changes to accommodate coding of diagnoses using the ICD–10–CM coding set and other important stakeholder concerns such as updating clinical concepts, and revised item wording and response categories to improve item clarity. This version included five (5) data items that required the use of ICD–10 codes.
Since OASIS–C1 was revised to incorporate ICD–10 coding, it is not feasible to implement the OASIS–C1/ICD–10 version prior to October 1, 2015, when ICD–10 is scheduled to be implemented. Due to this delay, we had to ensure the collection and submission of OASIS data continued, until ICD–10 could be implemented. Therefore, we have made interim changes to the OASIS–C1 data item set to allow use with ICD–9 until ICD–10 is adopted. The OASIS–C1/ICD–9 version was submitted to OMB for approval until the OASIS–C1/ICD–10 version could be implemented. A 6-month emergency approval was granted on October 7, 2014 and CMS subsequently applied for an extension. The extension of the OASIS–C1/ICD–9 version was reapproved under OMB control number 0938–0760 with a current expiration date of March 31, 2018. It is important to note, that this version of the OASIS will be discontinued once the OASIS–C1/ICD–10 version is approved and implemented. In addition, to facilitate the reporting of OASIS data as it relates to the planned implementation of ICD–10 on October 1, 2015, we submitted a new request for approval to OMB for the OASIS–C1/ICD–10 version under the Paperwork Reduction Act (PRA) process. We are requesting a new OMB control number for the proposed revised OASIS item as announced in the 30-day
We strive to promote high quality and efficiency in the delivery of health care to the beneficiaries we serve. Performance improvement leading to the highest quality health care requires continuous evaluation to identify and address performance gaps and reduce the unintended consequences that may arise in treating a large, vulnerable, and aging population. Quality reporting programs, coupled with public reporting of quality information, are critical to the advancement of health care quality improvement efforts.
We seek to adopt measures for the HH QRP that promotes better, safer, and more efficient care. Valid, reliable, relevant quality measures are fundamental to the effectiveness of our quality reporting programs. Therefore, selection of quality measures is a priority for CMS in all of its quality reporting programs.
The measures selected would address the measure domains as specified in the IMPACT Act and would be in alignment with the
• Better Care: Improve the overall quality of care by making healthcare more patient-centered, reliable, accessible, and safe.
• Healthy People, Healthy Communities: Improve the health of the U.S. population by supporting proven interventions to address behavioral, social, and environmental determinants of health in addition to delivering higher-quality care.
• Affordable Care: Reduce the cost of quality healthcare for individuals, families, employers, and government.
In addition, our measure selection activities for the HH QRP take into consideration input we receive from the Measure Applications Partnership (MAP), convened by the NQF, as part of the established CMS pre-rulemaking process required under section 1890A of the Act. The MAP is a public-private partnership comprised of multi-stakeholder groups convened for the primary purpose of providing input to us on the selection of certain categories of quality and efficiency measures, as required by section 1890A(a)(3) of the Social Security Act (the Act). By February 1st of each year, the NQF must provide that input to us. Input from the MAP is located at
We initiated an Ad Hoc MAP process for the review of the measures under consideration for implementation in
To meet the first specified application date applicable to HHAs under section 1899B(a)(2)(E) of the Act, which is October 1, 2017, we have focused on measures that:
• Correspond to a measure domain in sections 1899B(c)(1) or (d)(1) of the Act and are setting-agnostic: For example falls with major injury and the incidence of pressure ulcers;
• Are currently adopted for 1 or more of our PAC quality reporting programs, are already either NQF-endorsed and in use or finalized for use, or already previewed by the Measure Applications Partnership (MAP) with support;
• Minimize added burden on HHAs;
• Minimize or avoid, to the extent feasible, revisions to the existing items in assessment tools currently in use (for example, the OASIS); and
• Where possible, the avoidance duplication of existing assessment items.
In our selection and specification of measures, we employ a transparent process in which we seek input from stakeholders and national experts and engage in a process that allows for pre-rulemaking input on each measure, as required by section 1890A of the Act. This process is based on a private public partnership, and it occurs via the MAP. The MAP is composed of multistakeholder groups convened by the NQF, our current contractor under section 1890 of the Act, to provide input on the selection of quality and efficiency measures described in section 1890(b)(7)(B). The NQF must convene these stakeholders and provide us with the stakeholders' input on the selection of such measures. We, in turn, must take this input into consideration in selecting such measures. In addition, the Secretary must make available to the public by December 1 of each year a list of such measures that the Secretary is considering under Title XVIII of the Act. As discussed in section V.A. of this proposed rule 1899B(e)(3) provides that the pre-rulemaking process required by section 1890A of the Act applies to the measures required under section 1899B, subject to certain exceptions for expedited procedures or, alternatively, waiver of section 1890A. We initiated an ad hoc MAP process for the review of the quality measures under consideration for proposal, in preparation for adoption of those quality measures into the HH QRP that are required by the IMPACT Act, and that must be implemented by January 1, 2017. The List of Measures under Consideration (MUC List) under the IMPACT Act was made public on February 5, 2015. Under the IMPACT Act, these measures must be standardized so they can be applied across PAC settings and must correspond to measure domains specified in sections 1899B(c)(1) and (d)(1) of the IMPACT Act. The MAP reviewed each IMPACT Act-related quality measure proposed in this proposed rule for the HH QRP, in light of its intended cross-setting use. We refer to sections V.A. and V.C. of this proposed rule for more information on the MAP's recommendations. The MAP's final report, MAP Off-Cycle Deliberations 2015: Measures under Consideration to Implement Provisions of the IMPACT Act: Final Report, is available at
In the absence of NQF endorsement on measures for the home health setting, or measures that are not fully supported by the MAP for the HH QRP, we intend to propose for adoption measures that most closely align with the national priorities discussed above and for which the MAP supports the measure concept. Further discussion as to the importance and high-priority status of these measures in the HH setting is included under each quality measure proposal in this proposed rule. In addition, for measures not endorsed by the NQF, we have sought, to the extent practicable, to adopt measures that have been endorsed or adopted by a national consensus organization, recommended by multi-stakeholder organizations, and/or developed with the input of providers, purchasers/payers, and other stakeholders.
In the CY 2014 HH PPS final rule, (78 FR 72256–72320), we finalized a proposal to add two claims-based measures to the HH QRP, and stated that we would begin reporting the data from these measures to HHAs beginning in CY 2014. These claims based measures are: (1) Rehospitalization during the first 30 days of HH; and (2) Emergency Department Use without Hospital Readmission during the first 30 days of HH. In an effort to align with other updates to Home Health Compare, including the transition to quarterly provider preview reports, we have made the decision to delay the reporting of data from these measures until July 2015 (
(a) We are proposing one standardized cross-setting new measure for CY 2016 to meet the requirements of the IMPACT Act. The proposed quality measure that addresses the domain of skin integrity and changes in skin integrity is the National Quality Forum (NQF)-endorsed measure: Percent of Residents or Patients with Pressure Ulcers That Are New or Worsened (Short Stay) (NQF #0678) (
The IMPACT Act requires the specification of a quality measure to address skin integrity and changes in skin integrity in the home health setting by January 1, 2017. We are proposing the implementation of the quality measure NQF #0678, Percent of Residents or Patients with Pressure Ulcers that are New or Worsened (Short Stay) in the HH QRP as a cross-setting quality measure to meet the requirements of the IMPACT Act for the CY 2018 payment determination and subsequent years. This measure reports the percent of patients with Stage 2 through 4 pressure ulcers that are new or worsened since the beginning of the episode of care.
Pressure ulcers are high-volume in post-acute care settings and high-cost adverse events. According to the 2014 Prevention and Treatment Guidelines published by the National Pressure Ulcer Advisory Panel, European Pressure Ulcer Advisory Panel, and Pan Pacific Pressure Injury Alliance, pressure ulcer care is estimated to cost approximately $11 billion annually, and between $500 and $70,000 per individual pressure ulcer.
The IMPACT Act requires the specification of quality measures that are harmonized across PAC settings. This requirement is consistent with the NQF Steering Committee report, which stated that to understand the impact of pressure ulcers across settings, quality measures addressing prevention, incidence, and prevalence of pressure ulcers must be harmonized and aligned.
A TEP convened by our measure development contractor provided input on the technical specifications of this quality measure, including the feasibility of implementing the measure across PAC settings. The TEP was supportive of the implementation of this measure across PAC settings and applauded CMS's efforts to standardize this measure for cross-setting development. Additionally, the NQF MAP met on February 9, 2015 and
We propose that data for the standardized quality measure would be collected using the OASIS–C1 with submission through the Quality Improvement and Evaluation System (QIES) Assessment Submission and Processing (ASAP) system. HHAs began submitting data in January 2015 for the OASIS items used to calculate NQF #0678, the Percent of Residents, or Patients with Pressure Ulcers That Are New or Worsened (Short Stay), as part of the Home Health Quality Initiative to assess the number of new or worsened pressure ulcers in January 2015. By building on the existing reporting and submission infrastructure for HHAs, we intend to minimize the administrative burden related to data collection and submission for this measure under the HH QRP. For more information on HH reporting using the QIES ASAP system, refer to:
Data collected through the OASIS–C1 would be used to calculate this quality measure. Data items in the OASIS–C1 include M1308 (Current Number of Unhealed Pressure Ulcers at Each Stage or Unstageable) and M1309 (Worsening in Pressure Ulcer Status Since SOC/ROC). Data collected through the OASIS–C1 would be used for risk adjustment of this measure. We anticipate risk adjustment items would include, but is not limited to M1850 (Activities of Daily Living Assistance, Transferring), and M1620 (Bowel Incontinence Frequency). OASIS C1 items M1016 (Diagnoses Requiring Medical or Treatment Change Within past 14 Days), M1020 (Primary Diagnoses) and M1022 (Other Diagnoses) would be used to identify patients with a diagnosis of peripheral vascular disease, diabetes, or malnutrition. More information about the OASIS items is available in the OASIS Manual
The calculation of the proposed measure would be based on the items M1308 (Current Number of Unhealed Pressure Ulcers at Each Stage or Unstageable) and M1309 (Worsening in Pressure Ulcer Status Since SOC/ROC). The specifications and data items for NQF #0678, the Percent of Residents or Patients with Pressure Ulcers that are New or Worsened (Short Stay), are available at
We invite public comment on our proposal to adopt NQF #0678 Percent of Residents or Patients with Pressure Ulcers that are New or Worsened (Short Stay) for the HH QRP to fulfill the timeline requirements for implementation under the IMPACT Act, for CY2018 HH payment determination and subsequent years.
As part of our ongoing measure development efforts, we are considering a future update to the numerator of the quality measure NQF #0678, Percent of Residents or Patients with Pressure Ulcers that are New or Worsened (Short Stay). This update would hold providers accountable for the development of unstageable pressure ulcers and suspected deep tissue injuries (sDTIs). Under this proposed change the numerator of the quality measure would be updated to include unstageable pressure ulcers, including sDTIs that are new/developed while the patient is receiving home health care, as well as Stage 1 or 2 pressure ulcers that become unstageable due to slough or eschar (indicating progression to a full thickness [that is, stage 3 or 4] pressure ulcer) after admission. This would be consistent with the specifications of the “New and Worsened Pressure Ulcer” measure for HH patients presented to the MAP on the 2014 MUC list. At this time, we are not proposing the implementation of this change (that is, including sDTIs and unstageable pressure ulcers in the numerator) in the HH QRP, but are soliciting public feedback on this potential area of measure development.
Our measure development contractor convened a cross-setting pressure ulcer TEP that strongly recommended that CMS hold providers accountable for the development of new unstageable pressure ulcers and sDTIs by including these pressure ulcers in the numerator of the quality measure. Although the TEP acknowledged that unstageable pressure ulcers and sDTIs cannot and should not be assigned a numeric stage, panel members recommended that these be included in the numerator of NQF #0678, the Percent of Residents, or Patients with Pressure Ulcers That Are New or Worsened (Short Stay), as a new pressure ulcer if developed during a home health episode. The TEP also recommended that a Stage 1 or 2 pressure ulcer that becomes unstageable due to slough or eschar should be considered worsened because the presence of slough or eschar indicates a full thickness (equivalent to Stage 3 or 4) wound.
In addition, we are also considering whether body mass index (BMI) should be used as a covariate for risk-adjusting NQF #0678 in the home health setting, as is done in other post-acute care settings. We invite public feedback to inform our direction to include unstageable pressure ulcers and sDTIs in the numerator of the quality measure NQF #0678 Percent of Residents or Patients with Pressure Ulcers that are New or Worsened (Short Stay), as well as on the possible collection of height
(b) We have also identified four future, cross-setting measure constructs to potentially meet requirements of the IMPACT Act domains of: (1) All-condition risk-adjusted potentially preventable hospital readmission rates; (2) resource use, including total estimated Medicare spending per beneficiary; (3) discharge to community; and (4) medication reconciliation. These are shown in Table 22; we would like to solicit public feedback to inform future measure development of these constructs as it relates to meeting the IMPACT Act requirements in these areas.
(c) We are working with our measure development and maintenance contractor to identify setting-specific measure concepts for future implementation in the HH QRP that align with or complement current measures and new measures to meet domains specified in the IMPACT Act. In identifying priority areas for future measure enhancement and development, we take into consideration results of environmental scans and resulting gaps analysis for relevant home health quality measure constructs, along with input from numerous stakeholders, including the Measures Application Partnership (MAP), the Medicare Payment Advisory Commission (MedPAC), Technical Expert Panels, and national priorities, such as those established by the National Priorities Partnership, the HHS Strategic Plan, the National Strategy for Quality Improvement in Healthcare, and the CMS Quality Strategy. Based on input from stakeholders, CMS has identified several high priority concept areas for future measure development in Table 23.
These measure concepts are under development, and details regarding measure definitions, data sources, data collection approaches, and timeline for implementation would be communicated in future rulemaking. We invite feedback about these seven high priority concept areas for future measure development.
The HH conditions of participation (CoPs) at § 484.55(d) require that the
It is important to note that to calculate quality measures from OASIS data, there must be a complete quality episode, which requires both a Start of Care (initial assessment) or Resumption of Care OASIS assessment and a Transfer or Discharge OASIS assessment. Failure to submit sufficient OASIS assessments to allow calculation of quality measures, including transfer and discharge assessments, is a failure to comply with the CoPs.
HHAs do not need to submit OASIS data for those patients who are excluded from the OASIS submission requirements. As described in the December 23, 2005 Medicare and Medicaid Programs: Reporting Outcome and Assessment Information Set Data as Part of the Conditions of Participation for Home Health Agencies final rule (70 FR 76202), we defined the exclusion as those patients:
• Receiving only non-skilled services;
• For whom neither Medicare nor Medicaid is paying for HH care (patient receiving care under a Medicare or Medicaid Managed Care Plan are not excluded from the OASIS reporting requirement);
• Receiving pre- or post-partum services; or
• Under the age of 18 years.
As set forth in the CY 2008 HH PPS final rule (72 FR 49863), HHAs that become Medicare certified on or after May 31 of the preceding year are not subject to the OASIS quality reporting requirement nor any payment penalty for quality reporting purposes for the following year. For example, HHAs certified on or after May 31, 2014 are not subject to the 2 percentage point reduction to their market basket update for CY 2015. These exclusions only affect quality reporting requirements and do not affect the HHAs' reporting responsibilities as announced in the December 23, 2005 final rule, Medicare and Medicaid Programs; Reporting Outcome and Assessment Information Set Data as Part of the Conditions of Participation for Home Health Agencies (70 FR 76202).
In the CY 2014 HH PPS Final rule (78 FR 72297), we finalized a proposal to consider OASIS assessments submitted by HHAs to CMS in compliance with HH CoPs and Conditions for Payment for episodes beginning on or after July 1, 2012, and before July 1, 2013 as fulfilling one portion of the quality reporting requirement for CY 2014.
In addition, we finalized a proposal to continue this pattern for each subsequent year beyond CY 2014. OASIS assessments submitted for episodes beginning on July 1st of the calendar year 2 years prior to the calendar year of the Annual Payment Update (APU) effective date and ending June 30th of the calendar year one year prior to the calendar year of the APU effective date, fulfill the OASIS portion of the HH QRP requirement.
Section 1895(b)(3)(B)(v)(I) of the Act states that for 2007 and each subsequent year, the home health market basket percentage increase applicable under such clause for such year shall be reduced by 2 percentage points if a home health agency does not submit data to the Secretary in accordance with subclause (II) with respect to such a year. This pay-for-reporting requirement was implemented on January 1, 2007. In the CY 2015 HH PPS Final rule (79 FR 38387), we finalized a proposal to define the quantity of OASIS assessments each HHA must submit to meet the pay-for-reporting requirement.
We believe that defining a more explicit performance requirement for the submission of OASIS data by HHAs would better meet section 5201(c)(2) of the Deficit Reduction Act of 2005 (DRA), which requires that each home health agency shall submit to the Secretary such data that the Secretary determines are appropriate for the measurement of health care quality. Such data shall be submitted in a form and manner, and at a time, specified by the Secretary for purposes of this clause.
In the CY 2015 HH PPS Final rule (79 FR 38387), we reported information on a study performed by the Department of Health & Human Services, Office of the Inspector General (OIG) in February 2012 to: (1) Determine the extent to which HHAs met federal reporting requirements for the OASIS data; (2) to determine the extent to which states met federal reporting requirements for OASIS data; and (3) to determine the extent to which CMS was overseeing the accuracy and completeness of OASIS data submitted by HHAs. Based on the OIG report we proposed a performance requirement for submission of OASIS quality data, which would be responsive to the recommendations of the OIG.
In response to these requirements and the OIG report, we designed a pay-for-reporting performance system model that could accurately measure the level of an HHA's submission of OASIS data. The performance system is based on the principle that each HHA is expected to submit a minimum set of two matching assessments for each patient admitted to their agency. These matching assessments together create what is considered a quality episode of care, consisting ideally of a Start of Care (SOC) or Resumption of Care (ROC) assessment and a matching End of Care (EOC) assessment. However, it was determined that there are several scenarios that could meet this matching assessment requirement of the new pay-for-reporting performance requirement. These scenarios or quality assessments are defined as assessments that create a quality episode of care during the reporting period or could create a quality episode if the reporting period were expanded to an earlier reporting period or into the next reporting period.
Seven types of assessments submitted by an HHA fit this definition of a quality assessment. These are:
1. A Start of Care (SOC; M0100 = `01') or Resumption of Care (ROC; M0100 = `03') assessment that can be matched to an End of Care (EOC; M0100 = `06', `07', `08', or `09') assessment. These SOC/ROC assessments are the first assessment in the pair of assessments that create a standard quality of care episode describe in the previous paragraph.
2. An End of Care (EOC) assessment that can be matched to a Start of Care (SOC) or Resumption of Care (ROC) assessment. These EOC assessments are the second assessment in the pair of assessments that create a standard quality of care episode describe in the previous paragraph.
3. A SOC/ROC assessment that could begin an episode of care, but the assessment occurs in the last 60 days of the performance period. This is labeled as a Late SOC/ROC quality assessment. The assumption is that the EOC assessment will occur in the next reporting period.
4. An EOC assessment that could end an episode of care that began in the previous reporting period, (that is, an EOC that occurs in the first 60 days of the performance period). This is labeled as an Early EOC quality assessment. The
5. A SOC/ROC assessment that is followed by one or more follow-up assessments, the last of which occurs in the last 60 days of the performance period. This is labeled as an SOC/ROC Pseudo Episode quality assessment.
6. An EOC assessment is preceded by one or more follow-up assessments, the first of which occurs in the first 60 days of the performance period. This is labeled an EOC Pseudo Episode quality assessment.
7. A SOC/ROC assessment that is part of a known one-visit episode. This is labeled as a One-Visit episode quality assessment. This determination is made by consulting HH claims data.
SOC, ROC, and EOC assessments that do not meet any of these definitions are labeled as Non-Quality assessments. Follow-up assessments (that is, where the M0100 Reason for Assessment = `04' or `05') are considered Neutral assessments and do not count toward or against the pay-for-reporting performance requirement.
Compliance with this performance requirement can be measured through the use of an uncomplicated mathematical formula. This pay-for-reporting performance requirement metric has been titled as the “Quality Assessments Only” (QAO) formula because only those OASIS assessments that contribute, or could contribute, to creating a quality episode of care are included in the computation.
The formula based on this definition is as follows:
Our ultimate goal is to require all HHAs to achieve a pay-for-reporting performance requirement compliance rate of 90 percent or more, as calculated using the QAO metric illustrated above. In the CY 2015 HH PPS final rule (79 FR 66074), we proposed implementing a pay-for-reporting performance requirement over a three-year period. After consideration of the public comments received, we adopted as final our proposal to establish a pay-for-reporting performance requirement for assessments submitted on or after July 1, 2015 and before June 30, 2016 with appropriate start of care dates, HHAs must score at least 70 percent on the QAO metric of pay-for-reporting performance requirement or be subject to a 2 percentage point reduction to their market basket update for CY 2017.
HHAs have been statutorily required to report OASIS for a number of years and therefore should have many years of experience with the collection of OASIS data and transmission of this data to CMS. Given the length of time that HHAs have been mandated to report OASIS data and based on preliminary analyses that indicate that the majority of HHAs are already achieving the target goal of 90 percent on the QAO metric, we believe that HHAs would adapt quickly to the implementation of the pay-for-reporting performance requirement, if phased in over a three-year period.
In the CY2015 rule, we did not finalize a proposal to increase the reporting requirement in 10 percent increments over a two-year period until the maximum rate of 90 percent is reached, but instead proposed to analyze historical data to set the reporting requirements. To set the threshold for the 2nd year, we analyzed the most recently available data, from 2013 and 2014, to make a determination about what the pay-for-reporting performance requirement should be. Specifically, we reviewed OASIS data from this time period simulating the pay-for-reporting performance 70 percent submission requirement to determine the hypothetical performance of each HHA as if the pay-for-reporting performance requirement were in effect during the reporting period preceding its implementation. This analysis indicated a nominal increase of 10 percent each year would provide the greatest opportunity for successful implementation versus an increase of 20 percent from year 1 to year 2.
Based on this analysis, we propose to set the performance threshold at 80 percent for the reporting period from July 1, 2016 through June 30, 2017. For the reporting period from July 1, 2017 through June 30, 2018 and thereafter, we propose the performance threshold would be 90 percent.
We provided a report to each HHA of their hypothetical performance under the pay-for-reporting performance requirement during the 2014–2015 pre-implementation reporting period in June 2015. On January 1, 2015, the data submission process for OASIS converted from the current state-based OASIS submission system to a new national OASIS submission system known as the Assessment Submission and Processing (ASAP) System. On July 1, 2015, when the pay-for-reporting performance requirement of 70 percent goes into effect, providers would be required to submit their OASIS assessment data into the ASAP system. Successful submission of an OASIS assessment would consist of the submission of the data into the ASAP system with a receipt of no fatal error messages. Error messages received during submission can be an indication of a problem that occurred during the submission process and could also be an indication that the OASIS assessment was rejected. Successful submission can be verified by ascertaining that the submitted assessment data resides in the national database after the assessment has met all of the quality standards for completeness and accuracy during the submission process. Should one or more OASIS assessments submitted by a HHA be rejected due to an IT/servers issue caused by CMS, we may, at our discretion, excuse the non-submission of OASIS data. We anticipate that such a scenario would rarely, if ever, occur. In the event that a HHA believes, they were unable to submit OASIS assessments due to an IT/server issue on the part of CMS, the HHA should be prepared to provide any documentation or proof available, which demonstrates that no fault on their part contributed to the failure of the OASIS records to transmit to CMS.
The initial performance period for the pay-for-reporting performance requirement would be July 1, 2015 through June 30, 2016. Prior to and during this performance period, we have scheduled Open Door Forums and webinars to educate HHA personnel as needed about the pay-for-reporting performance requirement program and the pay-for- reporting performance QAO metric, and distributed individual provider preview reports. Additionally, OASIS Education Coordinators (OECs) would be trained to provide state-level instruction on this program and metric. We have already posted a report, which provides a detailed explanation of the methodology for this pay-for-reporting QAO methodology. To view this report, go to:
In the CY 2015 HH PPS final rule (79 FR 66031), we stated that the home health quality measures reporting requirements for Medicare-certified agencies include the Home Health Care CAHPS® (HHCAHPS) Survey for the CY 2015 Annual Payment Update (APU). We maintained the stated HHCAHPS data requirements for CY 2015 set out in previous rules, for the continuous monthly data collection and quarterly data submission of HHCAHPS data.
As part of the HHS Transparency Initiative, we implemented a process to measure and publicly report patient experiences with home health care, using a survey developed by the Agency for Healthcare Research and Quality's (AHRQ's) Consumer Assessment of Healthcare Providers and Systems (CAHPS®) program and originally endorsed by the NQF in March 2009 (NQF Number 0517) and recently NQF re-endorsed in 2015. The HHCAHPS survey is part of a family of CAHPS® surveys that asks patients to report on and rate their experiences with health care. The HHCAHPS Survey is approved under OMB Control Number 0938–1066 through May 31, 2017. The Home Health Care CAHPS® (HHCAHPS) survey presents home health patients with a set of standardized questions about their home health care providers and about the quality of their home health care.
Prior to the HHCAHPS survey, there was no national standard for collecting information about patient experiences that enabled valid comparisons across all HHAs. The history and development process for HHCAHPS has been described in previous rules and is also available on the official HHCAHPS Web site at
For public reporting purposes, we report five measures from the HHCAHPS Survey—three composite measures and two global ratings of care that are derived from the questions on the HHCAHPS survey. The publicly reported data are adjusted for differences in patient mix across HHAs. We update the HHCAHPS data on Home Health Compare on
• Patient care (Q9, Q16, Q19, and Q24);
• Communications between providers and patients (Q2, Q15, Q17, Q18, Q22, and Q23); and
• Specific care issues on medications, home safety, and pain (Q3, Q4, Q5, Q10, Q12, Q13, and Q14).
The two global ratings are the overall rating of care given by the HHA's care providers (Q20), and the patient's willingness to recommend the HHA to family and friends (Q25).
The HHCAHPS survey is currently available in English, Spanish, Chinese, Russian, and Vietnamese. The OMB number on these surveys is the same (0938–1066). All of these surveys are on the Home Health Care CAHPS® Web site,
All of the requirements about home health patient eligibility for the HHCAHPS survey and conversely, which home health patients are ineligible for the HHCAHPS survey are delineated and detailed in the
Home health patients are
• Are under the age of 18;
• Are deceased prior to the date the sample is pulled;
• Receive hospice care;
• Receive routine maternity care only;
• Are not considered survey eligible because the state in which the patient lives restricts release of patient information for a specific condition or illness that the patient has; or
• No Publicity patients, defined as patients who on their own initiative at their first encounter with the HHAs make it very clear that no one outside of the agencies can be advised of their patient status, and no one outside of the HHAs can contact them for any reason.
We stated in previous rules that Medicare-certified HHAs are required to contract with an approved HHCAHPS survey vendor. This requirement continues, and Medicare-certified agencies also must provide on a monthly basis a list of all their survey-eligible home health care patients served to their respective HHCAHPS survey vendors. Agencies are not allowed to influence at all how their patients respond to the HHCAHPS survey.
As previously required, HHCAHPS survey vendors are required to attend introductory and all update trainings conducted by CMS and the HHCAHPS Survey Coordination Team, as well as to pass a post-training certification test. Update training is required annually for all approved HHCAHPS survey vendors. We have approximately 30 approved HHCAHPS survey vendors. The most current list of approved HHCAHPS survey vendors is available at
We stated in prior final rules that all approved HHCAHPS survey vendors are required to participate in HHCAHPS oversight activities to ensure compliance with HHCAHPS protocols, guidelines, and survey requirements. The purpose of the oversight activities is to ensure that approved HHCAHPS survey vendors follow the
• Organizational Background and Staff Experience;
• Work Plan;
• Sampling Plan;
• Survey Implementation Plan;
• Data Security, Confidentiality and Privacy Plan; and
• Questionnaire Attachments.
As part of the oversight activities, the HHCAHPS Survey Coordination Team conducts on-site visits to all approved HHCAHPS survey vendors. The purpose of the site visits is to allow the HHCAHPS Coordination Team to observe the entire HHCAHPS Survey implementation process, from the sampling stage through file preparation and submission, as well as to assess data security and storage. The HHCAHPS Survey Coordination Team reviews the HHCAHPS survey vendor's survey systems, and assesses administration protocols based on the
• Survey management and data systems;
• Printing and mailing materials and facilities;
• Telephone call center facilities;
• Data receipt, entry and storage facilities; and
• Written documentation of survey processes.
After the site visits, HHCAHPS survey vendors are given a defined time period in which to correct any identified issues and provide follow-up documentation of corrections for review. HHCAHPS survey vendors are subject to follow-up site visits on an as-needed basis.
In the CY 2013 HH PPS final rule (77 FR 67094, 67164), we codified the current guideline that all approved HHCAHPS survey vendors fully comply with all HHCAHPS oversight activities. We included this survey requirement at § 484.250(c)(3).
In the CY 2015 HH PPS final rule (79 FR 66031), we stated that for the CY 2016 APU, we would require continued monthly HHCAHPS data collection and reporting for four quarters. The data collection period for CY 2016, APU includes the second quarter 2014 through the first quarter 2015 (the months of April 2014 through March 2015). Although these dates are past, we wished to state them in this proposed rule so that HHAs are again reminded of what months constituted the requirements for the CY 2016 APU. HHAs are required to submit their HHCAHPS data files to the HHCAHPS Data Center for the HHCAHPS data from the first quarter of 2015 data by 11:59 p.m., EST on July 16, 2015. This deadline is firm; no exceptions are permitted.
For the CY 2016 APU, we required that all HHAs that had fewer than 60 HHCAHPS-eligible unduplicated or unique patients in the period of April 1, 2013 through March 31, 2014 are exempted from the HHCAHPS data collection and submission requirements for the CY 2016 APU, upon completion of the CY 2016 HHCAHPS Participation Exemption Request form, and upon CMS verification of the HHA patient counts. Agencies with fewer than 60 HHCAHPS-eligible, unduplicated or unique patients in the period of April 1, 2013, through March 31, 2014, were required to submit their patient counts on the HHCAHPS Participation Exemption Request form for the CY 2016 APU posted on
We automatically exempt HHAs receiving Medicare certification after the period in which HHAs do their patient counts. HHAs receiving Medicare certification on or after April 1, 2014 are exempt from the HHCAHPS reporting requirement for the CY 2016 APU. These newly-certified HHAs did not need to complete a HHCAHPS Participation Exemption Request form for the CY 2016 APU.
For the CY 2017 APU, we require continued monthly HHCAHPS data collection and reporting for four quarters. The data collection period for the CY 2017, APU includes the second quarter 2015 through the first quarter 2016 (the months of April 2015 through March 2016). HHAs would be required to submit their HHCAHPS data files to the HHCAHPS Data Center for the second quarter 2015 by 11:59 p.m., EST on October 15, 2015; for the third quarter 2015 by 11:59 p.m., EST on January 21, 2016; for the fourth quarter 2015 by 11:59 p.m., EST on April 21, 2016; and for the first quarter 2016 by 11:59 p.m., EST on July 21, 2016. These deadlines will be firm; no exceptions will be permitted.
For the CY 2017 APU, we require that all HHAs that have fewer than 60 HHCAHPS-eligible unduplicated or unique patients in the period of April 1, 2014 through March 31, 2015 are exempted from the HHCAHPS data collection and submission requirements for the CY 2017 APU, upon completion of the CY 2017 HHCAHPS Participation Exemption Request form, and upon CMS verification of the HHA patient counts. Agencies with fewer than 60 HHCAHPS-eligible, unduplicated or unique patients in the period of April 1, 2014 through March 31, 2015, are required to submit their patient counts on the HHCAHPS Participation Exemption Request form for the CY 2017 APU posted on
We automatically exempt HHAs receiving Medicare certification after the period in which HHAs do their patient counts. HHAs receiving Medicare certification on or after April 1, 2015 are exempt from the HHCAHPS reporting requirement for the CY 2017 APU. These newly-certified HHAs did not need to complete a HHCAHPS Participation Exemption Request form for the CY 2017 APU.
For the CY 2018 APU, we require continued monthly HHCAHPS data collection and reporting for four quarters. The data collection period for the CY 2018, APU includes the second quarter 2016 through the first quarter 2017 (the months of April 2016 through March 2017). HHAs would be required to submit their HHCAHPS data files to the HHCAHPS Data Center for the second quarter 2016 by 11:59 p.m., EST on October 20, 2016; for the third quarter 2016 by 11:59 p.m., EST on January 19, 2017; for the fourth quarter 2016 by 11:59 p.m., EST on April 20, 2017; and for the first quarter 2017 by 11:59 p.m., EST on July 20, 2017. These deadlines will be firm; no exceptions will be permitted.
For the CY 2018 APU, we require that all HHAs that have fewer than 60 HHCAHPS-eligible unduplicated or unique patients in the period of April 1, 2015 through March 31, 2016 are exempted from the HHCAHPS data collection and submission requirements for the CY 2018 APU, upon completion of the CY 2018 HHCAHPS Participation Exemption Request form, and upon CMS verification of the HHA patient counts. Agencies with fewer than 60 HHCAHPS-eligible, unduplicated or unique patients in the period of April 1, 2015 through March 31, 2016, are required to submit their patient counts on the HHCAHPS Participation Exemption Request form for the CY 2018 APU posted on
We automatically exempt HHAs receiving Medicare certification after the period in which HHAs do their patient counts. HHAs receiving Medicare Certification on or after April 1, 2016 are exempt from the HHCAHPS reporting requirement for the CY 2018 APU. These newly-certified HHAs did not need to complete a HHCAHPS Participation Exemption Request form for the CY 2018 APU.
HHAs should monitor their respective HHCAHPS survey vendors to ensure that vendors submit their HHCAHPS data on time, by accessing their HHCAHPS Data Submission Reports on
We will continue HHCAHPS oversight activities as finalized in the CY 2014 rule. In the CY 2013 HH PPS final rule (77 FR 6704, 67164), we codified the current guideline that all approved HHCAHPS survey vendors must fully comply with all HHCAHPS oversight activities. We included this survey requirement at § 484.250(c)(3).
We propose to continue the OASIS and HHCAHPS reconsiderations and appeals process that we have finalized and that we have used for prior periods for the CY 2012, CY 2013, CY 2014, and CY 2015 APU determinations. We have described the reconsiderations process requirements in the CMS Technical Direction Letter that we sent to the affected HHAs, on or in late September. HHAs have 30 days from their receipt of the Technical Direction Letter informing them that they did not meet the OASIS and HHCAHPS requirements for the CY period, to send all documentation that supports their requests for reconsideration to CMS. It is important that the affected HHAs send in comprehensive information in their reconsideration letter/package because we would not contact the affected HHAs to request additional information or to clarify incomplete or inconclusive information. If clear evidence to support a finding of compliance is not present, the 2 percent reduction in the APU would be upheld. If clear evidence of compliance is present, the 2 percent reduction for the APU would be reversed. We notify affected HHAs by December 31st annually for the APU period that begins on January 1st. If we determine to uphold the 2 percent reduction, the HHA may further appeal the 2 percent reduction via the Provider Reimbursement Review Board (PRRB) appeals process. The PRRB contact information is provided to the HHAs receiving letters in December about the CMS reconsideration decisions.
Providers who wish to submit a reconsideration request should continue to follow the reconsideration and appeals process as finalized in the CY 2012, CY 2013, CY 2014, and CY 2015 Home Health Prospective Payment System Rate Update Final Rules.
We are not proposing any changes to the participation requirements, or to the requirements pertaining to the implementation of the Home Health CAHPS® Survey (HHCAHPS). We only updated the information to reflect the dates in the future APU years. We again strongly encourage HHAs to keep up-to-date about the HHCAHPS by regularly viewing the official Web site for the HHCAHPS at
Section 1895(b)(3)(B)(v)(III) of the Act and section 1899B(f) of the IMPACT Act states the Secretary shall establish procedures for making data submitted under subclause (II) available to the public. Such procedures shall ensure that a home health agency has the opportunity to review the data that is to be made public with respect to the agency prior to such data being made public. We recognize that public reporting of quality data is a vital component of a robust quality reporting program and are fully committed to ensuring that the data made available to the public be meaningful and that comparing performance across home health agencies requires that measures be constructed from data collected in a standardized and uniform manner. We also recognize the need to ensure that each home health agency has the opportunity to review the data before publication. Medicare home health regulations, as codified at § 484.250(a), requires HHAs to submit OASIS assessments and Home Health Care Consumer Assessment of Healthcare Providers and Systems Survey® (HHCAHPS) data to meet the quality reporting requirements of section 1895(b)(3)(B)(v) of the Act.
In addition, beginning April 1, 2015 HHAs began to receive Provider Preview Reports (for all Process Measures and Outcome Measures) on a quarterly, rather than annual, basis. The opportunity for providers to review their data and to submit corrections prior to public reporting aligns with the other quality reporting programs and the requirement for provider review under the IMPACT Act. We provide quality measure data to HHAs via the Certification and Survey Provider Enhanced Reports (CASPER reports), which are available through the CMS Health Care Quality Improvement and Evaluation System (QIES).
As part of our ongoing efforts to make healthcare more transparent, affordable, and accountable, the HH QRP has developed a CMS Compare Web site for home health agencies, which identifies home health providers based on the areas they serve. Consumers can search for all Medicare-certified home health providers that serve their city or ZIP code and then find the agencies offering the types of services they need. A subset of the HH quality measures has been publicly reported on the Home Health Compare (HH Compare) Web site since 2003. The selected measures that are made available to the public can be viewed on the HH Compare Web site located at
The Affordable Care Act calls for transparent, easily understood information on provider quality to be publicly reported and made widely available. To provide home health care consumers with a summary of existing quality measures in an accessible format, we plan to publish a star rating based on the quality of care measures for home health agencies on Home Health Compare starting in July 2015. This is part of our plan to adopt star ratings across all Medicare.gov Compare Web sites. Star ratings are currently publicly displayed on Nursing Home Compare, Physician Compare, the Medicare Advantage Plan Finder, and Dialysis Facility Compare, and they are scheduled to be displayed on Hospital Compare in 2015.
The Quality of Patient Care star rating methodology assigns each home health agency a rating between one (1) and five (5) stars, using half stars for adjustment and reporting. All Medicare-certified home health agencies are eligible to receive a Quality of Patient Care star rating providing that they have quality data reported on at least 5 out of the 9 quality measures that are included in the calculation.
Home health agencies would continue to have prepublication access to their agency's quality data, which enables each agency to know how it is
The Quality of Patient Care star ratings methodology was developed through a transparent process the included multiple opportunities for stakeholder input, which was subsequently the basis for refinements to the methodology. An initial proposed methodology for calculating the Quality of Patient Care star ratings was posted on the CMS.gov Web site in December 2014. CMS then held two Special Open Door Forums (SODFs) on December 17, 2014 and February 5, 2015 to present the proposed methodology and solicit input. At each SODF, stakeholders provided immediate input, and were invited to submit additional comments via the Quality of Patient Care star ratings Help Desk mailbox:
Additional information regarding the Quality of Patient Care star rating would be posted on the star ratings Web page at
While this proposed rule contains information collection requirements, this rule does not add new, nor revise any of the existing information collection requirements, or burden estimate. The information collection requirements discussed in this rule for the OASIS–C1 data item set had been previously approved by the Office of Management and Budget (OMB) on February 6, 2014 and scheduled for implementation on October 1, 2014. The extension of OASIS–C1/ICD–9 version was reapproved under OMB control number 0938–0760 with a current expiration date of March 31, 2018. This version of the OASIS will be discontinued once the OASIS–C1/ICD–10 version is approved and implemented. In addition, to facilitate the reporting of OASIS data as it relates to the implementation of ICD–10 on October 1, 2015, CMS submitted a new request for approval to OMB for the OASIS–C1/ICD–10 version under the Paperwork Reduction Act (PRA) process. CMS is requesting a new OMB control number for the proposed revised OASIS item as announced in the 30-day
Because of the large number of public comments we normally receive on
Section 1895(b)(1) of the Act requires the Secretary to establish a HH PPS for all costs of HH services paid under Medicare. In addition, section 1895(b)(3)(A) of the Act requires (1) the computation of a standard prospective payment amount include all costs for HH services covered and paid for on a reasonable cost basis and that such amounts be initially based on the most recent audited cost report data available to the Secretary, and (2) the standardized prospective payment amount be adjusted to account for the effects of case-mix and wage levels among HHAs. Section 1895(b)(3)(B) of the Act addresses the annual update to the standard prospective payment amounts by the HH applicable percentage increase. Section 1895(b)(4) of the Act governs the payment computation. Sections 1895(b)(4)(A)(i) and (b)(4)(A)(ii) of the Act require the standard prospective payment amount to be adjusted for case-mix and geographic differences in wage levels. Section 1895(b)(4)(B) of the Act requires the establishment of appropriate case-mix adjustment factors for significant variation in costs among different units of services. Lastly, section 1895(b)(4)(C) of the Act requires the establishment of wage adjustment factors that reflect the relative level of wages, and wage-related costs applicable to HH services furnished in a geographic area compared to the applicable national average level.
Section 1895(b)(3)(B)(iv) of the Act provides the Secretary with the authority to implement adjustments to the standard prospective payment amount (or amounts) for subsequent years to eliminate the effect of changes in aggregate payments during a previous year or years that was the result of changes in the coding or classification of different units of services that do not reflect real changes in case-mix. Section 1895(b)(5) of the Act provides the Secretary with the option to make changes to the payment amount otherwise paid in the case of outliers because of unusual variations in the type or amount of medically necessary care. Section 1895(b)(3)(B)(v) of the Act requires HHAs to submit data for purposes of measuring health care quality, and links the quality data submission to the annual applicable percentage increase.
Section 421(a) of the MMA requires that HH services furnished in a rural area, for episodes and visits ending on or after April 1, 2010, and before January 1, 2016, receive an increase of 3 percent of the payment amount otherwise made under section 1895 of the Act. Section 210 of the MACRA amended section 421(a) of the MMA to extend the 3 percent increase to the payment amounts for serviced furnished in rural areas for episodes and visits ending before January 1, 2018.
Section 3131(a) of the Affordable Care Act mandates that starting in CY 2014, the Secretary must apply an adjustment to the national, standardized 60-day episode payment rate and other amounts applicable under section 1895(b)(3)(A)(i)(III) of the Act to reflect factors such as changes in the number of visits in an episode, the mix of services in an episode, the level of intensity of services in an episode, the average cost of providing care per episode, and other relevant factors. In addition, section 3131(a) of the Affordable Care Act mandates that rebasing must be phased-in over a 4-year period in equal increments, not to exceed 3.5 percent of the amount (or amounts) as of the date of enactment (2010) under section 1895(b)(3)(A)(i)(III) of the Act, and be fully implemented in CY 2017.
The proposed HHVBP model would apply a payment adjustment based on
We have examined the impacts of this rule as required by Executive Order 12866 on Regulatory Planning and Review (September 30, 1993), Executive Order 13563 on Improving Regulation and Regulatory Review (January 18, 2011), the Regulatory Flexibility Act (RFA) (September 19, 1980, Pub. L. 96–354), section 1102(b) of the Act, section 202 of the Unfunded Mandates Reform Act of 1995 (UMRA, March 22, 1995; Pub. L. 104–4), Executive Order 13132 on Federalism (August 4, 1999), and the Congressional Review Act (5 U.S.C. 804(2)).
Executive Orders 12866 and 13563 direct agencies to assess all costs and benefits of available regulatory alternatives and, if regulation is necessary, to select regulatory approaches that maximize net benefits (including potential economic, environmental, public health and safety effects, distributive impacts, and equity). Executive Order 13563 emphasizes the importance of quantifying both costs and benefits, of reducing costs, of harmonizing rules, and of promoting flexibility. The net transfer impacts related to the proposed changes in payments under the HH PPS for CY 2016 are estimated to be −$350 million. The savings impacts related to the proposed HHVBP model are estimated at a total projected 5-year gross savings of $380 million assuming a very conservative savings estimate of a 6 percent annual reduction in hospitalizations and a 1.0 percent annual reduction in SNF admissions. In accordance with the provisions of Executive Order 12866, this regulation was reviewed by the Office of Management and Budget.
The update set forth in this rule applies to Medicare payments under HH PPS in CY 2016. Accordingly, the following analysis describes the impact in CY 2016 only. We estimate that the net impact of the proposals in this rule is approximately $350 million in decreased payments to HHAs in CY 2016. We applied a wage index budget neutrality factor and a case-mix weights budget neutrality factor to the rates as discussed in section III.C.3 of this proposed rule; therefore, the estimated impact of the 2016 wage index proposed in section III.C.3 of this proposed rule and the recalibration of the case-mix weights for 2016 proposed in section III.B. of this proposed rule is zero. The −$350 million impact reflects the distributional effects of the 2.3 percent HH payment update percentage ($420 million increase), the effects of the third year of the four-year phase-in of the rebasing adjustments to the national, standardized 60-day episode payment amount, the national per-visit payment rates, and the NRS conversion factor for an impact of −2.5 percent ($470 million decrease), and the effects of the −1.72 percent adjustment for nominal case-mix growth ($300 million decrease). The $350 million in decreased payments is reflected in the last column of the first row in Table 24 as a 0.1 percent decrease in expenditures when comparing CY 2015 payments to estimated CY 2016 payments.
The RFA requires agencies to analyze options for regulatory relief of small entities, if a rule has a significant impact on a substantial number of small entities. For purposes of the RFA, small entities include small businesses, nonprofit organizations, and small governmental jurisdictions. Most hospitals and most other providers and suppliers are small entities, either by nonprofit status or by having revenues of less than $7.5 million to $38.5 million in any one year. For the purposes of the RFA, we estimate that almost all HHAs are small entities as that term is used in the RFA. Individuals and states are not included in the definition of a small entity. The economic impact assessment is based on estimated Medicare payments (revenues) and HHS's practice in interpreting the RFA is to consider effects economically “significant” only if greater than 5 percent of providers reach a threshold of 3 to 5 percent or more of total revenue or total costs. The majority of HHAs' visits are Medicare-paid visits and therefore the majority of HHAs' revenue consists of Medicare payments. Based on our analysis, we conclude that the policies proposed in this rule will result in an estimated total impact of 3 to 5 percent or more on Medicare revenue for greater than 5 percent of HHAs. Therefore, the Secretary has determined that this HH PPS proposed rule will have a significant economic impact on a substantial number of small entities. Further detail is presented in Table 24, by HHA type and location.
With regards to options for regulatory relief, we note that in the CY 2014 HH PPS final rule we finalized rebasing adjustments to the national, standardized 60-day episode rate, non-routine supplies (NRS) conversion factor, and the national per-visit payment rates for each year, 2014 through 2017 as described in section II.C and III.C.3 of this proposed rule. Since the rebasing adjustments are mandated by section 3131(a) of the Affordable Care Act, we cannot offer HHAs relief from the rebasing adjustments for CY 2016. For the proposed reduction to the national, standardized 60-day episode payment amount of 1.72 percent for CY 2016 described in section III.B.2 of this proposed rule, we believe it is appropriate to reduce the national, standardized 60-day episode payment amount to account for the estimated increase in nominal case-mix in order to move towards more accurate payment for the delivery of home health services where payments better align with the costs of providing such services. In the alternatives considered section below, we note that we considered proposing the full 3.41 percent reduction to the 60-day episode rate in CY 2016 to account for nominal case-mix growth between CY 2012 and CY 2014. However, we instead proposed to reduce the 60-day episode rate by 1.72 percent in CY 2016 and 1.72 percent in CY 2017 to account for estimated nominal case-mix growth between CY 2012 and CY 2014.
Executive Order 13563 specifies, to the extent practicable, agencies should assess the costs of cumulative regulations. However, given potential utilization pattern changes, wage index changes, changes to the market basket forecasts, and unknowns regarding future policy changes, we believe it is neither practicable nor appropriate to forecast the cumulative impact of the rebasing adjustments on Medicare payments to HHAs for future years at this time. Changes to the Medicare program may continue to be made as a result of the Affordable Care Act, or new statutory provisions. Although these changes may not be specific to the HH PPS, the nature of the Medicare program is such that the changes may interact, and the complexity of the interaction of these changes would make it difficult to predict accurately the full scope of the impact upon HHAs for future years
In addition, section 1102(b) of the Act requires us to prepare a RIA if a rule may have a significant impact on the operations of a substantial number of small rural hospitals. This analysis must conform to the provisions of section 603 of RFA. For purposes of section 1102(b) of the Act, we define a small rural hospital as a hospital that is located outside of a metropolitan statistical area and has fewer than 100 beds. This proposed rule applies to HHAs. Therefore, the Secretary has determined that the HH PPS proposed rule will not have a significant economic impact on the operations of small rural hospitals.
To test the impact of upside and downside value-based payment adjustments, beginning in calendar year 2018 and in each succeeding calendar year through calendar year 2022, the proposed model would adjust the final claim payment amount for a home health agency for each episode in a calendar year by an amount equal to the applicable percent. For purposes of this proposed rule, we have limited our analysis of the economic impacts to the value-based incentive payment adjustments. Under the proposed model design, the incentive payment adjustments would be limited to the total payment reductions to home health agencies included in the model and would be no less than the total amount available for value-based incentive payment adjustment. Overall, the distributive impact of this proposed rule is estimated at $380 million for CY 2018–2022. Therefore, this proposed rule is economically significant and thus a major rule under the Congressional Review Act. The proposed model would test the effect on quality and costs of care by applying payment adjustments based on HHAs' performance on quality measures. This proposed rule was developed based on extensive research and experience with value-based purchasing models.
Guidance issued by the Department of Health and Human Services interpreting the Regulatory Flexibility Act considers the effects economically ‘significant’ only if greater than 5 percent of providers reach a threshold of 3 to 5 percent or more of total revenue or total costs. Among the over 1900 HHAs in the selected states that would be expected to be included in the proposed HHVBP model, we estimate that the maximum percent payment adjustment resulting from this proposed rule will only be greater than −5 percent for 10 percent of the HHAs included in the model (using the 8 percent maximum payment adjustment threshold applied in CY2021 and CY2022). As a result, only 2 percent of all HHA providers nationally would be significantly impacted, falling well below the RFA threshold. In addition, only HHAs that are impacted with lower payments are those providers that provide the poorest quality which is the main tenet of the model. This falls well below the threshold for economic significance established by HHS for requiring a more detailed impact assessment under the RFA. Thus, we are not preparing an analysis under the RFA because the Secretary has determined that this proposed rule would not have a significant economic impact on a substantial number of small entities.
In addition, section 1102(b) of the Act requires us to prepare a regulatory impact analysis if a rule may have a significant impact on the operations of a substantial number of small rural HHAs. This analysis must conform to the provisions of section 603 of the RFA. For purposes of section 1102(b) of the Act, we have identified less than 5 percent of HHAs included in the proposed selected states that primarily serve beneficiaries that reside in rural areas (greater than 50 percent of beneficiaries served). We are not preparing an analysis under section 1102(b) of the Act because the Secretary has determined that the proposed HHVBP model would not have a significant impact on the operations of a substantial number of small rural HHAs.
Section 202 of the Unfunded Mandates Reform Act of 1995 also requires that agencies assess anticipated costs and benefits before issuing any rule whose mandates require spending in any 1 year of $100 million in 1995 dollars, updated annually for inflation. In 2015, that threshold is approximately $144 million. This rule will have no consequential effect on state, local, or tribal governments or on the private sector.
Executive Order 13132 establishes certain requirements that an agency must meet when it promulgates a proposed rule (and subsequent final rule) that imposes substantial direct requirement costs on State and local governments, preempts state law, or otherwise has Federalism implications. Since this regulation does not impose any costs on state or local governments, the requirements of Executive Order 13132 are not applicable.
In accordance with the provisions of Executive Order 12866, this regulation was reviewed by the Office of Management and Budget.
This proposed rule sets forth updates for CY 2016 to the HH PPS rates contained in the CY 2015 HH PPS final rule (79 FR 66032 through 66118). The impact analysis of this proposed rule presents the estimated expenditure effects of policy changes proposed in this rule. We use the latest data and best analysis available, but we do not make adjustments for future changes in such variables as number of visits or case-mix.
This analysis incorporates the latest estimates of growth in service use and payments under the Medicare HH benefit, based primarily on preliminary Medicare claims data from 2014. We note that certain events may combine to limit the scope or accuracy of our impact analysis, because such an analysis is future-oriented and, thus, susceptible to errors resulting from other changes in the impact time period assessed. Some examples of such possible events are newly-legislated general Medicare program funding changes made by the Congress, or changes specifically related to HHAs. In addition, changes to the Medicare program may continue to be made as a result of the Affordable Care Act, or new statutory provisions. Although these changes may not be specific to the HH PPS, the nature of the Medicare program is such that the changes may interact, and the complexity of the interaction of these changes could make it difficult to predict accurately the full scope of the impact upon HHAs.
Table 24 represents how HHA revenues are likely to be affected by the policy changes proposed in this rule. For this analysis, we used an analytic file with linked CY 2014 HH claims data (as of December 31, 2014) for dates of service that ended on or before December 31, 2014, and OASIS assessments. The first column of Table 24 classifies HHAs according to a number of characteristics including provider type, geographic region, and urban and rural locations. The second column shows the number of facilities in the impact analysis. The third column shows the payment effects of proposed CY 2016 wage index. The fourth column shows the payment
The last column shows the combined effects of all the proposed policies for HH PPS. Overall, it is projected that aggregate payments in CY 2016 will decrease by 1.8 percent. As illustrated in Table 24, the combined effects of all of the changes vary by specific types of providers and by location. We note that some individual HHAs within the same group may experience different impacts on payments than others due to the distributional impact of the CY 2016 wage index, the extent to which HHAs had episodes in case-mix groups where the case-mix weight decreased for CY 2016 relative to CY 2015, the percentage of total HH PPS payments that were subject to the low-utilization payment adjustment (LUPA) or paid as outlier payments, and the degree of Medicare utilization.
Table 25 displays our analysis of the distribution of possible payment adjustments at the 5 percent, 6 percent and 8 percent rates that are being proposed in the model based on 2013–2014 data, providing information on the estimated impact of this proposed rule. We note that this impact analysis is based on the aggregate value of all 9 states identified in section IV.C.2. of this proposed rule by applying the proposed state selection methodology.
Table 26 displays our analysis of the distribution of possible payment adjustments based on 2013–2014 data, providing information on the estimated impact of this proposed rule. We note that this impact analysis is based on the aggregate value of all nine states (identified in section IV.C.2. of this proposed rule) by applying the proposed state selection methodology.
If our methodology is finalized as proposed, all Medicare-certified HHAs that provide services in Massachusetts, Maryland, North Carolina, Florida, Washington, Arizona, Iowa, Nebraska, and Tennessee will be required to compete in this model. However, should the methodology we propose in this rule change as a result of comments received during the rulemaking process, it could result in different states being selected for the model. In such an event, we would apply the final methodology and announce the selected states in the final rule. The estimates presented here may also change accordingly.
Value-based incentive payment adjustments for the estimated 1,900 plus HHAs in the proposed selected states that would compete in the HHVBP model are stratified by the size as defined in section F. For example, Arizona has 31 HHAs that do not provide services to enough beneficiaries to be required to complete CAHPS
The next columns provide the distribution of scores by percentile; we see that the value-based incentive percentage payments for home health agencies in Arizona range from −3.3 percent at the 10th percentile to +5.0 percent at the 90th percentile, while the value-based incentive payment at the 50th percentile is 0.56 percent.
The smaller-volume HHA cohorts table identifies that some consideration will have to be made for MD, WA and TN where there are too few HHAs in the smaller-volume cohort and would be included in the larger-volume cohort without being measured on HHCAHPS.
Table 27 provides the payment adjustment distribution based on proportion of dual-eligible beneficiaries, average case mix (using HCC scores), proportion that reside in rural areas, as well as HHA organizational status. Besides the observation that higher proportion of dually-eligible beneficiaries serviced is related to better performance, the payment adjustment distribution is consistent with respect to these four categories.
The TPS score and the payment methodology at the state and size level were calculated so that each home health agency's payment adjustment was calculated as it would be in the model. Hence, the values of each separate analysis in the tables are representative of what they would be if the baseline year was 2013 and the performance year was 2014.
There were 1,931 HHAs in the nine selected states out of 1,991 HHAs that were found in the HHA data sources which yielded the sufficient measures to be included in the model. It is expected that a certain number of HHAs will not be subject to the payment adjustment because they may be servicing too small of a population to report on an adequate number of measures to calculate a TPS.
As described in section III.B.2 of this proposed rule, we considered proposing to reduce the national, standardized 60-day episode payment rate by 3.41 percent in CY 2016 to account for nominal case-mix growth between CY 2012 and CY 2014. If we were to reduce the national, standardized 60-day episode payment rate by 3.41 percent, we estimate that the aggregate impact would be a net decrease of $650 million in payments to HHAs, resulting from a $470 million decrease (−2.5 percent) due to the third year of the Affordable Care Act mandated rebasing adjustments, a $420 million increase (2.3 percent) due to the home health payment update percentage, and a $600 million decrease due to reducing the national, standardized 60-day episode payment rate by 3.41 percent. However, instead of proposing a one-time reduction in the national, standardized 60-day episode payment rate of 3.41 percent in CY 2016 to account for nominal case-mix growth from CY 2012 through CY 2014, we proposed to reduce the national, standardized 60-day episode payment rate by 1.72 percent in CY 2016 and 1.72 percent in CY 2017 to account for nominal case-mix growth from CY 2012 through CY 2014 as outlined in section III.B.2 of this proposed rule.
Section 3131(a) of the Affordable Care Act mandates that starting in CY 2014, the Secretary must apply an adjustment to the national, standardized 60-day episode payment rate and other amounts applicable under section 1895(b)(3)(A)(i)(III) of the Act to reflect factors such as changes in the number of visits in an episode, the mix of services in an episode, the level of intensity of services in an episode, the average cost of providing care per episode, and other relevant factors. In addition, section 3131(a) of the Affordable Care Act mandates that rebasing must be phased-in over a 4-year period in equal increments, not to exceed 3.5 percent of the amount (or amounts) as of the date of enactment (2010) under section 1895(b)(3)(A)(i)(III) of the Act, and be fully implemented in CY 2017. Therefore, in the CY 2014 HH PPS final rule (78 FR 77256), we finalized rebasing adjustments to the national, standardized 60-day episode payment amount, the national per-visit rates and the NRS conversion factor. As we noted in the CY 2014 HH PPS final rule, because section 3131(a) of the Affordable Care Act requires a four year phase-in of rebasing, in equal increments, to start in CY 2014 and be fully implemented in CY 2017, we do not have the discretion to delay, change, or eliminate the rebasing adjustments once we have determined that rebasing is necessary (78 FR 72283).
Section 1895(b)(3)(B) of the Act requires that the standard prospective payment amounts for CY 2016 be increased by a factor equal to the applicable HH market basket update for those HHAs that submit quality data as required by the Secretary. For CY 2016, section 3401(e) of the Affordable Care Act, requires that, in CY 2015 (and in subsequent calendar years), the market basket update under the HHA prospective payment system, as described in section 1895(b)(3)(B) of the Act, be annually adjusted by changes in economy-wide productivity. Beginning in CY 2015, section 1895(b)(3)(B)(vi)(I) of the Act, as amended by section 3401(e) of the Affordable Care Act, requires the application of the productivity adjustment described in section 1886(b)(3)(B)(xi)(II) of the Act to the HHA PPS for CY 2015 and each subsequent CY. The −0.6 percentage point productivity adjustment to the
We invite comments on the alternatives discussed in this analysis.
As required by OMB Circular A–4 (available at
Table 28 provides our best estimate of the decrease in Medicare payments under the proposed HHVBP model.
In conclusion, we estimate that the net impact of the HH PPS proposals in this rule is a decrease in Medicare payments to HHAs of $350 million for CY 2016. The $350 million decrease in estimated payments to HHAs for CY 2016 reflects the distributional effects of the 2.3 percent CY 2016 HH payment update percentage ($420 million increase), the proposed reduction to the national, standardized 60-day episode payment rate in CY 2016 of 1.72 percent to account for nominal case-mix growth ($300 million decrease), and the third year of the 4-year phase-in of the rebasing adjustments required by section 3131(a) of the Affordable Care Act of −2.5 percent ($470 million decrease). This analysis, together with the remainder of this preamble, provides an initial Regulatory Flexibility Analysis.
In conclusion, we estimate there will be no net impact of the proposals in this rule in Medicare payments to HHAs for CY 2016. However, the overall economic impact of the HHVBP model provision is an estimated $380 million in total savings from a reduction in unnecessary hospitalizations and SNF usage as a result of greater quality improvements in the HH industry over the life of the proposed model.
Executive Order 13132 on Federalism (August 4, 1999) establishes certain requirements that an agency must meet when it promulgates a final rule that imposes substantial direct requirement costs on state and local governments, preempts state law, or otherwise has Federalism implications. We have reviewed this proposed rule under the threshold criteria of Executive Order 13132, Federalism, and have determined that it will not have substantial direct effects on the rights, roles, and responsibilities of states, local or tribal governments.
Health facilities, Medicare
Emergency medical services, Health facilities, Health professions, Medicare, and Reporting and recordkeeping requirements.
Health facilities, Health professions, Medicare, and Reporting and recordkeeping requirements.
For the reasons set forth in the preamble, the Centers for Medicare & Medicaid Services proposes to amend 42 CFR chapter IV as set forth below:
Secs. 1102 and 1871 of the Social Security Act (42 U.S.C. 1302 and 1395hh).
(e) * * *
(1) * * *
(iii) Discharge with goals met and/or no expectation of a return to home health care and the patient returns to home health care during the 60 day episode.
Secs. 1102 and 1871 of the Social Security Act (42 U.S.C. 1302 and 1395hh).
Secs 1102 and 1871 of the Social Security Act (42 U.S.C. 1302 and 1395(hh)) unless otherwise indicated.
(d)
(2) The PEP adjustment will not apply in situations of transfers among HHAs of common ownership. Those situations will be considered services provided under arrangement on behalf of the originating HHA by the receiving HHA with the common ownership interest for the balance of the 60-day episode. The common ownership exception to the transfer PEP adjustment does not apply if the beneficiary moves to a different MSA or Non-MSA during the 60-day
(3) If the intervening event warrants a new 60-day episode payment and a new physician certification and a new plan of care, the initial HHA receives a partial episode payment adjustment reflecting the length of time the patient remained under its care. A partial episode payment adjustment is determined in accordance with § 484.235.
(e)
(a) * * *
(3) For CY 2011, the adjustment is 3.79 percent.
(4) For CY 2012, the adjustment is 3.79 percent.
(5) For CY 2013, the adjustment is 1.32 percent.
(6) For CY 2016 and CY 2017, the adjustment is 1.72 percent in each year.
(a) CMS updates the unadjusted national 60-day episode payment rate on a fiscal year basis (as defined in section 1895(b)(1)(B) of the Act).
(b) For 2007 and subsequent calendar years, in accordance with section 1895(b)(3)(B)(v) of the Act, in the case of a home health agency that submits home health quality data, as specified by the Secretary, the unadjusted national prospective 60-day episode rate is equal to the rate for the previous calendar year increased by the applicable home health market basket index amount.
(c) For 2007 and subsequent calendar years, in accordance with section 1895(b)(3)(B)(v) of the Act, in the case of a home health agency that does not submit home health quality data, as specified by the Secretary, the unadjusted national prospective 60-day episode rate is equal to the rate for the previous calendar year increased by the applicable home health market basket index amount minus 2 percentage points. Any reduction of the percentage change will apply only to the calendar year involved and will not be taken into account in computing the prospective payment amount for a subsequent calendar year.
(b) The outlier threshold for each case-mix group is the episode payment amount for that group, or the PEP adjustment amount for the episode, plus a fixed dollar loss amount that is the same for all case-mix groups
(e) The fixed dollar loss amount and the loss sharing proportion are chosen so that the estimated total outlier payment is no more than 2.5 percent of total payment under home health PPS.
(f) The total amount of outlier payments to a specific home health agency for a year may not exceed an amount equal to 10 percent of the total payments to the specific agency under home health PPS for the year.
This subpart is established under section 1115A(a)(1) of the Act (42 U.S.C. 1315a), which authorizes the Secretary to test innovative payment and service delivery models to improve coordination, quality, and efficiency of health care services furnished under Title XVIII.
As used in this subpart—
(1) For CY 2018 and 2019, 5 percent.
(2) For CY 2020, 6 percent.
(3) For CY 2021 and 2022, 8 percent.
(1) That has a current Medicare certification; and,
(2) Is being reimbursed by CMS for home health care delivered within any of the states specified in accordance with CMS's selection methodology.
(a)
(b) Nine states are selected in accordance with CMS's selection methodology. All Medicare-certified HHAs that provide services in Massachusetts, Maryland, North Carolina, Florida, Washington, Arizona, Iowa, Nebraska, and Tennessee will be required to compete in this model.
(a) Medicare-certified home health agencies will be evaluated using a starter set of quality measures.
(b) Medicare-certified home health agencies in selected states will be required to report information on New Measures, as determined appropriate by the Secretary, to CMS in the form, manner, and at a time specified by the Secretary.
(c) Medicare-certified home health agencies in selected states will be required to collect and report such information as the Secretary determines is necessary for purposes of monitoring and evaluating the HHVBP model under section 1115A(b)(4) of the Act (42 U.S.C. 1315a).
A Medicare-certified home health agency's Total Performance Score for a model year is calculated as follows:
(a) CMS will award points to the Medicare-certified home health agency for performance on each of the applicable measures in the starter set, other than New Measures.
(b) CMS will award points to the Medicare-certified home health agency for reporting on each of the New Measures in the starter set, worth up to ten percent of the Total Performance Score.
(c) CMS will sum all points awarded for each applicable measure in the starter set, weighted equally at the individual measure level, to calculate a value worth up to 90 percent of the Total Performance Score.
(d) The sum of the points awarded to a Medicare-certified HHA for each applicable measure in the starter set and the points awarded to a Medicare-certified HHA for reporting data on each New Measure is the Medicare-certified HHA's Total Performance Score for the calendar year.
CMS will determine a payment adjustment up to the maximum applicable percentage, upward or downward, under the HHVBP model for each Medicare-certified home health agency based on the agency's Total Performance Score using a linear exchange function. Payment adjustments made under the HHVBP model will be calculated as a percentage of otherwise-applicable payments for home health services provided under section 1895 of the Act (42 U.S.C. 1395fff).
(a)
(b)
(c)