Loading...
HomeMy WebLinkAbout20250306INT to Staff 8 Attachment.pdf DOE/EE-0829 Y O 3 W z z O U Energy Efficiency Program U u Impact Evaluation Guide U pW v Evaluation, Measurement, and Verification Working Group LL u_ W December 2012 W Ujz W J Q UjU 0 (n LU 0 . o 0 Qp 9 W� I o o The State and Local Energy Efficiency Action Network is a state and local effort facilitated by the federal government that helps states, utilities, and other local stakeholders take energy efficiency to scale and achieve all cost-effective energy efficiency by 2020. Learn more at www.seeaction.energy.gov ABOUT THIS DOCUMENT This Energy Efficiency Program Impact Evaluation Guide describes the common terminology,structures,and approaches used for determining(evaluating)energy and demand savings as well as avoided emissions and other non-energy benefits resulting from facility(non-transportation)energy efficiency programs that are implemented by local governments,states,utilities,private companies,and nonprofits.While this guide does not recommend specific approaches,it provides context,planning guidance, and discussion of issues that determine the most appropriate evaluation objectives and best practices approaches for different efficiency portfolios. By using standard evaluation terminology and structures and best practices approaches,evaluations can support the adoption,continuation,and expansion of effective efficiency actions. The primary audiences for this guide are energy regulators;public and private energy efficiency portfolio administrators such as utilities,nonprofit organizations,and government agencies; program implementers;and evaluators looking for guidance on the following: • The evaluation process and approaches for determining program impacts • Planning evaluation efforts • Key issues associated with establishing evaluation frameworks for improving the efficacy of energy efficiency portfolios, documenting the impacts of such portfolios,and comparing demand-and supply-side resources. Introductory portions and appendices are also intended for policymakers seeking general information about efficiency program impact evaluation as well as the basic principles of process and market evaluations and cost-effectiveness analyses.Although the guide is not directly intended for expert evaluation practitioners who can rely on more detailed and specific resources that are referred to in this guide,it offers introductions to and summaries of evaluation topics that can be useful for explaining concepts and standard practices to clients, new staff,stakeholders,and others who could benefit from a refresher on principles. This 2012 version of the guide is an update to the 2007 National Action Plan for Energy Efficiency Model Energy Efficiency Program Impact Evaluation Guide. Prepared by Steve Schiller,Schiller Consulting,Inc.www.epa.gov/eeactionplan. www.seeaction.energy.gov December 2012 The Energy Efficiency Program Impact Evaluation Guide was developed as a product of the State and Local Energy Efficiency(EPA)Action Network(SEE Action),which is facilitated by the U.S. Department of Energy and the U.S.Environmental Protection Agency.Content does not imply an endorsement by the individuals or organizations that are part of SEE Action working groups or reflect the views,policies,or other- wise of the federal government. This effort was funded by the Permitting,Siting,and Analysis Division of the U.S. Department of Energy's Office of Electricity Delivery and Energy Reliability under Lawrence Berkeley National Laboratory Contract No. DE-ACO2-05CH11231. If this document is referenced,it should be cited as: State and Local Energy Efficiency Action Network.2012.Energy Efficiency Program Impact Evaluation Guide. Prepared by Steven R.Schiller,Schiller Consulting, Inc.,www.seeaction.energy.gov. FOR MORE INFORMATION For more information regarding the Energy Efficiency Program Impact Evaluation Guide,please contact: Michael Li Carla Frisch U.S. Department of Energy U.S. Department of Energy Email: Michael.Li@ee.doe.gov Email:Carla.Frisch@ee.doe.gov Regarding the State and Local Energy Efficiency Action Network,please contact: Johanna Zetterberg U.S. Department of Energy Email:Johanna.Zetterberg@ee.doe.gov December 2012 www.seeaction.energy.gov Acknowledgments The following people provided key input to the development of this version of the guide by providing materials,sidebars,and/or extremely valuable review and input on draft versions(*denotes member of SEE Action's EM&V Working Group): • Jeffrey Brown, Robyn DeYoung,and Nikolaas Dietsch,U.S. Environmental Protection Agency • Tom Eckman, Northwest Power Planning Council • Hilary Forster,Consortium for Energy Efficiency* • Donald Gilligan,National Association of Energy Services Companies* • Fred Gordon, Energy Trust of Oregon* • Dennis Hartline, Maryland Energy Administration* • M.Sami Khawaja, Hossein Haeri,Tina Jayaweera,and David Sumi,The Cadmus Group • Stephen Kromer, Kromer Engineering • Jennifer Meissner, New York State Energy Research and Development Authority* • Mike Messenger, Itron • Julie Michals*and Elizabeth Titus, Northeast Energy Efficiency Partnerships • Peter Miller, Natural Resources Defense Council* • Jane Peters, Research Into Action • Mitch Rosenberg, DNV KEMA • Lisa Skumatz,Skumatz Economic Research Associates • Rodney Sobin,Alliance to Save Energy* • Annika Todd, Lawrence Berkeley National Laboratory • Carol Zabin, University of California, Berkeley. Dr. Khawaja provided substantial input into several sections of this guide,as did the EPA staff on the Avoided Emissions Chapter;their input is therefore particularly acknowledged. Mark Wilson provided editing services. Publication and graphics were provided by the National Renewable Energy Laboratory. v www.seeaction.energy.gov December 2012 List of Acronyms EERS:energy efficiency resource standard ACEEE:American Council for an Energy-Efficient Economy EGU:electric generating unit(a power plant) AEA:American Evaluation Association EM&V:evaluation,measurement,and verification ANSI:American National Standards Institute EPA:U.S.Environmental Protection Agency ASHRAE:American Society of Heating, Refrigerating and ER:emission rate Air-Conditioning Engineers ESCO:energy services company ETD:Energy Trust of Oregon EUL:effective useful life BAU:business as usual BM:build margin(for electric generating units) Btu:British thermal units FEMP: Federal Energy Management Program FERC:Federal Energy Regulatory Commission CAIR:Clean Air Interstate Rule CALMAC:California Measurement Advisory Council GHG:greenhouse gas C&S:(efficiency)codes and standards CDD:cooling degree day CDM:(United Nations Framework Convention on Climate Change) HDD:heating degree day Clean Development Mechanism HERS:Home Energy Rating System CEMS:Continuous Emission Monitoring Systems HHV:higher heating value CFL:compact fluorescent light bulb HVAC:heating,ventilation,and air conditioning CMVP:Certified Measurement and Verification Professionals COz:carbon dioxide CPUC:California Public Utility Commission IPCC:Intergovernmental Panel on Climate Change CSA:conditional savings analysis IPM:integrated planning model CV:contingent valuation IPMVP:International Performance Measurement and Verification Protocol Cx:commissioning IRP:integrated resource planning ISO:independent system operator or International Organization for Standardization DEER: Database for Energy Efficiency Resources(California) DOE: U.S.Department of Energy DR:demand response kW:kilowatt DSM:demand-side management kWh: kilowatt-hour E&T:education and training lb: pound ECM:energy conservation measure EE:energy efficiency December 2012 www.seeaction.energy.gov v M&V:measurement and verification RASS:residential appliance saturation studies MARAMA:Mid-Atlantic Regional Air Management Association RCx: retro-commissioning MMBtu:million Btu RCT:randomized controlled trial MT:market transformation RD&D:research,development,and demonstration MW:megawatt REED: Regional Energy Efficiency Database MWh:megawatt-hour RFP:request for proposal RFQ: request for qualifications RGGI: Regional Greenhouse Gas Initiative NAESCO: National Association of Energy Service Companies RIM:ratepayer impact test NEB:non-energy benefit RTF:Regional Technical Forum NEEA: Northwest Energy Efficiency Alliance NEI:non-energy impact S NERC: North American Reliability Corporation SCT:societal cost test NOMAD: naturally occurring market adoption rate SIP:State(air pollution reduction)Implementation Plan NOx:nitrogen oxide SOz:sulfur dioxide NPV:net present value SPM:(California)Standard Practice Manual NTG:net-to-gross SPT:standardized project tracking NTGR:net-to-gross ratio TBE:theory-based evaluation O&M:operations and maintenance T&D:transmission and distribution OM:operating margin(for electric generating units) TRC:total resource cost test TRM:technical reference manual PACT:program administrator cost test PCT:participant cost test UMP:Uniform Methods Project PMP:Performance Measurement Protocols for Commercial Buildings(ASHRAE) PSC:Public Service Commission VCS:Verified Carbon Standard PUC: Public Utilities Commission WBCSD:World Business Council for Sustainable Development QAG:quality assurance guideline WRI:World Resources Institute QA/QC:quality assurance/quality control QEM:quasi-experimental methods vi www.seeaction.energy.gov December 2012 Table of Contents Acknowledgments........................................................................................................................................................................................................iv Listof Acronyms.............................................................................................................................................................................................................v Listof Tables.................................................................................................................................................................................................................. A Listof Figures............................................................................................................................................................................................................... xii ExecutiveSummary.........................................................................................................................................................................................xiii ES.1 Topics Covered by this Guide and Intended Audiences.............................................................................................................................xiii ES.2 Definition and Importance of Energy Efficiency Evaluation.......................................................................................................................xiii ES.3 Impact Evaluation Metrics.......................................................................................................................................................................... xiv ES.4 Energy Savings Estimates and Uncertainty:"How Good is Good Enough?"..............................................................................................xv ES.S Evaluation and the Efficiency Program Process...........................................................................................................................................xv ES.5.1 Planning Impact Evaluations............................................................................................................................................................ xvi ES.5.2 Implementing Impact Evaluations................................................................................................................................................... xvi ES.5.3 Evaluation Planning Characteristics and Frameworks.....................................................................................................................xvii ES.5.4 Evaluation Planning Issues..............................................................................................................................................................xviii Chapter1 Introduction....................................................................................................................................................................................1-1 1.1 Guide Objective............................................................................................................................................................................................1-1 1.2 Subjects Covered in This Guide....................................................................................................................................................................1-1 1.3 Guide Structure and How to Use This Guide...............................................................................................................................................1-2 1.4 Source Documents........................................................................................................................................................................................1-2 Chapter 2 Energy Efficiency Program Evaluation Overview.............................................................................................................................2-1 2.1 Energy Efficiency Program Categories and Definitions................................................................................................................................2-1 2.1.1 Policy Context of Energy Efficiency Programs....................................................................................................................................2-1 2.1.2 Efficiency Program Categories...........................................................................................................................................................2-1 2.1.3 Savings Hierarchy...............................................................................................................................................................................2-2 2.2 Program Evaluation Categories and Impact Evaluation Definitions............................................................................................................2-3 2.2.1 Evaluation Definitions........................................................................................................................................................................2-4 2.3 Objectives and Importance of Evaluation....................................................................................................................................................2-5 Chapter3 Impact Evaluation Basics.................................................................................................................................................................3-1 3.1 Impact Evaluation Process............................................................................................................................................................................3-1 3.1.1 Verification and Documentation........................................................................................................................................................3-1 3.1.2 The Evaluator's Role and Reporting of Savings..................................................................................................................................3-2 3.2 Energy and Non-Energy Benefit Evaluation Metrics....................................................................................................................................3-3 December 2012 www.seeaction.energy.gov vii 3.3 Fundamental Energy Efficiency Impact Evaluation Concepts:The Counterfactual and Managing Uncertainty........................................3-4 3.3.1 The Counterfactual.............................................................................................................................................................................3-4 3.3.2 Baselines.............................................................................................................................................................................................3-4 3.4 Determining Energy and Demand Savings...................................................................................................................................................3-5 3.4.1 Noncontrol Group Impact Evaluation Approaches............................................................................................................................3-5 3.4.2 Control Group Approaches:Large-Scale Consumption Data Analysis (Randomized Controlled Trials and Quasi-Experimental Methods)..................................................................................................3-6 3.4.3 Approaches for Determining Net Energy and Demand Savings........................................................................................................3-7 Chapter4 Calculating Energy Savings..............................................................................................................................................................4-1 4.1 Measurement and Verification Approach....................................................................................................................................................4-1 4.1.1 Measurement and Verification Approaches......................................................................................................................................4-3 4.1.2 M&V Option A: Retrofit Isolation—Key Parameter Measurement...................................................................................................4-5 4.1.3 M&V Option B:Retrofit Isolation—All Parameter Measurement.....................................................................................................4-5 4.1.4 M&V Option C:Whole-Facility Analyses............................................................................................................................................4-6 4.1.5 M&V Option D:Calibrated Simulation...............................................................................................................................................4-6 4.2 Deemed Savings Approach...........................................................................................................................................................................4-7 4.3 Large-Scale Consumption Data Analysis(Randomized Controlled Trials and Quasi-Experimental Methods)............................................4-9 4.3.1 Randomized Controlled Trial(RCT) Methods.....................................................................................................................................4-9 4.3.2 Quasi-Experimental Methods(QEMs).............................................................................................................................................4-10 4.3.3 Analysis Techniques..........................................................................................................................................................................4-11 4.4 Selecting an Energy Savings Evaluation Approach.....................................................................................................................................4-11 4.4.1 M&V Approach.................................................................................................................................................................................4-12 4.4.2 Deemed Savings Approach..............................................................................................................................................................4-12 4.4.3 Large-Scale Consumption Data Analysis Approach.........................................................................................................................4-13 Chapter5 Determining Net Energy Savings.....................................................................................................................................................5-1 5.1 Definitions and Uses of Net Savings.............................................................................................................................................................5-1 5.1.1 Defining Net Savings..........................................................................................................................................................................5-1 5.1.2 Importance and Uses of Net Savings.................................................................................................................................................5-3 5.2 Approaches for Determining Net Savings....................................................................................................................................................5-3 5.2.1 Self-Reporting Surveys.......................................................................................................................................................................5-5 5.2.2 Enhanced Self-Reporting Surveys......................................................................................................................................................5-6 5.2.3 Stipulated Net-to-Gross Ratio............................................................................................................................................................5-7 5.2.4 Selecting a Net Savings Evaluation Approach and Timing.................................................................................................................5-7 5.3 Issues and Nuances of Net Savings..............................................................................................................................................................5-7 5.3.1 Free Ridership....................................................................................................................................................................................5-8 5.3.2 Spillover and Market Effects..............................................................................................................................................................5-8 5.3.3 Baselines and Net Savings..................................................................................................................................................................5-9 5.4 Summary.......................................................................................................................................................................................................5-9 viii www.seeaction.energy.gov December 2012 Chapter 6 Calculating Avoided Air Emissions...................................................................................................................................................6-1 6.1 Energy Efficiency and Avoided Emissions Summary....................................................................................................................................6-1 6.2 Special Considerations for Calculating Avoided Emissions from Efficiency Programs.................................................................................6-2 6.2.1 Additionality.......................................................................................................................................................................................6-2 6.2.2 Assessment Boundary Issues: Primary and Secondary Effects/Direct and Indirect Emissions.........................................................6-3 6.2.3 Special Issues for Capped Pollutants under Cap-and-Trade Programs..............................................................................................6-4 6.2.4 Avoided Emissions Calculations for Different Objectives..................................................................................................................6-5 6.3 Direct Onsite Avoided Emissions Calculations.............................................................................................................................................6-5 6.4 Emission Factors for Grid-Connected Electric Generating Units...................................................................................................................6-6 6.4.1 The Electricity Generation Mix...........................................................................................................................................................6-6 6.4.2 Using Avoided Emission Factors and Scenario Analyses for Electric Grid Analyses..........................................................................6-8 6.5 Selecting an Approach for Calculating Avoided Emissions........................................................................................................................6-14 6.6 Additional Resources on Avoided Emissions Calculations.........................................................................................................................6-14 Chapter7 Impact Evaluation Considerations...................................................................................................................................................7-1 7.1 Selecting Baselines.......................................................................................................................................................................................7-1 7.1.1 Existing Conditions Baselines.............................................................................................................................................................7-1 7.1.2 Codes and Standards Baseline...........................................................................................................................................................7-2 7.1.3 Common Practice Baseline................................................................................................................................................................7-2 7.1.4 Defining Baselines for Specific Program Types..................................................................................................................................7-2 7.2 Determining Demand Savings......................................................................................................................................................................7-4 7.3 Persistence of Savings...................................................................................................................................................................................7-6 7.3.1 Definitions..........................................................................................................................................................................................7-6 7.3.2 Rebound.............................................................................................................................................................................................7-7 7.3.3 Determining Persistence....................................................................................................................................................................7-7 7.4 Controlling Uncertainty................................................................................................................................................................................7-8 7.4.1 Systematic Errors/Bias........................................................................................................................................................................7-9 7.4.2 Random Errors...................................................................................................................................................................................7-9 7.5 Evaluation Budgets: Balancing the Value of Evaluation Results and Uncertainty.....................................................................................7-11 7.5.1 Using Impact Evaluations to Manage Risk.......................................................................................................................................7-11 7.5.2 Budget-Setting Considerations.........................................................................................................................................................7-12 7.5.3 State Evaluation Budgets..................................................................................................................................................................7-14 7.6 Evaluation Principles...................................................................................................................................................................................7-14 7.7 Using Impact Evaluations for Program Feedback.......................................................................................................................................7-18 7.8 Evaluation for Demand Forecasting and Energy Resource Planning.........................................................................................................7-19 7.9 Determining Non-Energy Benefits.............................................................................................................................................................7-20 7.10 Impact Evaluation for Unique Energy Efficiency Program Types.............................................................................................................7-24 7.10.1 Residential Behavior-Based Programs...........................................................................................................................................7-24 7.10.2 Education and Training(E&T) Programs........................................................................................................................................7-25 December 2012 www.seeaction.energy.gov x 7.10.3 Market Transformation(MT)Programs.........................................................................................................................................7-25 7.10.4 Codes and Standards Programs.....................................................................................................................................................7-26 7.10.5 Demand Response(DR)Programs.................................................................................................................................................7-28 7.10.6 Greenhouse Gas(GHG)Mitigation Programs................................................................................................................................7-28 Chapter8 Impact Evaluation Planning.............................................................................................................................................................8-1 8.1 Integration of Evaluation into the Portfolio Implementation Cycle.............................................................................................................8-1 8.2 Hierarchy of Evaluation Planning and Reporting Documents......................................................................................................................8-3 8.2.1 Evaluation Planning Documents........................................................................................................................................................8-3 8.2.2 Evaluation Reports.............................................................................................................................................................................8-4 8.3 Principles and Issues That Determine the Scope of an Impact Evaluation..................................................................................................8-5 8.3.1 Defining Evaluation Objectives and Metrics......................................................................................................................................8-6 8.3.2 Evaluator Roles and Selection............................................................................................................................................................8-8 8.3.3 Setting the Boundary and Granularity of Reported Results............................................................................................................8-10 8.3.4 Schedule for Evaluation and Reporting............................................................................................................................................8-11 8.3.5 How Are Evaluated Savings Estimates Applied—Looking Back Or Going Forward?.......................................................................8-11 8.3.6 Data Management Strategies..........................................................................................................................................................8-12 8.3.7 Addressing Disputes.........................................................................................................................................................................8-13 8.4 Portfolio Cycle Impact Evaluation and M&V Plan Outlines........................................................................................................................8-13 8.4.1 Portfolio Cycle Impact Evaluation Plan and Report Outlines...........................................................................................................8-13 8.4.2 Project-Specific M&V Plan and Report Outlines..............................................................................................................................8-15 8.4.3 Checklist of Planning Decisions for an Impact Evaluation...............................................................................................................8-15 AppendixA:Glossary.....................................................................................................................................................................................A-1 Appendix B:Other Evaluation Categories and Approaches............................................................................................................................B-1 B.1 Process,Market Effects,and Cost-Effectiveness Evaluations...................................................................................................................... B-1 B.1.1 Process Evaluations........................................................................................................................................................................... B-1 B.1.2 Market Effects Evaluations................................................................................................................................................................ B-3 B.1.3 Cost-Effectiveness Analyses.............................................................................................................................................................. B-4 B.2 Top-Down Impact Evaluation...................................................................................................................................................................... B-9 AppendixC:Resources...................................................................................................................................................................................C-1 CABackground...................................................................................................................................................................................................C-1 C.2 Primary Impact Evaluation Resources..........................................................................................................................................................C-1 C.2.1 General Efficiency Evaluation Resource Websites.............................................................................................................................C-1 C.2.2 Select Impact Evaluation Resources..................................................................................................................................................C-1 C.2.3 U.S.Environmental Protection Agency Non-Energy Benefits and Avoided Emissions Calculation References...............................C-2 C.2.4Technical Reference Manual(TRMs)Resources................................................................................................................................C-3 C.3 Measurement and Verification Resources...................................................................................................................................................C-5 References.....................................................................................................................................................................................................D-1 x www.seeaction.energy.gov December 2012 List of Tables Table 1.1:Summary of Guide Content and Intended Audience for Each Part of the Guide.............................................................................1-3 Table 2.1:Summary of Evaluation Categories and Types..................................................................................................................................2-4 Table4.1:IPMVP Options Summary..................................................................................................................................................................4-4 Table 4.2:Applications for Each IPMVP M&V Option......................................................................................................................................4-14 Table 4.3:Factors Affecting the Cost of Each M&V Option..............................................................................................................................4-14 Table 4.4:Key Project Characteristics and Applicability for Different M&V Options.......................................................................................4-15 Table 5.1:Audiences for Net Energy Savings.....................................................................................................................................................5-4 Table 5.2:Example of Assigning Free Ridership Percentages............................................................................................................................5-6 Table 6.1:Choosing an Avoided Grid Electricity Emissions Quantification Approach.......................................................................................6-9 Table 7.1:Standard Practices for Selection of Baselines for Common Program Categories.............................................................................7-3 Table 7.2:Electric and Gas EM&V Expenditure and Budget Dollars,United States and Canada(millions USD)*..........................................7-15 Table 7.3:Ratepayer-Funded Energy Efficiency Budgets and EM&V Expenditures for Selected States.........................................................7-16 Table 7.4:EM&V 2008 Funding Levels and Allocation Among Activities........................................................................................................7-17 Table 8.1:Maryland Public Service Commission EM&V,Data Tracking,and Reporting Act Roles and Responsibilities for Utilities and Statewide Evaluation Consultants........................................................................................8-16 Table 8.2:Energy Efficiency M&V Plan Contents:General Components.........................................................................................................8-20 Table 8.3:Energy Efficiency Project-Specific and Measure-Specific M&V Plan Contents...............................................................................8-21 Table 8.4:Evaluation Planning Checklist .........................................................................................................................................................8-22 TableB.1:Program Evaluation Types................................................................................................................................................................. B-1 TableB.2:Elements of Typical Process Evaluations.......................................................................................................................................... B-2 Table B.3:Approaches for Assessing Market Effects,Including Attribution .................................................................................................... B-5 Table B.4:The Five Principal Cost-Effectiveness Tests Used in Energy Efficiency............................................................................................. B-7 Table B.S:Description of Benefits and Costs Included in Each Cost-Effectiveness Test.................................................................................... B-8 Table B.6:Summary of Benefits and Costs Included in Each Cost-Effectiveness Test....................................................................................... B-9 Table CA:United States Technical Reference Manuals......................................................................................................................................C-3 December 2012 www.seeaction.energy.gov xi List of Figures FigureESA:Evaluation objectives ..................................................................................................................................................................... xiv Figure ES.2:Workflow and reporting for planning,implementing,and evaluating efficiency programs .........................................................xv Figure ES.3:Evaluation activities workflow........................................................................................................................................................ xvi Figure ESA:Evaluation is integral to a typical cyclic planning-implementation-evaluation process...............................................................xviii Figure ES.5:Hierarchy of EM&V planning documents........................................................................................................................................xix Figure 2.1:Hierarchy of energy efficiency activities..........................................................................................................................................2-2 Figure 2.2:Continuum of energy efficiency actions...........................................................................................................................................2-6 Figure 3.1:True program savings:the counterfactual for a residential household energy efficiency program...............................................3-4 Figure 3.2:Energy consumption before,during,and after a project is implemented......................................................................................3-5 Figure 4.1:M&V option selection flowchart....................................................................................................................................................4-16 Figure 5.1:Single program evaluation components..........................................................................................................................................5-2 Figure 5.2:Multiple program evaluation overlaps.............................................................................................................................................5-2 Figure 6.1:Example electricity load duration curve...........................................................................................................................................6-7 Figure6.2:U.S.EPA eGrid subregions..............................................................................................................................................................6-10 Figure 7.1: Incremental value of information versus incremental cost of evaluation ....................................................................................7-13 Figure 7.2:Components of C&S savings estimate............................................................................................................................................7-28 Figure 8.1:Program implementation cycle with high-level evaluation activities..............................................................................................8-2 Figure B.1:Comparison of bottom-up versus top-down evaluation.............................................................................................................. B-10 Figure B.2:Graphical illustration of estimation of market gross savings........................................................................................................ B-10 xii www.seeaction.energy.gov December 2012 Executive Summary ES.1 TOPICS COVERED BY THIS GUIDE Documents from the U.S.Department of Energy's(DOE'S)Uniform AND INTENDED AUDIENCES Methods Project(UMP)serve as a companion set to this guide and This Energy Efficiency Program Impact Evaluation Guide describes and include model evaluation plans for specific energy efficiency measures provides guidance on approaches for determining and documenting and program categories(e.g.,residential lighting,refrigerators, energy and non-energy benefits resulting from end-use energy commercial cooling).' efficiency programs and portfolios of programs. It specifically focuses ES.2 DEFINITION AND IMPORTANCE OF on impact evaluations for programs designed to reduce facility(e.g., ENERGY EFFICIENCY EVALUATION home,commercial building,factory)energy consumption and/or demand as well as related air emissions.This guide's objective is to Energy efficiency evaluation includes any of a range of assessment support the implementation of effective energy efficiency actions by studies and other activities aimed at determining the effects of providing information on standard procedures and best practices for an energy efficiency program. Evaluations can document program planning and conducting evaluations and reporting results.To this performance,operations,changes in energy efficiency markets,and end,this guide accomplishes the following: cost-effectiveness.There are three broad categories of efficiency • Defines a systematic evaluation planning and program evaluations:impact evaluations,process evaluations,and implementation process market evaluations.Although this guide focuses on impact evalua- tions,it is helpful to know the purposes and goals of all three: • Describes several standard approaches for determining • Impact evaluations:assessments that determine and document energy and demand savings(as well as avoided emissions the direct and indirect benefits of an energy efficiency program. and other non-energy impacts) Impact evaluation involves real-time and/or retrospective • Defines key terms related to energy efficiency evaluation assessments of the performance and implementation of an • Provides guidance on key evaluation issues efficiency program or portfolio of programs.Program benefits, • Lists publicly available energy efficiency evaluation resources. or impacts,can include energy and demand savings and non- energy benefits(sometimes called co-benefits,with examples The programs primarily addressed in this guide are voluntary;that being avoided emissions,health benefits,job creation and local is,program participants choose to take the efficiency actions as a economic development,energy security,transmission and result of some form of inducement.This guide does not focus on, distribution benefits,and water savings).Impact evaluations also but does touch on,evaluating mandatory requirements for efficiency support cost-effectiveness analyses aimed at identifying relative such as found in codes and standards.Similarly,the guide only briefly program costs and benefits of energy efficiency as compared to other addresses evaluating programs for which energy savings are an energy resources,including both demand-and supply-side options. indirect benefit,such as contractor training programs. Process evaluations:formative,systematic assessments of an energy efficiency program.They document program operations The audiences for this guide are program designers,implementers, and identify and recommend improvements that are likely to administrators,evaluators,and public agency officials who oversee increase the program's efficiency or effectiveness for acquiring and implement energy efficiency programs. Introductory portions of energy efficiency resources, preferably while maintaining high this guide are intended for policymakers seeking information about the basic principles of impact evaluation.Those looking for just the levels of participant satisfaction. basics may want to read only through Chapter 3 and refer to the Market evaluations:assessments of structure or functioning of appendices for overviews of other evaluation types,definitions,and a market,the behavior of market participants,and/or market references.Some readers who are new to evaluation assignments changes that result from one or more program efforts. Market can benefit from reading the entire document,while others may evaluation studies may include estimates of the current market benefit from focusing on the evaluation issues and planning chapters role of energy efficiency(market baselines),as well as the (Chapters 7 and 8,respectively)and using the rest of the document potential role of efficiency in a local,state, regional,or national as a reference.Although the guide is not intended for expert evalu- market(potential studies). Market evaluation studies indicate ation practitioners,they may find it useful for explaining evaluation how the overall supply chain and market for energy efficiency prod- concepts to those without their expertise. ucts works and how they have been affected by a program(s). December 2012 www.seeaction.energy.gov xiii Market evaluations are critical for,but not exclusively used for, Many energy efficiency evaluations are oriented toward developing programs with market transformation elements and objectives. retrospective estimates of energy savings attributable to a program Examples of market evaluations are potential studies,baselines to demonstrate in regulatory proceedings that public or energy studies,and market effects studies. consumer funds were properly and effectively spent. Beyond documenting savings and attribution,though,is the role of evaluation Evaluations have three primary objectives,as shown in Figure ESA: in improving programs and providing a basis for future savings • Document the benefits(i.e.,impacts)of a program and determine estimates in resource plans.Therefore,evaluation both fosters whether the subject program(or portfolio of programs)met its goals more effective programs and justifies increased levels of investment • Identify ways to improve current and future programs through in energy efficiency as a long-term,reliable energy resource. Perhaps determining why program-induced impacts occurred the imperative for conducting evaluation is best described by a quote • Support energy demand forecasting and resource planning by attributed to John Kenneth Galbraith:"Things that are measured understanding the historical and future resource contributions tend to improve:'z of energy efficiency as compared to other energy resources. ES.3 IMPACT EVALUATION METRICS One or more of the following three metrics are usually reported as FIGURE ESA: Evaluation objectives the output of impact evaluations: • Estimates of gross(energy and/or demand)savings.These are the changes in energy consumption and/or demand that result directly from program-related actions taken by participants in an efficiency program,regardless of why they participated. • Estimates of net(energy and/or demand)savings.These are the changes in energy consumption or demand that are attributable to an energy efficiency program.The primary,but UNDERSTANDnot exclusive,considerations that account for the difference DOCUMENT • - • between net and gross savings are free riders(i.e.,those who IMPACTS PROGRAM I would have implemented the same or similar efficiency projects, PERFORMANCE to one degree or another,without the program now or in the near future)and participant and non-participant spillover (i.e.,savings that result from actions taken as a result of a SUPPORTIENERGY programs's influence but which are not directly subsidized or RESOURCErequired by the program).Net savings may also include consider- ation of market effects(changes in the structure of a market). Determining net savings involves separating out the impacts that are a result of influences other than the program being evaluated,such as consumer self-motivation or effects of prior and/or other programs.Given the range of influences on consumers'energy consumption and the complexity in separat- EVALUATION SUPPORTS SUCCESSFUL ing out both short-term and long-term market effects caused by EFFICIENCY PROGRAMS the subject programs(and other programs),attributing changes Documenting the benefits of efficiency,using credible and to one cause(i.e.,a particular program)or another can be quite transparent methods,is a key component of successfully complex.This is compounded by a general lack of consensus implementing and expanding the role of efficiency in providing among policymakers and regulators on which short-term and secure,stable,reliable,clean,and reasonably priced energy. long-term market influences and effects should be considered Therefore,evaluation is not an end unto itself but an effective when determining net savings. Net savings are discussed in tool for supporting the adoption,continuation,and expansion of Chapter 5. energy efficiency programs,and thus the efficient use of energy. xiv www.seeaction.energy.gov December 2012 • Estimates of non-energy benefits(NEBs).These are the impacts Two principles are important when considering"how good is good associated with program implementation or participation enough":(1)energy efficiency investments should be cost effective, aside from energy and demand savings.These results can be and(2)evaluation investments should consider risk management positive or negative.Some examples include reduced emissions principles and thus balance the costs of evaluation against the value and environmental benefits,productivity improvements,jobs of the information derived from evaluation(i.e.,evaluation should created and local economic development, reduced utility also be cost effective).The value of the information is directly related customer disconnects,greater comfort for building occupants, to the risks of underestimating or overestimating the benefits(savings) lower maintenance costs due to better equipment,or increased and costs associated with efficiency investments.These risks might maintenance costs due to new and more complex systems. be associated with errors of commission or errors of omission.An NEBs are discussed in Section 7.9. error of commission might be overestimating savings,which in turn can result in continuing programs that are not cost effective and/or ESA ENERGY SAVINGS ESTIMATES overpaying contractors,administrators,and participants.An error of AND UNCERTAINTY: "HOW GOOD IS omission,on the other hand,might be associated with underestimating GOOD ENOUGH?" savings or not implementing efficiency actions because of the difficulty Each of the bullets in Section ES.3 above defines an"estimate" in documenting savings,both of which can result in underinvesting in versus an exact value.This is because energy and demand savings as efficiency and relying on other energy resources that have their own well as non-energy benefits resulting from efficiency actions cannot risks and uncertainties,such as fuel costs and environmental impacts. be directly measured. Instead,savings and benefits are based on counterfactual assumptions.Using counterfactual assumptions ES.5 EVALUATION AND THE EFFICIENCY implies that savings are estimated to varying degrees of accuracy PROGRAM PROCESS by comparing the situation (e.g.,energy consumption) after a As shown in Figure ES.2,the efficiency program process consists of program is implemented(the reporting period)to what is assumed planning,implementing,and evaluating activities.Throughout this to have been the situation in the absence of the program(the process,savings values are typically indicated based on estimates "counterfactual"scenario,known as the baseline).For energy impacts, prepared as part of each activity.One way to describe these savings the baseline and reporting period energy use are compared,while is with the following classifications,also displayed in Figure ES.2: controlling(making adjustments)for factors unrelated to energy Projected savings:values reported by a program implementer efficiency actions,such as weather or building occupancy.These or administrator before the efficiency activities are completed adjustments are a major part of the evaluation process; how they are determined can vary from one program type to another and from one evaluation approach to another. FIGURE ES.2: Workflow and reporting for planning, Because the indicated values are estimates,their use as a basis implementing, and evaluating efficiency programs for decision making can be challenged if their sources and level of accuracy are not described.Therefore,evaluation results, like any estimate,should be reported as"expected values";that is, PLAN PROJECTED based on the impact evaluation,values are expected to be correct PROGRAMS within an associated level of certainty. Minimizing uncertainty and balancing evaluation costs with the value of the indicated evaluation Ilk information are at the heart of the evaluation process and leads to perhaps the most fundamental evaluation question:"How good is IMPLEMENT CLAIMED good enough?"This question is a short version of asking(1)what PF PROGRAMS level of certainty is required for energy savings estimates resulting from evaluation activities,and(2)is that level of certainty properly Ilk balanced against the amount of effort(e.g., resources,time,money) used to obtain that level of certainty? EVALUATE EVALUATED PROGRAMS SAVINGS December 2012 www.seeaction.energy.gov xv • Claimed savings:values reported by a program implementer or ES.5.2 Implementing Impact Evaluations administrator after the efficiency activities have been completed The impact evaluation is conducted through the following steps: • Evaluated savings:values reported by an independent third- 1. Verify actual implementation of the program,for example, party evaluator after the efficiency activities and impact evaluation by confirming installation and proper operation of the energy have been completed.The designation of"independent"and efficiency measures.This usually also includes auditing and "third-party" is determined by those entities involved in the validating assumptions used in the program planning process use of the evaluations and may include evaluators retained,for and checking program tracking databases,project applications, example, by the program administrator or a regulator. and other documentation and related data records for accurate With respect to the evaluation activities,they can also be described recording of information. as consisting of three phases:planning,implementation,and report- 2• Determine first-year program energy(and demand)savings ing,as shown in Figure ES.3 and described in the next subsections. using one of the following approaches(which are further defined and described in Chapters 3 and 4): ES.5.1 Planning Impact Evaluations a. Measurement and verification(M&V):a project-by-project The following provide the basic steps in planning impact evaluations: approach involving estimating energy and/or demand savings by 1. Define the evaluation objectives and metrics in the context of determining the savings for a representative sample of projects the evaluated program's(or portfolio's)intended benefits,risks, and applying these projects'savings to the entire population and policy objectives. (i.e.,the program).Options for conducting M&V are defined in 2. Select appropriate evaluation approach(es)and prepare a the International Performance Measurement and Verification program evaluation plan that takes into account the critical Protocol(IPMVP)and include two end-use metering options, evaluation issues and the expectation for reliability(certainty) billing regression analysis,and computer simulation.This of evaluated impacts. approach determines gross savings values;net savings can 3. Define data collection requirements. be determined with program-wide adjustments to the gross savings values. FIGURE ES.3: Evaluation activities workflow PLANNING OBJECTIVES AND METRICS VERIFICATION EVALUATED SAVINGS EVALUATION DETERMINE FIRST-YEAR APPROACHES GROSS AND/OR NET SAVINGS DATA DETERMINE LIFETIME COLLECTION SAVINGS DETERMINE NON-ENERGY BENEFITS DETERMINE COST EFFECTIVENESS xvi www.seeaction.energy.gov December 2012 b.Deemed savings values:stipulations based on historical 6. Determine the program's cost-effectiveness using one or more and verified data(in some cases using the results of prior M&V of the common cost-effectiveness tests. Inputs into these tests studies).Similarly,deemed savings calculations are standardized are the lifecycle net or gross energy and demand savings and algorithms. Both deemed savings values and deemed savings possibly one or more non-energy benefits.See Appendix B for calculations should only be used with well-defined energy effi- an overview of cost-effectiveness analyses. ciency measures that have documented and consistent savings values.This approach determines gross savings values or net The evaluation approaches described in this guide are often referred savings values,if net-to-gross ratios are included in the deemed to as"bottom-up"approaches because they add up the savings savings values or calculations. from measures and projects to determine program impacts,and C.Large-scale consumption data analysis:uses metered energy they add up the impacts of programs to determine total portfolio use data to compare the energy use of the program participants impacts.Another evaluation category,called"top-down,"uses with the energy use of a control group.The control group can be approaches that rely on energy consumption data or per-unit energy either program nonparticipants,as is the case with randomized consumption indicators(e.g.,energy consumption per-unit of output controlled trials,or participants,as is the case with some quasi- or per person)defined by market sector, utility service territory,or experimental methods. If the program participants are used, a geographic region(e.g.,a state or region).Top-down evaluation their energy use before the program and after the program are is not commonly used for evaluation of efficiency programs and compared;in effect,this means that each participant is his/ portfolios,although interest in the approach is growing,and it has her own non random control group.All of these methods can advantages over bottom-up evaluations.A section of Appendix B provide results that are either gross or net savings values. covers top-down evaluation. In some cases,the three approaches listed above are combined, E$.5.3 Evaluation Planning Characteristics particularly the deemed savings and M&V approaches.Portfolios and Frameworks of programs also often use different approaches for different The following are best practice characteristics for evaluations: programs to determine total portfolio savings.Multiple-year Evaluation is integral to a typical cyclic planning-implementa- programs may also conduct detailed measurement-based studies tion-evaluation process.Therefore,evaluation planning is part (e.g.,M&V)for one year of the program and then apply the of the program planning process,including the alignment of savings values(deemed savings)for other program years. implementation and evaluation budgets and schedules.This is 3. Convert,as needed,first-year gross program energy(and done so that evaluation efforts can support program imple- demand)savings to first-year net program savings using a range mentation and provide timely evaluation results for improving of possible considerations as discussed in Chapters 3 and 5. existing programs and informing future program and energy 4. Determine lifetime savings,which are the expected energy resource planning.See Figure ES.4. (and demand)savings over the lifetime of the measures that The evaluation process is designed to support the policy goals are implemented in the efficiency program.These savings are of the energy efficiency programs being evaluated by providing usually calculated by multiplying the first-year annual energy appropriate documentation of progress toward the goals, use reduction associated with the subject measures by the as well as feedback required by program administrators and expected life of these measures with possible consideration implementers to continuously improve the programs and plan of factors such as performance degradation or in some cases future efforts. consideration of rebound(an increased level of service that Evaluation budgets and resources are adequate to support, is accompanied by an increase in energy use as a result of a over the entire evaluation,the evaluation goals and the level program).Section 7.3 discusses savings persistence. of quality(certainty)expected in the evaluation results. 5. Determine non-energy benefits(NEBs)using a range of Reported values for metrics are those that are"most likely" subjective and objective analytical tools. Determining avoided and not biased to be overly conservative or overly aggressive. emissions,which is the primary NEB addressed in this guide, Evaluations use the planning and implementation structure is discussed in Chapter 6. Evaluating other NEBs is discussed described in this guide,as well as the definitions provided for in Section 7.9. evaluation terms. • Energy and demand savings calculations follow one or more of the approaches defined in this guide. December 2012 www.seeaction.energy.gov xvii FIGURE ESA Evaluation is integral to a typical Portfolio cycle EM&V plan.This plan indicates the major evalua- cyclic planning-implementation-evaluation process tion activities that will be conducted during the evaluation cycle (typically one,two,or three years). It includes the budget and allocation among the programs,measures,and market sectors, as applicable. 14j.4 Evaluation activity-specific detailed plans.Evaluation plans Z ,� are created for each of the major evaluation activities(typically k, the evaluation of an energy efficiency program but may include studies such as market assessments)in a given cycle prior to 0 the time each activity is launched. • Project-specific plans. Project-specific plans may be required a N for custom project sites that are analyzed and inspected. Also complementary to this hierarchy of planning documents is a reporting structure that can include individual site evaluation reports,program reports,and annual portfolio reports. Another typical resource document for large-scale efficiency portfolios 4o1 (such as those for a state or regional consumer-funded efficiency 44(14TE PR0GRAMS program)is a technical reference manual(TRM).A TRM is a database of standardized,state-or region-specific deemed savings calculations and associated deemed savings values for well-documented energy efficiency measures.Energy efficiency program administrators and • Evaluations are complete,readable,fair,accurate,transparently implementation contractors use TRMs to reduce evaluation costs documented, relevant,and actionable,as well as balanced and uncertainty. between certainty of results and costs to achieve the results. ES.5.4 Evaluation Planning Issues They also follow the American Evaluation Association's guiding The evaluation requirements described in each of the planning principles,which are listed in Section 7.6. documents listed above are determined by the program objectives, With the above characteristics in mind,individual entities can regulatory mandates(if any),expectations for quality(i.e.,reliability) define their own policy-specific program evaluation requirements. of the evaluation results,available budgets,timing of reporting dead- Jurisdictions such as states can establish and document their evalua- lines,intended uses of the evaluation results,and other factors that lion requirements in a hierarchy of documents.A useful structure of can vary across jurisdictions and programs.In this guide(Chapter 8), planning documents includes the following(see Figure ES.S): 14 key evaluation planning issues are presented and discussed to help define policy-specific program evaluation requirements: • Evaluation framework.A framework is a primary docu- 1. What are the policy and/or regulatory goals that are the basis ment that lays out evaluation principles,metrics,allowable for the efficiency programs,and what are the evaluation approaches,definitions,and metrics for determination of gross objectives, metrics,and research issues that support the and/or net savings,reporting requirements,schedules,and program policies and/or regulations? the roles and responsibilities of various entities.An evaluation 2. What are the evaluation principles that drive the effort? framework document tends to be"fixed"for several years, 3. What is the scale and budget of the evaluation effort? but of course can be updated periodically. It often sets the expectations for the content and scope of the other evaluation 4. Who will conduct the evaluations,how is an independent documents.This is perhaps the principle document that all evaluation defined,and what are the relative EM&V roles stakeholders can focus on and provide high-level input to—the between implementers,evaluators, regulators,stakeholders, and others? "forest versus the trees"of evaluation planning. xviii www.seeaction.energy.gov December 2012 5. Is performance determined on the basis of net or gross savings?What factors are included in defining net savings? 6. What are the baselines against which savings are determined? 7. What is the reporting"boundary"?Are transmission and distribution(T&D)losses included,and how"granular"will the results be? 8. What are the schedules for implementing the evaluation and reporting? 9. What impact evaluation approaches will be used? 10. What are expectations for savings determination certainty (confidence and precision)? 11. Which cost-effectiveness tests will be used? 12. How are evaluated savings estimates applied—looking back/going forward? 13. What are the data management strategies? 14. How are disputes addressed? FIGURE ES.S: Hierarchy of EM&U planning documents TIMEFRAME COVERAGE Multiple EM&V Region, State, Year FRAMEWORK or Program Administrator Annual or PORTFOLIO Region, State, Multiple Year CYCLE EM&V or Program PLAN Administrator As Required EVALUATION Program (e.g., annual) RESEARCH PLAN or Portfolio As Required Project (e.g., annual) or Site December 2012 www.seeaction.energy.gov xix Executive Summary: Notes 1"Uniform Methods Project."(2012). U.S.Department of Energy. wwwl.eere.energy.gov/deployment/ump.html. 2 Although,as discussed in this guide,this sentiment needs to be tempered with a quote that some attribute to Albert Einstein: "Not everything that can be counted counts,and not everything that counts can be counted." xx www.seeaction.energy.gov December 2012 Chapter 1 Introduction 1.1 GUIDE OBJECTIVE EXPECTATIONS AFTER READING Jurisdictions and organizations(e.g.,state agencies,regulatory bodies, THIS GUIDE utilities,efficiency portfolio administrators)can use this guide as both a primer on efficiency impact evaluation and for defining their After reading this guide,the reader will be able to define own institution-specific,general evaluation requirements as well the basic objectives,structure,and evaluation approaches as specific impact evaluation requirements.While each jurisdiction or that can be used to plan and conduct impact evaluations of entity will need to define its own evaluation requirements,this guide efficiency programs.The reader will also be able to support provides a structure,a set of evaluation approaches,suggestions on and provide input for an energy efficiency evaluation framework key evaluation issues,and definitions that can be applied to a variety (general guidance documents)and review impact evaluation of policy situations. plans and reports. Applying the information in this guide can be particularly helpful This guide provides the following: for jurisdictions and organizations just starting or ramping up their efficiency and evaluation activities. By using standard approaches • Policy-neutrals descriptions and guidance for planning and and terminology,developed through 30-plus years of efficiency conducting impact evaluations of end-use efficiency programs6 program evaluation experience,costs for starting up an evaluation to determine energy and demand savings effort and moving"up the learning curve"can be reduced.Use of • Information on determining energy and demand savings, common approaches and terminology can also support comparison as well as avoided emissions that result from energy of efficiency programs in different jurisdictions and facilitate the efficiency programs implementation of"cross-border"energy efficiency and/or green- . Discussions about issues encountered with planning and house gas and other air emissions mitigation programs. implementing impact evaluations 1.2 SUBJECTS COVERED IN THIS GUIDE • A planning process for impact evaluations including a recommended hierarchy of documents and evaluation reports This 2012 guide is an update to the 2007 National Action Plan for • Background on other types of energy efficiency evaluations Energy Efficiency Model Energy Efficiency Program Impact Evaluation Guide.3 It includes new and updated material based on feedback • A glossary of evaluation terms received on the 2007 guide and lessons learned from impact evaluations • A list of other reference documents and resources on energy conducted during the last five years. efficiency evaluation. This guide focuses on bottom-up evaluations4 of the impacts—pri- In practical terms,evaluation planners can use this guide to do marily energy,demand,and emissions savings—of energy efficiency the following: programs implemented in facilities,and for which energy and demand . Define the questions and hypotheses that the evaluation effort savings are the primary objectives.Therefore,the guide helps users is intended to address determine the end-use fossil fuel(e.g.,natural gas)and electricity a Identify appropriate evaluation approaches and methods savings from programs that encourage lighting,space conditioning, that,from a budgetary perspective,balance the value of the process approaches,and similar energy efficiency strategies in resi- information provided by impact evaluations with the costs to dential,commercial,institutional,and industrial facilities.In addition, provide such information at an acceptable level of accuracy while not a focus of this guide,some guidance is provided in Chapter 7 on documenting non-energy benefits and evaluating market transfor- ' Set realistic expectations among the evaluation process mation,behavior,training,and behavior-based programs.Appendix stakeholders regarding the nature and practical value of results B also has sections on market effects and process evaluations,top-down to be delivered,the timing of when evaluation results can be evaluations,and cost-effectiveness analyses. available,and the expected quality of quantitative estimates of program impacts • Set appropriate schedules and budgets that reflect the desired level of certainty expected in the results. December 2012 www.seeaction.energy.gov 1-1 •PLANNING ISSUES Members of the energy efficiency community looking for the following: While reading this guide's first seven chapters,keep in — Common terminology and definitions mind the 14"evaluation planning"issues listed in — A central reference that provides guidance and also lists the Executive Summary and addressed in Chapter 8 with publicly available best practices resources respect to preparing an evaluation plan. — An understanding of the mechanisms for determining the potential value of energy efficiency as an emissions avoid- ance strategy It is also important to indicate what the guide does not cover: Expert evaluation practitioners looking to provide introductions • It is not sufficiently detailed to be the only resource for plan- and summaries of evaluation topics to those who do not have ning or conducting evaluations of specific programs. Rather, their expertise. the guide provides high-level guidance,identifies issues,and 1.4 SOURCE DOCUMENTS directs users to resources for defining policy and program- specific requirements and details. For example,it does not The information in this document is a summary of definitions, describe specific data collection and analysis options,although approaches,and best practices developed during more than Appendix C does list documents where this information can be 30 years of energy efficiency program implementation and evaluation. found for various program types and technologies.? This experience and expertise is documented in numerous guides, • It is not a guide describing how to perform feasibility studies protocols,papers,and reports.More information on these documents or potential studies,which are intended to assess potential and other evaluation resources is included in footnoted references savings and benefits from future energy efficiency projects or throughout the guide and in Appendix C. programs, respectively. Instead,this guide can be used to help define and conduct studies that inform on what has been or is being accomplished with existing programs. USING THIS GUIDE 1.3 GUIDE STRUCTURE AND HOW TO USE Policymakers and those looking for the"basics": Read the THIS GUIDE Executive Summary and first three chapters and refer to the Table 1.1 at the end of this chapter provides a summary of the guide appendices for overviews of other evaluation types,defini- contents and suggestions for which chapters different audiences will tions,and references. find of interest. Experienced evaluation planners:Go straight to the evaluation This guide's intended audience includes the following: considerations chapter(Chapter 7)and the planning chapter • Program and evaluation managers looking for basic guidance— (Chapter 8)and use the rest of the document as a reference. or a "road map"—on approaches and key issues including: Readers new to evaluation or energy efficiency:Read the — Defining and documenting net and gross energy and entire document. demand savings — Documenting avoided emissions — Comparing demand-and supply-side resources • Energy system resource planners and demand forecasters looking for how end-use efficiency impact evaluation strategies and results can be effectively used in resource planning efforts • Program designers looking to understand how their programs will be evaluated and the benefits they can receive from evaluations • Policymakers and regulators looking for a basic understanding of evaluation objectives,processes,and issues 1-2 www.seeaction.energy.gov December 2012 TABLE 1.1: Summary of Guide Content and Intended Audience for Each Part of the Guide INTENDEDPART CHAPTER . Part 1 Executive Summary Readers interested in a brief The Executive Summary provides an overview of summary and introduction to impact evaluation,with discussion of the importance impact evaluation and types of evaluations,the impact evaluation process,key issues,and evaluation planning. Part 2 Chapter 1: Introduction Readers who want an overview The Introduction describes the guide's objective, of evaluation and the key a review of what is and is not covered in the guide, Chapter 2: Energy aspects of impact evaluation and recommendations for how to use the guide. Efficiency Program Chapter 2 describes the objectives of evaluation Evaluation Overview and provides definitions of different efficiency Chapter 3: Impact program types and evaluation categories.Chapter 3 Evaluation Basics summarizes the basics,processes,and approaches associated with impact evaluation. Part 3 Chapter 4:Calculating Readers who want additional Chapter 4 covers the three categories of approaches Energy Savings detail on impact evaluation for determining energy and demand savings.Chapter approaches 5 defines net savings terms and uses,briefly describes Chapter 5: Determining methods for determining net savings,and discusses Net Energy Savings issues associated with the use and calculation of Chapter 6:Calculating net savings.Chapter 6 provides approaches for Avoided Air Emissions determining avoided air emissions associated with efficiency programs. Part 4 Chapter 7: Impact Program implementers, Chapter 7 provides background on topics associated Evaluation evaluators,and managers/ with implementing impact evaluations that are Considerations regulators of evaluations not covered in other chapters,such as persistence, Chapter 8: Impact looking for guidance on key demand savings,controlling uncertainty, non-energy evaluation issues and planning benefits,program costs,and evaluation issues Evaluation Planning of evaluations as well as readers associated with certain unique program types. with a background in evaluation may want to go directly to these Chapter 8"brings it all together"and describes how chapters the basics and details described in earlier chapters can be used to plan evaluation efforts. Part 5 Appendix A:Glossary Readers interested in standard The appendices provide resources and further energy efficiency evaluation background on evaluation topics. Appendix B:Other definitions and reference Evaluation Categories materials used in the evaluation and Approaches industry as well as summaries Appendix C: Resources of process,market evaluations, References cost-effectiveness analyses,and top-down evaluation December 2012 www.seeaction.energy.gov 1-3 Chapter 1 : Notes 3 National Action Plan for Energy Efficiency.(2007).Model Energy Efficiency Program Impact Evaluation Guide. Prepared by Steven R. Schiller,Schiller Consulting, Inc.www.epa.gov/eeactionplan. 4"Bottom-up"evaluation involves adding up the savings from measures,projects,and programs to estimate total portfolio impacts.Another evaluation category,termed"top-down,"refers to approaches that rely on energy consumption data or per-unit energy consumption indicators(e.g.,energy consumption per-unit of output or per person)defined for a market sector,utility service territory, or jurisdiction as the starting point for savings determination using macro-economic approaches. 5 Because the guide is a policy-neutral document,evaluation plans must address any jurisdiction-specific policy requirements. 6 The guide does not cover transportation-related energy efficiency programs. 7 In addition to guidance documents,the planning and implementa- tion of impact evaluation activities requires skilled and experienced practitioners. 1-4 www.seeaction.energy.gov December 2012 Chapter 2 Energy Efficiency Program Evaluation Overview Chapter 2 provides an overview of the three main categories of energy efficiency program evaluation.The chapter also makes the distinction between evaluations for individual energy efficiency projects and multifaceted efficiency programs. Because this guide focuses on end-use energy efficiency program evaluation, some background on different program categories is also provided.The last sections cover the importance and objectives of energy efficiency evaluation. 2.1 ENERGY EFFICIENCY PROGRAM 2.1.2 Efficiency Program Categories CATEGORIES AND DEFINITIONS There are many types of energy efficiency programs and several Before describing the different categories of evaluation and its approaches to differentiating them.One approach is to divide importance and objectives,this section provides some context programs into two categories:voluntary and mandatory.Mandatory regarding categories of efficiency programs that are the subject programs involve codes and standards that require mandated levels of evaluations,as well as definitions of the savings hierarchy. of efficiency in buildings and/or products(e.g.,equipment or appli- ances).Voluntary programs involve a wide range of mechanisms to 2.1.1 Policy Context of Energy Efficiency Programs incent consumers to use energy more efficiently. The evaluation of energy efficiency programs should support the policy goals of the programs.Thus,understanding policy goals,and Voluntary programs can be defined as including the following the context in which the programs are being implemented,affects subcategories and objectives: program evaluation. Policy goals can vary widely;however,a major- Resource acquisition.The primary objective of this program ity of states now have policies in place that establish specific energy category is to directly achieve energy and/or demand savings, savings targets for energy efficiency programs provided to customers and possibly avoid emissions,through specific actions.This by their utilities or related organizations.$The following are three category includes activities such as rebate and direct-install common ways in which states have set their efficiency goals—by programs for energy-efficient equipment,specific operational legislation,regulation,or voter initiative. or maintenance actions(e.g.,boiler tune-ups, building • All cost-effective energy efficiency.This requires that the state commissioning),and behavior-based programs that encourage or its utilities must acquire all energy efficiency measures that consumers to adopt energy and demand savings practices. are less expensive than energy supply options. Funding for These later programs typically include outreach,education, efficiency programs can be drawn from dedicated utility bill rewards, benchmarking,and/or feedback elements. surcharges and/or the budgets that utilities would otherwise Market transformation(MT).The primary objective of this use to procure a more expensive energy supply. program category is to change the way in which energy • Energy efficiency resource standard(EERS).An EERS requires efficiency markets operate(e.g.,how manufacturers,distribu- that a percentage of the resources used by utilities to supply tors, retailers,consumers,and others sell and buy energy- their customers must come from energy efficiency.An EERS related products and services),which tends to result in more policy can mandate that the utility increase the percentage indirect energy and demand savings. Education and training of energy efficiency incrementally over a number of years (E&T)programs and programs that support the development (e.g., 1%per year)or achieve a specific target percentage of or compliance with codes and standards(C&S)are examples by a future date(e.g.,20%by 2025): of market transformation activities.These programs indirectly • Target spending budget.This requires that an efficiency result in energy savings.To a large extent,all programs can program administrator spend a certain amount of money on be considered market transformation programs in that they energy efficiency portfolios and maximize energy or peak sav- involve a change in how energy efficiency activities take place ings within these portfolio budgets. in the marketplace. • Multiple objectives.Programs can include some or all of the above-listed objectives. December 2012 www.seeaction.energy.gov 2-1 APPLICATIONSSOME 2.1.3 Savings Hierarchy EFFICIENCY EVALUATIONS The starting point for evaluating energy and demand savings,at least with bottom-up evaluation approaches,is a savings hierarchy for • Utility-administered energy efficiency programs energy efficiency actions,as shown in Figure 2.1.This figure shows • Government efficiency programs,either for public the energy efficiency actions in the following order: facilities or for private-sector incentive programs Energy efficiency measure:at an end-use energy consumer • Independent system operator(ISO) programs to reduce facility,an installed piece of equipment or system;a strategy demand(e.g.,a forward capacity market). intended to affect consumer energy use behaviors;or modification • Air pollution and greenhouse gas mitigation programs of equipment,systems,or operations that reduces the amount that rely on efficiency actions of energy that would otherwise have been used to deliver an equivalent or improved level of end-use service. Examples • Private company programs include lighting retrofits, HVAC retrofits,and commissioning. • Energy service company contracts Project:an activity or course of action involving one or multiple energy efficiency measures at a single facility or site. Examples include home retrofits and commercial new construction projects. This guide focuses on documenting the impacts of resource acquisi- Program:an activity,strategy,or course of action undertaken tion programs,including directly achieved energy and demand by a program implementer or administrator. Each program savings and related emission reductions.Section 7.10 of this guide is defined by a unique combination of program strategy, briefly discusses evaluation of market transformation programs, market segment,marketing approach,and energy efficiency including education and training and codes and standards programs. measure(s). Programs consist of a group of projects with similar It should be noted that while a program may have just one primary characteristics and installed in similar applications. Examples objective,there are often secondary objectives that are integral include a utility program to install energy-efficient lighting in to the program's overall success.This is frequently the case when commercial buildings,a developer's program to build a resource acquisition and market transformation objectives are subdivision of homes that exceed common practice,or a state's involved.With respect to impact evaluation,it is more important to effort to improve compliance with energy efficiency codes. focus on the performance goals to be assessed and establish metrics than to categorize individual program types. End-use(consumer)energy efficiency is part of the very general FIGURE 2.1: Hierarchy of energy efficiency activities category of activities known as demand-side management(DSM). Demand-side management programs are designed to encourage consumers to modify their level and pattern of energy use.Another category of DSM is demand response(DR),defined by the Federal Energy Regulatory Commission(FERC)as"a reduction in the consumption of electric energy by customers from their expected consumption in response to an increase in the price of electric energy or to incentive PROGRAMS payments designed to induce lower consumption of electric energy."9 Demand response programs employ energy rate design(pricing), customer incentives,and technology to enable customers to change their demand in response to system conditions or prices. PROJECTS While this guide does not specifically address DR programs,the basic evaluation approaches and planning process explained here can be applied to DR with the understanding that the emphasis for DR program MEASURES evaluation is demand savings.Demand savings definitions and evaluation techniques are highlighted in Sections 7.2 and 7.10.5. 2-2 www.seeaction.energy.gov December 2012 • Portfolio:either(1)a collection of similar programs addressing WHY CONDUCT EVALUATIONS? the same market(e.g.,a portfolio of residential programs), technology(e.g.,motor efficiency programs),or mechanisms The reasons to do an evaluation can be summarized in two (e.g.,loan programs),or(2)the set of all programs adminis- words:improvement and accountability. Evaluations provide tered by one organization,such as a utility. information that can help improve programs and demonstrate 2.2 PROGRAM EVALUATION CATEGORIES internal and external accountability for the use of resources. AND IMPACT EVALUATION DEFINITIONS Program evaluations provide timely information to improve The variety of evaluation activities that are associated with energy not only program implementation,but also the design of efficiency can be categorized in several different ways,one of which future programs and individual energy efficiency projects. is to define evaluations as either formative or outcome. Formative They can answer the following questions: evaluations are associated with helping efficiency programs be as • Are the program and the projects that make up the effective as possible.Outcome evaluations are associated with program achieving their goals?If so,how and why? documenting program results. However,the most common way . How well has the program/project worked? to categorize efficiency evaluations is as impact,process,or market • What changes are needed to improve the evaluations.These are defined as follows(with the first two program/project? described in more detail in Appendix B). • Impact evaluations:outcome evaluations of the changes • What is the program's impact on actual projects attributable to an energy efficiency program.While impact and future projects? evaluations usually focus on determining the quantity of • Should the program/project be replicated,adjusted, changes in energy use and demand associated with a program, or cancelled? the calculation of non-energy benefits(or co-benefits)such Program evaluations also provide an understanding of as avoided emissions and job creation that directly or indirectly result from a program can also be an output of impact evaluations. the following: Impact evaluations often support cost-effectiveness analyses • Program approaches that are most and least effective, that document the relationship between the value of program as well as how to improve future programs results(i.e.,energy,demand,and emission savings)and the • Where to focus for greater savings costs incurred to achieve those benefits.Cost-effectiveness • Actual values that can be used in future estimates of (sometimes called cost-benefit)analyses may also take into benefits(e.g.,estimates of energy savings per square account market evaluation results considering a program's foot of office space). short-and long-term market effects. • Process evaluations:systematic assessments of an energy efficiency program.Their purpose is to document program Market evaluations:a very broad category of activities that operations and identify and recommend improvements to document aspects of the marketplace with respect to energy increase the program's efficiency or effectiveness for acquiring efficiency.One particular type is a market effects evaluation, energy resources while maintaining high levels of participant which characterizes changes in the structure or functioning of satisfaction. For example,process evaluations can include an a market or the behavior of market participants that resulted assessment of program delivery,from design to implementation, from one or more program efforts. Market effects evaluations to identify bottlenecks,successes,failures,constraints,and can include projections of impacts that a market could have on potential improvements.Timeliness in identifying opportunities future energy efficiency efforts. If the evaluation's goal is to for improvement is essential to making corrections along the assess cost-effectiveness for stakeholders or regulators,excluding way.Process evaluations also provide a backdrop for interpreting the measurement of market effects could result in underesti- the results of impact evaluations. mating(or possibly overestimating)a program's overall benefits or cost-effectiveness. December 2012 www.seeaction.energy.gov 2-3 TABLE 2.1: Summary of Evaluation Categories and Types EVALUATIONEVALUATION PHASE AT WHICH IT IS . ASSESSMENT - . - Market Assessment Analyses Market, Portfolio, Program (includes characterization, baseline) Pre-program Planning Phase Potential or Feasibility Analyses Portfolio, Program, Project Formative Portfolio, Program, Project Implementation Phase Process Evaluations Portfolio, Program and Ongoing and/or After Program Implementation Market Effects Assessments Portfolio, Program Impact Evaluations Program, Project, Measure Implementation Phase— Outcomes Ongoing and/or After Market Effects Evaluations Market, Portfolio Program Implementation Cost-Effectiveness Analyses Portfolio, Program,Project The following are example questions that could be used to implicitly bundled with impact evaluation.Table 2.1 summarizes determine market effects: these categories of efficiency evaluation,although not all of these — Did a program encourage more vendors to offer energy- different evaluations are necessary for every program or portfolio. efficient products,and will there thus be future efficiency 2.2.1 Evaluation Definition benefits associated with such increased availability of products? Evaluation is the conduct of any of a wide range of assessment studies and other activities aimed at determining the effects of a — Did a voluntary incentive program prove that a new tech- program(or a portfolio of programs).This includes understanding nology is viable,cost effective,and accepted by consumers, or documenting program performance,program or program-related and therefore make it possible for this technology to be markets and market operations,program-induced changes in energy included in a future building code or appliance standard? efficiency markets,levels of demand or energy savings,or program There are other types of market evaluations:market assessment cost-effectiveness.While this guide focuses on evaluations of studies used to determine current practices for the purposes individual programs,the basic concepts can be applied to portfolios. of establishing measure,project,or program baselines;and Measurement and verification(M&V)is another term often used potential studies used to estimate the technical,economic,or when discussing analyses of energy efficiency activities. M&V can be market-based potential of increasing the amount of energy a stand-alone activity or it can be a subset of program impact evaluation. efficiency for various products and services. In either case,it is associated with the documentation of energy While this document focuses on impact evaluation,all types of (and/or demand)savings at individual sites or projects using one formative and outcome evaluations are not mutually exclusive,and or more options that can involve measurements,engineering calcula- there are benefits to undertaking more than one type at a time and tions,statistical analyses,and/or computer simulation modeling. integrating the data collection and analyses functions.Thus, process These options are defined in the International Performance evaluation and market effects evaluation often end up explicitly or Measurement and Verification Protocol(IPMVP).10 Generally speaking, 2-4 www.seeaction.energy.gov December 2012 the differentiation between evaluation and project M&V is that of contractors and administrators responsible for implementing evaluation is associated with programs(or portfolios)and M&V is efficiency programs. associated with projects.Contractors and the owners of facilities tend 2. Help understand why program-induced effects occurred to be interested in only M&V on their own project(s),while program and identify ways to improve current programs and future administrators are interested in evaluation of their programs and programs.The role of evaluation can go well beyond simply portfolios. documenting savings to actually improving current and future As discussed in later chapters, M&V is also one of the three programs.If applied concurrently with program implementation, approaches used for program evaluation,where M&V techniques evaluations can provide information in real time to allow for as-needed course corrections. Evaluation fosters more-effective are typically used to determine the savings from a sample of projects programs and can justify increased levels of energy efficiency (versus a census),with the results applied to the entire program investment as along-term, reliable energy resource. Perhaps population of projects.The other two evaluation approaches are the imperative for conducting evaluation is best described by deemed savings(which does not involve any project-specific mea- a quote attributed to John Kenneth Galbraith:"Things that are surement)and large-scale consumption data analysis. Both are measured tend to improve:' typically applied to all of the projects(or sites)in a program. 3. Support energy demand forecasting and resource planning The term evaluation,measurement,and verification(EM&V)is by understanding the historical and future effects of energy frequently seen in evaluation literature and is a catchall term for efficiency as compared to other energy supply and demand- determining both program and project impacts. side resources.As efficiency has become a more important energy resource in state and regional energy plans,an objective 2.3 OBJECTIVES AND IMPORTANCE of impact evaluations can be to support state and regional OF EVALUATION energy forecasting and resource-planning efforts.Understanding The first step in the evaluation process is to define the objectives and supporting the needs of forecasters and planners(and for given evaluation activities.This leads to defining the metrics, their data formats and definitions)can thus be an important assumptions,approaches,budgets,and other characteristics of the consideration when defining end-use efficiency program evaluation effort. Evaluations have three overarching objectives: evaluation metrics and reporting requirements. In addition, Document the benefits/impacts of a program and determine evaluation can support resource planning through projections 1. whether the program(or portfolio of programs)met its goals. of non-energy benefits,specifically emissions profiles for Rigorous evaluations help ensure that programs are cost-effective planning how to meet air quality and greenhouse gas and that benefits(e.g.,energy savings,avoided emissions) mitigation objectives(see Chapter 6). are both "real"and sustained over time.This often includes There are several technical and policy barriers to the full use assessment of compliance with regulatory requirements of cost-effective energy efficiency,and to the incorporation of associated with programs funded by the public(or energy efficiency programs into energy resource portfolios.One of these consumers). Energy efficiency impact evaluations are oriented barriers is proving that energy efficiency"can be counted on"or is toward developing retrospective estimates of energy savings "real."Consistent,complete,accurate,and transparent evaluation attributable to a program,in a manner that is defensible mechanisms for documenting energy and demand savings,as well as in regulatory proceedings conducted to ensure that public non-energy benefits such as avoided emissions,address this barrier. funds are properly and effectively spent. Regulators support Indeed,having effective evaluation policies, processes,and trained evaluation activities because of their interest in documenting personnel in place to document the energy and non-energy benefits total savings,assessing the cost-effectiveness of efficiency of energy efficiency programs is critical to the success of energy compared to generation alternatives,and assessing savings efficiency,emission,and climate change-mitigation programs that attribution(e.g.,the contributions of efficiency portfolio must prove their value and worthiness for continued investment. administrators in achieving savings versus the influences of common practice,end-user self-motivation,or codes and Evaluation is thus not a goal unto itself;it should be viewed as one standards).With respect to this last objective,evaluation can part of a continuous,and usually cyclic,process of program planning, also be used explicitly for retrospectively determining the implementation,and evaluation.The results of impact evaluation performance(and resulting payments, incentives,or penalties) studies do not stand alone but are used as inputs into planning December 2012 www.seeaction.energy.gov 2-5 and improving future programs.As shown in Figure 2.2,there is a continuum of strategies associated with moving energy efficiency actions from the research,development,and demonstration(RD&D) stage to an increasing level of adoption,and to ultimately have them become standard practice or be mandated through codes and standards.With public policy orientated toward accelerating the success and fulfillment of these strategies and greater energy savings,evaluation is a tool that supports the acceleration through documentation,feedback,and energy resource planning. FIGURE 2.2: Continuum of energy efficiency actions DEPLOYMENT TRANFORMED MARKETS Voluntary implementation . Standard practice of projects and programs: or outreach, education, • Codes and standards and subsidies. RD&D Research For example: Development • Incenting customers, distributors, Demonstration and manufacturers • Mass market and individual market strategies 2-6 www.seeaction.energy.gov December 2012 Chapter 2 : Notes 8 Nowak,S.; Kushler, M.;Sciortino, M.;York,D.;Witte,P. (June 2011). Energy Efficiency Resource Standards:State and Utility Strategies for Higher Energy Savings.American Council for an Energy-Efficient Economy(ACEEE). Report Number U113.www.aceee.org/ resea rch-report/u 113. 9 Federal Energy Regulatory Commission(FERC).(June 2010). National Action Plan on Demand Response. Docket No.AD09-10. Prepared by FERC staff.www.ferc.gov/legal/staff-reports/ 06-17-10-demand-response.pdf. 10 Efficiency Valuation Organization(EVO). International Performance Measurement and Verification Protocol(IPMVP).(multiple dates). www.evo-world.org.The IPMVP is an international M&V guidance document. It is discussed in Chapter 4 and references are in Appendix C. December 2012 www.seeaction.energy.gov 2-7 Chapter 3 Impact Evaluation Basics Chapter 3 is a stand-alone introduction to the basic concepts associated with energy efficiency impact evaluation. It introduces the evaluation process,the role of evaluators versus administrators of efficiency programs, and some key metrics. Also presented is the concept of savings determination based on a counterfactual situation and the fundamental resulting issue of evaluation being the balancing of evaluation costs with the reliability of savings estimates.The chapter concludes with a brief introduction to the approaches used for determining energy and demand savings: deemed savings, measurement and verification, and large-scale consumption data analysis. 3.1 IMPACT EVALUATION PROCESS EVALUATION PLANNING ISSUES Impact evaluations determine program-specific induced effects, which include reductions in energy use(such as kilowatt-hours[kWh] Chapter 8 discusses the evaluation planning process. and therms)and demand(kilowatts[kW]),and non-energy benefits The planning process is used to decide which(or which such as avoided air emissions.The basic steps in the evaluation combination)of the evaluation metrics and approaches process are as follows: defined in this guide should be used. • Set the program evaluation objectives in the context of the program policy objectives. • Select an impact evaluation savings determination approach, Verification would involve confirming that the replaced lamps are define baseline scenarios,and prepare a plan that takes into 100 watts,that the new CFLs are 23 watts,and that the CFLs were account the critical issues. installed and are working.As a result of this verification,it can be • Determine energy and demand savings. confirmed that the project has the potential to save energy,with the amount of energy saved dependent on how many hours the lamps • Determine non energy benefits(as needed). operate. Determining how many hours the lamps operate,in this • Report the evaluation results and,as appropriate,work with case,would be the"documentation"activity(or what some might program administrators to implement recommendations for call the"measurement,"although as discussed in Section 3.3,savings current or future program improvements and/or resource cannot be literally measured). planners and demand forecasters to support their efforts. In this guide,verification is defined formally as an independent The program evaluation process begins with defining and assessing assessment that a program has been implemented per the program the evaluation objectives.Well-defined objectives indicate what design.For example,the objectives of measure installation verification data need to be collected or developed during the evaluation effort are to confirm(1)the installation rate(number of units installed), and the scope and scale of effort required for meeting the objectives (2)that the installation meets reasonable quality standards,and(3) (e.g.,the cost of obtaining the desired information,schedules,labor that the measures meet the program eligibility requirements and requirements).A key to successful evaluation is the comparison of are operating correctly with the potential to generate the predicted the costs of evaluation with the value of the information that will come savings.For some programs,it may be that verifying the potential to from the evaluation, possibly through an iterative planning process generate savings is all that is needed to meet the evaluation objectives, that balances cost and value. while in many other situations,both verification and documentation 3.1.1 Verification and Documentation of the actual savings value will be required. Within the impact evaluation process there tends to be two types of Verification may include one-time or multiple activities over the major activities:verification and documentation.These may be more estimated life of the measures and can overlap with commissioning aptly called"verifying the potential to generate savings"and"documenting (Cx)or retro-commissioning(RCx)activities,which have similar (determining)the actual savings! To illustrate the difference between objectives—to ensure the installed equipment is working correctly the two,consider a project involving replacement of 100-watt and per design. However,verification may not go"as far"as Cx or incandescent lamps with 23-watt compact fluorescent lamps(CFLs). RCx to ensure the operation of the energy efficiency measure(s). 3-1 www.seeaction.energy.gov December 2012 BASIC IMPACT EVALUATION CONCEPTS activities are completed.These are typically estimates of savings prepared for program and/or portfolio planning purposes. • Impact evaluations are used for determining achieved • Claimed savings:values reported by an implementer or program effects. administrator after the subject energy efficiency activities • Savings cannot be directly measured,only indirectly have been completed. determined by comparing energy use and demand • Evaluated savings:savings estimates reported by an independent after a program is implemented to what they would third-party evaluator after the subject energy efficiency activities have been had the program not been implemented and an impact evaluation have been completed.These can (i.e.,the baseline). differ from claimed savings in that an independent third-party • Successful evaluations harmonize the costs incurred evaluator,to an agreed-to level of rigor,has conducted evaluation with the value of the information received;in other and/or verification activities. words,they appropriately balance risk management, uncertainty,and cost considerations. The implementers and/or administrators usually prepare their projected savings estimates and claimed savings estimates.They will thus conduct their own evaluation activities,using their own evalua- tion staff or consultants,for purposes such as confirming any incen- 3.1.2 The Evaluator's Role and Reporting of Savings tive payments to program participants or contractors and preparing There are several entities involved in the actual implementation of documentation for internal and external reporting.11 If an indepen- efficiency projects and programs,and they may each conduct their dent third-party evaluator is used,that evaluator will then conduct own impact evaluation activities.These entities include the end-use some level of evaluation(verification only or verification and their energy consumers who have projects installed in their facilities, own data collection/analysis to determine savings)for preparation of designers,contractors,and program implementers and administra- its own evaluation reports and a realization rate12 comparing evalu- tors.Some of these entities may only be interested in verification ated savings with projected savings estimates and/or claimed savings activities or just their own individual projects'measurement and estimates.Of course,the evaluator can and should be brought into verification,while others will conduct complete program evaluations. the process before any of this work is conducted to participate in The following definitions help to explain these roles. defining the roles and responsibilities of the administrators,imple- Entities: menters,and evaluators as well as reporting requirements.The • Administrator:an entity selected by a regulatory or other designation of"independent"and"third-party"is determined by government organization to contract for and administer an those entities involved in the use of the evaluations and may include energy efficiency portfolio within a specific geographic region and/or market.Typical administrators are investor-owned or public utilities, nonprofits,and state government agencies. QUALITY ASSURANCE An administrator could also be a private entity that hires a company,such as an energy services company(ESCO),to The impact evaluation approaches described in this guide are implement its efficiency program(s). based on new and unique analyses of energy and demand savings. However,often there is documentation on energy • Implementer:an entity selected and contracted with,or and demand savings from analyses prepared independently qualified by,a program administrator to provide products of the subject impact evaluation.Even though such documentation and/or services to consumers,either directly or indirectly. was not necessarily prepared per predetermined evaluation • Independent third-party evaluator:an entity that conducts requirements,it may be sufficient for meeting the evaluation evaluations and is designated to be independent of the objectives. Using existing documentation in combination with implementer and administrator. quality assurance guidelines(QAG)can significantly reduce overall program/evaluation costs. Essentially,a QAG can help Reported Savings: determine whether indicated savings,and the assumptions • Projected savings:values reported by an implementer or and rigor used to prepare the existing documentation,can be administrator prior to the time the subject energy efficiency used in place of new evaluation efforts. December 2012 www.seeaction.energy.gov 3-2 evaluators retained,for example,by the administrator or a regulator. Defining the relative roles of the administrator,implementer,and independent third-party evaluator is another important activity of Savings,or more accurately stated"savings estimates," the planning process.Section 8.3.2 of this guide discusses evaluator from energy efficiency measures, projects,programs,and roles and selection. portfolios are reported at various times in the lifecycle of 3.2 ENERGY AND NON-ENERGY BENEFIT the efficiency activity and with varying degrees of cer- EVALUATION METRICS tainty.Savings are most commonly reported at two major milestones—prior to and after the implementation of the For energy and demand savings(and conceptually for non-energy activity.Savings can also be indicated as first-year,annual, benefits)the primary metrics are known as gross energy savings and/or lifetime energy or demand savings values.They also and net energy savings.In this guide,based on industry standard can be indicated as gross savings and/or net savings values. practice,gross and net energy(and demand)savings are defined Different jurisdictions currently have different names for as follows: savings reports,what they contain,and whether and what • Gross energy savings:the change in energy consumption and/ adjustments or evaluation activities take place between or demand that results directly from program-related actions preimplementation and postimplementation. Ideally,the taken by participants in an efficiency program, regardless of terms and methods in this guide should be applied. However, why they participated.This is the physical change in energy use whenever savings are reported,it is critical that the basis for after taking into account factors not caused by the efficiency the values indicated be made clear. actions(e.g.,changes in weather or building occupancy). • Net energy savings:the change in energy consumption and/or demand that is attributable to a particular energy When energy or demand savings are reported,they are typically efficiency program. Estimating net energy savings typically estimated for the first year of a program,for a specific number of involves assessing free ridership and spillover,although this years(e.g.,5,10,15),or for the life of the program's measures. guide discusses additional considerations. In the efficiency Measure life is the length of time that an energy efficiency measure industry,free ridership refers to the portion of energy savings is expected to be functional and generating savings. It is a function that participants would have achieved in the absence of the of equipment life and measure persistence.Equipment life is the program through their own initiatives and expenditures number of years that a measure is installed and will operate until (i.e.,the participant would have undertaken the energy-saving failure.Measure persistence refers to the duration of an energy- activity anyway).Spillover refers to the program-induced consuming measure,taking into account business turnover,early adoption of measures by nonparticipants and participants retirement of installed equipment,and other reasons measures who did not claim financial or technical assistance for addi- might be removed or discontinued.Measure life is sometimes tional installations of measures supported by the program. For referred to as expected useful life(EUL). instance,a participant undertakes additional energy efficiency measures due to positive experience with the program,or a The other two main metric categories are non-energy benefits nonparticipant undertakes such measures based on observing and cost-effectiveness. a program participant's results. Net savings estimates also • Non-energy benefits(NEBs):the identifiable—although sometimes include consideration of market effects. sometimes unquantified—non-energy impacts associated with program implementation or participation;also referred to as The difference between these two metrics is associated with(1) non-energy impacts(NEI)or co-benefits. Examples of NEBs attribution of the savings—in other words,the determination of include environmental benefits,productivity improvements, whether the savings were caused by the program being studied jobs created,reduced program administrator debt and disconnects, (entirely or partially)or by other influences such as prior year and higher comfort and convenience levels of participants.The programs or other programs/influences operating at the same time value of NEBs is most often positive,but may also be negative as the program;and(2)differences in how different entities(e.g., (e.g.,the cost of additional maintenance associated with a regulatory bodies)define net and gross savings.Approaches for sophisticated energy-efficient control system),which is why determining gross and net savings are summarized in this chapter, some practitioners prefer the term NEls. Potential benefits of with additional information provided in Chapters 4 and 5. efficiency to the energy system(e.g.,price stability,grid reliability, 3-3 www.seeaction.energy.gov December 2012 and power quality)are"energy-related"but are also often put FIGURE 3.1: True program savings: the counterfactual into this general category of NEBs.The primary NEB addressed for a residential household energy efficiency program in this guide is avoided air emissions(see Chapter 6).Section 7.9 has more information on calculating non-energy benefits. • Cost-effectiveness:an indicator of the relative performance TRUE or economic attractiveness of any energy efficiency investment PROGRAM or practice relative to energy supply resources. It is another SAVINGS metric that is commonly used when reporting the results of impact evaluations.In the energy efficiency field,the present value of the estimated benefits produced by an energy efficiency program is compared with the estimated total costs of the program in order to determine whether the proposed investment or measure is desirable from a variety of perspectives(e.g., whether the estimated benefits exceed the estimated costs from a societal perspective or from a program participant perspective). 3.3 FUNDAMENTAL ENERGY EFFICIENCY IMPACT EVALUATION CONCEPTS: THE COUNTERFACTUAL AND MANAGING Households Counterfactual: UNCERTAINTY in Program Exact Same Households, 3.3.1 The Counterfactual Same Time Period, No Program In theory,the true energy savings from an energy efficiency program is the difference between the amount of energy that participants in the program use relative to the amount of energy those same partici- (e.g.,through measurement),but to determine the savings(the blue pants would have used had they not been in the program(during the shaded area),the energy use that would have occurred without the same time period).This baseline is called the counterfactual scenario project(the green line)has to be estimated in order to determine a (see Figure 3.1). However,in practice,we can never observe how value for energy savings. much energy those participants would have used had they not been in the program,because at any given time a participant must either As discussed in Chapter 2,an objective of program evaluation is be in the program or not.Thus,there is no direct way of measuring to produce energy and demand savings values(and,as desired, energy(demand)savings,because(1)it is not possible to measure a associated non-energy benefits). However,as noted above,these participant's energy use with and without the program,at the same values are always going to be estimates;the use of these estimates time;and(2)one cannot measure the absence of energy use. as a basis for decision making can be called into question if their sources and level of accuracy are not analyzed and described. Defining this counterfactual scenario represents the fundamental Therefore,evaluation results,like any estimate,should be reported concept and the greatest challenge to documenting the benefits as"expected values"with an associated level of uncertainty.Most of energy efficiency.This challenge is met with impact evaluations of the remainder of this guide describes the approaches,issues,and measuring energy consumption—but,the savings themselves will planning processes that should be considered when addressing the always be estimates.The savings estimate is the difference between counterfactual challenge and undertaking impact evaluations. (1)actual energy consumption after a project or program is imple- mented,and(2)what energy consumption would have occurred 3.3.2 Baselines during the same period,by the same participants,had the efficiency The baseline is the counterfactual scenario,determined on the basis project/program not been implemented. of a number of considerations:the evaluation approach being used, the type of project being implemented,site-specific issues,and The graph in Figure 3.2 summarizes this estimation process.The blue broader policy-orientated considerations.These considerations usu- line represents energy use of a building before,during,and after an ally result in one of three different types of baselines being selected efficiency project is implemented.This energy use can be known for the impact evaluation:existing conditions,common practice,or December 2012 www.seeaction.energy.gov 3-4 FIGURE 3.2: Energy consumption before, during, and after a project is implemented BEFORE PROJECT INSTALLED � AFTER PROJECT INSTALLED • Estimated Energy Use Without Efficiency Project l W 0 cc W 4� Energy Savings Z W Energy Use Before Efficiency Project Energy Use After Efficiency Project TIME codes/standards. Baselines are discussed throughout this guide with Control group approaches.These are the large-scale consumption respect to the evaluation approaches and planning for evaluation data analysis approaches that are also described in Chapter 4. activities.The first section of Chapter 7 provides more information With these approaches,a comparison group's energy use is on selecting baselines. compared with the energy use of program participants.These approaches,in most cases,generate estimates of net savings, 3.4 DETERMINING ENERGY AND taking into consideration free ridership and participant spillover, DEMAND SAVINGS but do not take into account nonparticipant spillover and long- The three impact evaluation approaches used to determine energy term market effects,which some jurisdictions include in the net and demand savings can be grouped into two conceptual frameworks: savings determination. noncontrol group approaches and control group approaches. 3.4.1 Noncontrol Group Impact Evaluation Approaches • Noncontrol group approaches.These are the deemed savings The following are brief summaries of the two noncontrol group and M&V approaches defined below and in greater detail approaches:measurement and verification and deemed savings. in Chapter 4.With these approaches,pre-project(or pre- program)baseline energy use is defined using one or more Measurement and verification(M&V).Measurement and of a variety of different methods.This baseline is compared verification is the process of using measurements to reliably with post-project(or post-program)energy use measurements determine energy and/or demand savings created within an or assumptions to estimate savings.These noncontrol group individual facility.The International Performance Measurement approaches generate estimates of gross savings,which require and Verification Protocol(IPMVP),an international M&V adjustments to determine net savings. guidance document,defines four M&V options used in the 3-5 www.seeaction.energy.gov December 2012 efficiency industry:two end-use metering approaches,energy installed correctly,number of point-of-sale CFLs that were sold). use data(billing data)regression analysis,and calibrated This approach is only valid for projects with fixed operating computer simulation. conditions and well-known,documented stipulation values.This • Deemed savings.Deemed savings are based on stipulated approach involves multiplying the number of installed measures values,which come from historical savings values of typical proj- by the estimated(or deemed)savings per measure. ects.A typical source of such historical values are prior year M&V or large-scale consumption data analysis studies.Unlike the A variant of deemed savings is the deemed savings calculation, M&V approach,with the use of deemed savings,there are no(or which is one or more agreed-to(stipulated)engineering algorithm(s) very limited)measurement activities;instead,only the number used to calculate the energy and/or demand savings associated of measures implemented is verified(e.g.,number of motors with an installed energy efficiency measure.These calculations may include stipulated assumptions for one or more parameters in the algorithm,but typically they require users to input data associated MEASUREMENT AND VERIFICATION with the actual installed measure into the algorithm(s). VERSUS DEEMED SAVINGS 3.4.2 Control Group Approaches: Large-Scale For simple,well-defined,efficiency measures whose Consumption Data Analysis (Randomized Controlled performance characteristics and use conditions are well Trials and Quasi-Experimental Methods) known and consistent,a deemed savings approach may be A reliable(precise,unbiased)approach for estimating energy savings appropriate.Since they are stipulated and,by agreement, from efficiency programs is to measure the difference between the fixed during the period for which savings are reported(e.g., energy use of facilities(e.g.,houses)participating in a program(the first year or lifetime),deemed savings can help alleviate some "treatment group")and that of a similar comparison group of non- of the guesswork in program planning and design;in effect, participating facilities(the"control group")during the same period they minimize one type of risk by providing certainty. of time.The two generic categories of control group approaches are However,deemed savings can result in another form of randomized controlled trials and quasi-experimental methods: risk if not properly developed and applied—overestimates Randomized controlled trials(RCTs).In an RCT,a study population or underestimates of savings if the projects or products do (e.g.,single-family houses in Seattle that have electric heat)is not perform as expected.This can occur,for example,if the defined and randomly assigned to either the treatment group deemed savings value is incorrectly calculated or the deemed or the control group. Energy use(consumption)data must be savings value was simply applied to the wrong type of application. collected for all of the project sites in the treatment and control group in order to estimate energy savings.The energy savings Measurement-based approaches are more appropriate estimate is then calculated by comparing the difference between for more complex efficiency projects or for project with the measured energy use(or preferably the difference between significant savings variability(i.e.,those with a significant the measured change in energy use)of the treatment house- amount of savings,or"risky"savings,or with no history or holds and the energy use of the control households during the analysis or metering on which to base a deemed savings value). same period. Measurement-based approaches are also more rigorous than • Quasi-experimental methods.Unlike RCTs,with quasi-exper- deemed savings approaches and involve site data collection imental methods the assignment of the control group is not during the period of evaluation for at least the most critical totally random.Thus,quasi-experimental methods, relative to variables.These approaches add to evaluation costs but may RCTs,often suffer from selection bias and may produce biased provide more accurate savings values. estimates of energy savings. However,because of the difficulty Also,deemed savings can be used together with some and costs of conducting RCTs,quasi-experimental approaches monitoring of one or two key parameters in an engineering are more common than RCTs,with perhaps the most common calculation. For example,in a high-efficiency motor program, being the"pre-post"approach.With this approach,sites in the actual operating hours could be monitored over a full work treatment group after they were enrolled in the program are cycle.This approach is consistent with IPMVP Option A,which compared with the same sites'historical energy use prior to is described in Chapter 4. program enrollment. In effect,this means that each site in the treatment group is its own nonrandom control group. December 2012 www.seeaction.energy.gov 3-6 With these approaches,statistical analyses are conducted on the It is not unusual for combinations of these approaches to be used. energy use data(typically collected from the meter data reported on For example,rigorous randomized controlled trials may be used monthly utility bills)and other important independent variable data every three years,with self-reported or deemed NTG ratios used for (e.g.,weather)for those in the control and treatment groups.These the other program years.More information about determining net approaches are primarily used for programs with relatively homog- savings is provided in Chapter 5. enous participants and measures,when project-specific analyses are not required or practical,but could,at least in theory,be considered for every type of program.Example applications are large-scale THE COUNTERFACTUAL SCENARIO weatherization programs and residential behavior-based programs. A counterfactual analysis occurs when a person modifies 3.4.3 Approaches for Determining Net Energy a factual antecedent(a thing or event that existed before and Demand Savings or logically precedes another)and then assesses the conse- For the noncontrol group approaches where gross savings are quences of that modification.A person may imagine how determined,the difference between net and gross savings is an outcome could have turned out differently if the factual specified as a net-to-gross(NTG)ratio.The following common situation,or what led to it,did not occur.This may seem approaches are used to determine net savings: daunting,but for energy efficiency impact evaluations,this is • Stipulated net-to-gross ratios.These are ratios that are simply defining what the energy use(or demand,emissions, multiplied by the gross savings to obtain an estimate of net number of jobs,etc.)would have been if the program had not savings and are based on historical studies of similar programs. been implemented. Sources of stipulations can cover a wide range,from simply using"negotiated guesses"to historical values to structured The fact that energy and demand savings,as well as related expert judgment panels.This is the least expensive approach. non-energy benefits,from efficiency efforts cannot be directly • Self-reporting surveys and enhanced self-reporting surveys. measured results in analyses based on a counterfactual scenario. Enhanced surveys include interviews and documentation It is counterfactual because savings are not measured,but review and analysis.These are moderately expensive approaches. rather estimated to varying degrees of accuracy by comparing energy consumption after a program is implemented(the • Panel of trade allies.A significant number of trade allies reporting period)with what is assumed to have been the provide information on their recent projects,whether the energy consumption(and demand)in the absence of the projects are in the subject program or not,to assess the project program(the baseline or the counterfactual scenario). program's impact on incented and nonincented program The baseline and reporting period energy use and demand energy efficiency measures. are compared and adjusted so that only program effects are • Large-scale consumption data analysis approaches(ran- considered when determining savings.These adjustments are domized controlled trial methods and quasi-experimental a major part of the evaluation process and can vary from one methods).When a control group of nonparticipants is used, program type to another and from one evaluation approach the savings indicated are"net"of free riders and participant to another. spillover.These are discussed in Chapter 4. • Cross-sectional studies.These studies are comparisons of market share of targeted technologies or behaviors between a baseline area not served by the program and the area served by the program. • Top-down evaluations.These evaluations use state, regional, or national data at a sector level to assess the extent to which markets for energy-efficient products and services have been affected by programs. 3-7 www.seeaction.energy.gov December 2012 Chapter 3: Notes 11 In many administrator organizations,these internal evaluations are called"measurement and verification"or just"verification"and are conducted by the same team that implements the programs. 12 Realization rate is used in several contexts for comparing one savings estimate with another.The primary and most meaningful application is the ratio of evaluated gross savings to claimed gross savings(versus comparing net and gross savings estimates,which is best defined with a net-to-gross ratio). Basis for the ratio not being 1.0 can include several considerations such as the following:(1) adjustments for data errors,(2)differences in implemented measure counts as a result of verification activities,and/or(3)other differ- ences revealed through the evaluation process,such as with respect to baseline assumptions. December 2012 www.seeaction.energy.gov 3-8 Chapter 4 Calculating Energy Savings As discussed in Chapter 3, there is no direct way of measuring energy savings, because one cannot measure the absence of energy use. However, the absence of energy use (i.e., savings) can be estimated. Within the efficiency evaluation industry there are three generic classifications of savings determination approaches, and the following sections describe each of them in more detail than what was summarized in earlier chapters: measurement and verification, deemed savings, and large-scale consumption data analysis (with the use of control groups).These descriptions are intended to be overviews,with additional information sources referenced in Appendix C.The last section of this chapter describes some criteria for selecting an approach. Supporting information in Chapter 5 provides information on determining net savings. Chapter 7 provides information on other impact evaluation topics related to calculating energy savings (and demand and non-energy benefits), and Chapter 8 discusses planning evaluation efforts. 4.1 MEASUREMENT AND VERIFICATION extensively as a program evaluation approach,primarily for"custom APPROACH projects"whose savings are dependent on the technologies applied Measurement and verification(M&V)is the determination of gross and/or the situations in which they are applied. For program evalua- energy savings at individual sites or projects using one or more tion,M&V involves the following activities: methods that can involve measurements in combination with • Selecting a representative sample of projects in a specific engineering calculations,statistical analyses,and/or computer efficiency program,although in some programs all of the simulation modeling. projects may be selected for M&V(a census) Measurement and verification is a project-based approach to • Determining the savings of each project in the sample,which is determining savings. Its genesis was in the efficiency performance done with M&V activities consisting of the following: contracting industry,starting in the 1970s.With performance • Development of a M&V plan contracting,a contractor implements an efficiency project for a client — Meter calibration and installation and,for long-term mea- via a contractual arrangement that includes a savings(performance) surements,maintenance guarantee or an arrangement that payment to the contractor is — Data gathering and screening dependent on the savings achieved.These arrangements required — Computations of savings with measured data "measurement and verification"to determine what level of savings — Quality assurance reviews and reporting were being achieved at the client's facility. By the early 1990s,the growth of the performance contracting industry was constrained, Applying the sample projects'savings to the entire population (i.e.,the program). in part,by the lack of robust methodologies for verifying project savings,which,in turn,restricted the ability of the project finance industry to participate in the market.To this end,the efficiency industry(represented by the National Association of Energy Service UNIFORM METHODS PROJECT Companies)14 worked with the U.S.Department of Energy(DOE),the American Society of Heating, Refrigerating and Air-Conditioning A source of more detailed impact evaluation information is Engineers(ASHRAE),and other stakeholders in the mid-90s to DOE's Uniform Methods Project(UMP),13 which provides develop guidelines,which became the basis for the International model evaluation plans for specific energy efficiency Performance Measurement and Verification Protocol(IPMVP),15 measures and project categories.These UMP documents described below. contain additional information and specific examples that As publicly funded efficiency programs became more prevalent in apply the concepts presented in this guide and include the 1990s,the energy efficiency industry adopted M&V techniques examples of the three impact evaluation approaches and concepts into the growing number of evaluations that were presented in the following sections. being conducted for efficiency programs.Thus,M&V is now used 4-1 www.seeaction.energy.gov December 2012 FIELD INSPECTIONS OF ENERGY With M&V,energy(and/or demand)savings are determined for a EFFICIENCY MEASURES project by comparing energy use(and/or demand)before and after implementation of the energy efficiency measures.Thus,the following Not all of the evaluation approaches described in this chapter fundamental algorithm applies for energy(and demand)savings: require field inspections,but typically there are some physical Energysavings= Baseline energyuse (Reporting period assessments for at least a sample of the individual projects in g ( )—( p g energy use)±(Baseline adjustments) a program(i.e.,field activities).As part of a broader verification process,field inspections ensure that the measures installed Baseline energy use:the energy consumption that would have meet appropriate specifications and that the projects included occurred without implementation of the energy efficiency in a program have the potential to generate savings.This activity.When discussed in terms of specific projects,where potential to generate savings can be verified through observa- energy saving is the metric of interest,it is sometimes called tion,inspections,and spot or short-term metering conducted preinstallation energy use. immediately before and after project installation.These field • Reporting period energy use:the energy consumption that activities can also be conducted at regular intervals during the occurs within the time frame following implementation of reporting period to verify a project's continued potential to an energy efficiency activity during which savings are to be generate savings.The field activities are an inherent part of determined.When discussed in terms of specific projects,it is the data collection aspects of the M&V approach,though they sometimes called postinstallation energy use. may be considered"add-ons"to the other approaches. • Baseline adjustments:factors that modify baseline energy or demand values to account for independent variables(such as M&V also includes all field activities dedicated to collecting site weather)that influence energy use.These adjustments account information,including equipment counts,observations of field for conditions in the reporting period that are different from conditions,building occupant or operator interviews,measurements the conditions during the baseline period but are not a result of parameters,and metering and monitoring. of the project activity.They distinguish properly determined savings from a simple comparison of energy use before and The industry's primary M&V resource is the IPMVP,which is an after implementation of a program.By accounting for independent international end-use,energy efficiency M&V guidance document. variables that are,or are not,beyond the control of the program It provides a framework for conducting M&V,and most important, implementer or energy consumer,the adjustments term brings defines four M&V options that are used in the efficiency industry. energy use in the two time periods to the same set of condi- The options involve metering of all relevant parameters,metering tions.Common examples of adjustment include the following: of key parameters,energy use(billing)data regression analysis, — Weather corrections(e.g., if the program involves heating and/or computer simulation.Complementing the IPMVP are the or air-conditioning systems in buildings) Federal Energy Management Program(FEMP)M&V Guidelines.16 — Occupancy levels and hours(e.g., if the program involves These guidelines,and companion technical notes,provide more lighting retrofits in hotels or office buildings) details on M&V for specific measure and technology applications. — Production levels e.( g.,if the program involves energy A third important M&V resource is ASHRAE Guideline 14(2002): efficiency improvements in factories). Measurement of Demand and Energy Savings.17 This guideline provides technical detail on subjects such as metering.A new One of the other considerations for adjustments is unique to version of Guideline 14 is expected to be available in early 2013. program evaluation. In almost all bilateral performance contracts between a contractor and its client,the baseline is considered to be The following subsections provide material mostly from the 2010 whatever"exists in the facility"before the efficiency measures are version of the IPMVP(EVO 10000-1:2010).They cover the basic implemented. However,in many programs using public or energy M&V concept(algorithm)for calculating gross energy savings at the consumer funds,the baseline may be defined as common practice or project level and introduce the four IPMVP Options—A,B,C,and as required by a code or standard(in order to not give credit for what D—using the descriptions found in the IPMVP. would have been done without the project activity). In this situation, the savings calculated from an end-use consumer perspective may December 2012 www.seeaction.energy.gov 4-2 be different from the savings determined for a government agency The magnitude of such interactive effects,if significant,should be or regulatory body.Thus,a single project may involve two different considered,and a method developed to estimate them under the sets of savings calculations—one that supports the contract between savings determination process. the energy services contractor and the customer(using a baseline of existing conditions in the customer facility),and a second that The four IPMVP options(A, B,C,and D)provide a flexible set of supports the customer's claim to a utility rebate(using a baseline of methods for evaluating project energy/demand savings.Having current codes and standards).See Chapter 7 for more discussion of four options provides a range of methods for determining energy/ the nuances of selecting baselines. demand savings with varying levels of savings certainty and cost. A particular option is chosen based on the specific features of each 4.1.1 Measurement and Verification Approaches project,including the following: If M&V is a part of the evaluation process,at least some M&V details . Energy efficiency measure technologies employed and the end will need to be specified in the evaluation planning documents,and uses in which they are applied if sampling is used,a basis for selecting the sample of specific project • Complexity, particularly in terms of interactive effects with sites at which M&V activities will take place will be needed.In addition, multiple measures and energy-using systems as M&V is a project-specific approach to evaluation,each project evaluated will need to have a project-specific M&V plan.There are • Potential for changes in key factors during the baseline and/or two types of project-specific M&V plans: prescriptive method plans reporting periods and generic method plans. • Uncertainty of the project savings as compared to the value of • Prescriptive method plans.For project types with significant project savings M&V"experience"and well-understood determinants of • Value of understanding the performance of the measures savings(e.g., lighting and motor retrofits),there are established (e.g.,for a new technology). M&V procedures,example plans,and standardized algorithms. The FEMP M&V Guidelines contain prescriptive approaches The options differ in their approach to the level,duration,and type for documenting savings for several common energy efficiency of baseline and reporting period measurements.The options also measures,as does ASHRAE Guideline 14.The DOE LIMP materials differ in terms of measurement boundaries: also include several standardized approaches to documenting • Measurement boundaries with Options A and B are made at savings.The UMP documents do so in the context of M&V as the end use,system level(e.g.,lighting,HVAC). an evaluation approach and are thus more germane to the . Measurement boundaries with Options C and D are at the subject of this guide than the FEMP and ASHRAE documents whole-building or whole-facility level. which are more specific. • Generic method plans.There are conceptual approaches applicable Additionally,they differ in terms of type of measurements and to a variety of project types for which prescriptive M&V methods their duration: are not available(e.g.,comprehensive building retrofits and • Option A involves using a combination of both stipulations and industrial energy efficiency measures).The FEMP and ASHRAE measurements of the key factors needed to calculate savings guidelines contain several generic M&V approaches. in engineering models. Data collection tends to involve either spot-measurements or short-term measurements.18 One of the other important aspects of M&V is defining a measurement Options B and C involve using spot,short-term,or continuous boundary.The measurement boundary might be a single piece of measurements in engineering models(Option B)or regression equipment(e.g.,the replaced motor in a factory),a system(e.g.,the analyses(Option C). entire lighting system retrofitted in a commercial building),or the whole facility(e.g.,a home that has undergone a complete retrofit). • Option D may include spot,short-term,or continuous measure- Any energy effects occurring beyond the measurement boundary ments to calibrate computer simulation models. are called"interactive effects."A typical interactive effect is the The four generic M&V options are summarized in Table 4.1. decrease in air-conditioning requirements or increase in space heat- ing requirements that can result from a lighting retrofit,which by its nature reduces the amount of heat produced by a lighting system. 4-3 www.seeaction.energy.gov December 2012 TABLE 4.1: IPMVP Options Summary IPMVP OPTION CALCULATED A.Retrofit Isolation:Key Parameter Measurement Engineering calculation of baseline A lighting retrofit in which power draw is Savings are determined by field measurement of and reporting period energy from: the key performance parameter that is the key performance parameter(s),which define the • Short-term or continuous measured periodically.Estimate operating energy use of the energy conservation measures measurement of key hours of the lights based on building (ECMs)affected system(s)and/or the success of operating parameter(s); schedules and occupant behavior. the project. estimated values Measurement frequency ranges from short-term to • Routine and nonroutine continuous,depending on the expected variations adjustments as required. in the measured parameter and the length of the reporting period. Parameters not selected for field measuring are estimated.Estimates can be based on historical data,manufacturer's specification,or engineering judgment. Documentation of the source or justification of the estimated parameter is required. The plausible savings error arising from estimation rather than measurement is evaluated. B.Retrofit Isolation:All Parameter Measurement Short-term or continuous Application of a variable-speed drive and Savings are determined by field measurement of the measurement of baseline and controls to a motor to adjust pump flow. energy use of the ECM-affected system. reporting period energy,and/or Measure electric power with a kW meter engineering computations using installed on the electrical supply to the Measurement frequency ranges from short-term to measurements of proxies of energy motor,which reads the power every minute. continuous,depending on the expected variations in use.Routine and nonroutine In the baseline period,this meter is in place the savings and length of the reporting period. adjustments as required. for a week to verify constant loading.The meter is in place throughout the reporting period to track variations in power use. C.Whole Facility Analysis of whole facility baseline Multifaceted energy management program Savings are determined by measuring energy use and reporting period(utility) affecting many systems in a facility. Measure at the whole-facility or subfacility level. meter data. energy use with the gas and electric utility Routine adjustments as required, meters for a 12-month baseline period and Continuous measurements of the entire facility's throughout the reporting period. energy use are taken throughout the reporting using techniques such a simple comparison or regression analysis. period. Nonroutine adjustments as required. D.Calibrated Simulation Energy use simulation,calibrated Multifaceted energy management Savings are determined through simulation of the with hourly or monthly utility program affecting many systems in a energy use of the whole facility or a subfacility. billing data.(Energy end-use facility but where no meter existed in metering may be used to help the baseline period. Simulation routines are demonstrated to adequately refine input data.) model actual energy performance measured in the Energy use measurements,after facility. installation of gas and electric meters, are used to calibrate a simulation. This option usually requires considerable skill in calibrated simulation. Baseline energy use,determined using the calibrated simulation,is compared to a simulation of reporting period energy use. Source:Efficiency Valuation Organization(EVO).International Performance Measurement and Verification Protocol.(2010).IPMVP,EVO 10000-1:2010.www.evo-world.org. December 2012 www.seeaction.energy.gov 4-4 4.1.2 M&V Option A: Retrofit Isolation—Key from a simple lighting retrofit(less complex)may be more accurately Parameter Measurement determined with Option A than could the savings from a chiller Option A involves project-or system-level M&V assessments in retrofit(more complex). which the savings associated with a particular project can be isolated at the"end-use"(e.g.,ventilation system,lighting system)level.With Also true with Options A and B is that measurement of all end-use this option, key performance parameters or operational parameters equipment or systems may not be required if statistically valid can be measured during the baseline and reporting periods.However, sampling is used. For example,the operating hours for a selected some parameters are stipulated rather than measured.This level of group of lighting fixtures and the power draw from a subset of verification may suffice for types of projects in which a single parameter representative constant-load motors may be metered. represents a significant portion of the savings uncertainty. Savings determinations under Option A can be less costly than Under Option A,energy and demand savings are calculated using under other options because the cost of measuring one or two "engineering models.' These models are essentially groups of equations parameters is usually less than measuring all of the parameters. defining energy use as a function of various inputs—often simple However,because some stipulation is allowed under this option, spreadsheet models—and involve the development of estimates of care is needed to review the engineering design and installation to energy and demand savings based on the following: ensure that the stipulations are realistic,applicable,and achievable • Assumptions concerning operating characteristics of the (i.e.,the equipment can truly perform as assumed).This can be done equipment or facilities in which the equipment is installed, through"desk reviews"of data but is more reliably done as part of which are informed by measurements(from spot to continuous). the verification process,where site inspections check the efficiency measure characteristics and collect site data.At defined intervals Examples are power draws(wattage)of light fixtures or fan during the reporting period,the installation can be reinspected to motors and efficiencies of air conditioners(kWh/ton)and verify the equipment's continued existence and its proper operation heaters(Btu out/Btu in). and maintenance.Such reinspections will ensure continuation of the • Assumptions about how often the equipment is operated or potential to generate predicted savings and validate stipulations and what load it serves. Examples are operating hours of lights or prior savings estimates. fixed-speed fans and air-conditioning loads(tons)or heater loads(Btu). 4.1.3 M&V Option B: Retrofit Isolation—All Parameter Measurement The most straightforward application of engineering models involves Option B,as with Option A,involves project or system-level(end-use) using savings algorithms that calculate energy use for the subject M&V assessments with performance and operational parameters end use(e.g.,cooling system or lighting system).Savings are then measured at the component or system level.Option B also involves estimated by changing the model parameters that are affected by procedures for verification activities that are the same as Option A. program participation.With Option A,at least one of the key model In addition,savings calculations,as with Option A,involve the use of parameters must be measured.The parameters not measured are engineering models. However,unlike Option A,Option B does not stipulated based on assumptions or analysis of facility historical data allow stipulations of any major factors that would have a significant or manufacturer's data on the affected baseline and/or project equip- influence on energy or demand savings. ment.It is appropriate to use a stipulated factor only if supporting data demonstrate that its value is not subject to fluctuation over Thus,Option B requires additional and often longer-term measurements the term of analysis and it is demonstrably applicable to the project. compared to Option A.These include measurements of both equip- ment operating characteristics(as may be required under Option A) This option and Option B are best applied to programs that involve and relevant performance factors(which may not be required under retrofitting equipment or replacing failed equipment with efficient Option A).Commonly measured parameters include operating hours models.All end-use technologies can be verified using Option A or B; for lighting and HVAC equipment,wattage for lighting and HVAC however,the validity of this option is considered inversely propor- equipment,and flow rates and pressure for various compressed-air tional to the complexity of the measure and the variability of its applications.Spot or short-term measurements may be sufficient savings(e.g.,Option A is not a very reliable option for energy to characterize the baseline condition.Short-term or continuous management system retrofits that involve complex building, measurements of one or more parameters take place after project environment,user,and operator interactions).Thus,the savings installation to determine energy use during the reporting period. 4-5 www.seeaction.energy.gov December 2012 All end-use technologies can be verified with Option B,but determining use Options B,C,or D as needed).For programs targeting integrated energy savings using Option B can be more difficult than doing so whole-building approaches to energy efficiency,utility bill analysis with Option A.And,as noted above for Option A,the difficulty and can be used to statistically evaluate persistence.One useful tool for cost increase as measurement complexity and savings variability this purpose is the U.S. Environmental Protection Agency's(EPA's) increase.The savings,however,are typically more reliable than ENERGY STAR' Portfolio Manager(see sidebar on the next page). those determined with Option A. It should be noted that the term billing analysis is often used generically 4.1.4 M&V Option C:Whole-Facility Analyses to describe any analytic methodology used to determine project or Option C involves use of whole-building meters or submeters to program energy savings based on the use of the energy consumption assess the energy performance of a building or facility.These meters data contained in consumer billing data.It compares billing data are typically the ones used for utility billing,although other meters, from program participants over a period of time before the energy if properly calibrated,can also be used.With this option,energy efficiency measures are installed at customer sites to billing data for consumption from the baseline period is compared with energy a comparable period of time afterward. If used to describe a project- consumption data(usually derived from energy bills)from the based measurement and verification approach,it is equivalent to reporting period.Option C also involves procedures for verification the IPMVP Option C:Whole Facility Analysis.If billing analysis is used to activities that are the same as Option A. describe a program-based evaluation approach,it is comparable to the large-scale consumption data analysis approach,which involves Whole-building or facility-level metered data are evaluated using billing data from both participants and nonparticipants(control group). techniques ranging from simple bill data comparisons to multivariate regression analysis.Option C regression methods can be powerful 4.1.5 M&V Option D: Calibrated Simulation tools for determining savings,while simple bill comparison methods Option D involves calibrated computer simulation models of systems, are strongly discouraged.The latter approach does not account for system components,or whole-facility(usually residential or commercial independent variables such as weather. buildings)energy consumption to determine project energy savings. While in theory simulation can involve the use of any computer For the regression analyses to be accurate,all substantive explana- analysis tools,such as spreadsheets,these calibrated simulations are tory(independent)variables that affect energy consumption need to typically associated with complex building analysis tools that model be monitored during the performance period.Substantive variables heating,cooling,lighting,ventilation,and other energy flows as well may include weather,occupancy schedules,industrial throughput, as water use and sometimes onsite air emissions. Examples of such control set points,and operating schedules.Most applications programs are DOE-2 and EnergyPlus.19 The quality of the savings of Option C require at least 9 to 12 months of monthly baseline estimate depends on how well the simulation models are calibrated (preinstallation)meter data and at least 9 to 12 months of monthly and how well they reflect actual performance. data from the reporting period(postinstallation). All end-use technologies can be verified with Option C. However,this INTERACTIVE FACTORS AND LEAKAGE option is intended for projects in which savings are expected to be large enough to be distinguishable from the random or unexplained Interactive effects are those that an energy efficiency energy variations normally found at the level of the whole-facility measure has on energy use in a facility,but which are meter.The larger the savings,or the smaller the unexplained varia- indirectly associated with the measure.For example, tions in the baseline consumption,the easier it will be to identify reduction in lighting loads through an energy-efficient savings. In addition,the longer the period of savings analysis after lighting retrofit will reduce air conditioning and/or increase project installation,the less significant the impact of short-term heating requirements,since there is less heat generated unexplained variations.Typically,savings should be more than 10% by the energy-efficient lights.When energy efficiency of the baseline energy use so that they can be separated from the programs have interactive effects beyond a single building "noise"in baseline data. and start to affect energy supply and distribution systems, Option C is the most common form of M&V for some multi-measure there can be implications for calculating avoided emissions and other related co-benefits. In this situation of wide-scale building energy efficiency retrofits in the performance contracting interactive effects,the term leakage is used. industry(although ESCOs use Option A as much as possible and then December 2012 www.seeaction.energy.gov 4-6 U.S. EPA!S PORTFOLIO MANAGER 4.2 DEEMED SAVINGS APPROACH Deemed savings values,also called stipulated savings values,are One tool that can be used to analyze facility utility billing estimates of energy or demand savings for a single unit of an meter data is U.S. EPA's Portfolio Manager(PM).Over installed energy efficiency measure that(1)has been developed from 300,000 buildings have been benchmarked with PM,which data sources(such as prior metering studies)and analytical methods provides a consistent framework and metric that building that are widely considered acceptable for the measure and purpose, energy managers can use to track,measure,and monitor and(2)is applicable to the situation being evaluated. Individual whole-building energy use. PM employs a methodology parameters or calculation methods can also be deemed;for example, that is consistent with IPMVP Option C.It aggregates all the effective useful life of a measure or the annual operating hours of meter data from a building so that performance changes can light fixtures in an elementary school classroom.Common sources be assessed at the whole-facility level.Savings are deter- of deemed savings values are previous evaluations and studies that mined at the building level to promote system-wide energy involved actual measurements and analyses. Deemed savings values reductions.Additionally,because the PM approach combines are used for both planning and evaluation purposes. multiple meters,it accounts for differences among fuel types.This is done by converting site meter data into source Deemed savings are used to stipulate savings values for projects with energy consumption.www.energystar.gov/portfoliomanager. well-known and documented savings values.Examples are energy- efficient appliances such as washing machines,computer equipment, and refrigerators,and lighting retrofit projects with well-understood Typically,reporting period energy use data are compared with the operating hours. Many performance contracting projects document baseline computer simulation energy use prediction(using reporting their savings with deemed savings,and it is also a popular evaluation period independent variable values)to determine energy savings, approach for many efficiency programs because of both the relatively although simulation results(reporting period)to simulation results low cost of using deemed savings and the certainty of savings values (baseline)are also used to determine savings. that all parties can rely on for their own purposes. Models are often calibrated by comparing simulation results with The use of deemed values in a savings calculation is thus essentially historical data to ensure that the models have accurately captured an agreement between the involved parties to an evaluation to the operating characteristics of the building. Manufacturer's data, accept a stipulated value or a set of assumptions for use in determining spot measurements,or short-term measurements may be collected the baseline or reporting period energy consumption.With the to characterize baseline and reporting period conditions and operating deemed savings approach,it is increasingly common to hold the schedules.The collected data serve to link the simulation inputs to stipulated value constant regardless of what the actual value is during actual operating conditions.The model calibration is accomplished the term of the evaluation. If certain requirements are met(e.g., by comparing simulation results with end-use or whole-building verification of installation and performance,satisfactory commis- data.Whole-building models usually require at least 9 to 12 months sioning results,and sufficient equipment or system maintenance), of preinstallation data for baseline model calibration. However,these models are sometimes calibrated with only reporting period data so that they can be used with new construction projects for which no BUILDING ENERGY baseline data exist. SIMULATION PROGRAMS Any end-use technology can be verified with Option D if the drop in consumption is larger than the associated simulation modeling error. For about 40 years,engineers and scientists have been This option can be used in cases in which there is a high degree of developing computerized models that describe how the interaction among installed energy systems or the measurement energy use of buildings changes in response to independent of individual component savings is difficult.Option D is commonly variables such as weather.The sophistication and complexity used with new construction energy efficiency programs,where the of these models is quite varied.To learn about some of the baseline is typically modeled using standard practice or building building simulation models that are publicly available,visit code requirements to define what would have occurred without the http://simulationresearch.lbl.gov/resources. efficiency activity. 4-7 www.seeaction.energy.gov December 2012 the project savings are considered to be confirmed.The stipulated to be applied to calculated savings(e.g.,net-to-gross values),source savings for each verified installed project are then summed to documentation,specified assumptions,and other relevant material generate a program savings value. Installation might be verified by to support the calculation of measure and program savings as well as physical inspection of a sample of projects or perhaps just an audit the application of such values and algorithms in appropriate applica- of receipts.Savings can also be verified for"persistence"with tions. For example,a value for operating hours in an elementary periodic inspections that verify that the retrofits are still in place classroom with no summer hours should not be applied to a high and functioning. school classroom with summer sessions. A variant of deemed savings is the deemed savings calculation, As of the date of this guide's publication,there are approximately which is an agreed-to(stipulated)set of engineering algorithm(s) 17 TRMs in use across the United States.These include state and used to calculate the energy and/or demand savings associated with regional TRMs,which are listed in Appendix C.In a recent SEE Action an installed energy efficiency measure.These calculations are devel- report on TRMs,21 it was shown that these resources are very valu- oped from common practice that is widely considered acceptable able,but there are a wide variation in methodologies for estimating for the subject measure and its specific application. It may include savings and the actual values.Some TRMs include information based stipulated assumptions for one or more parameters in the algorithm, on prior year evaluations including,in some cases,rigorous metering but typically it requires users to input data associated with the actual and analysis,and thus these TRMs contain robust(reliable)savings installed measure into the algorithm(s). values. Many others have values based on analyses(e.g.,using com- puter simulations or engineering algorithms),with consideration in The use of deemed savings is quite popular for evaluations of energy their calculations of waste heat factors,in-service rates,and partial consumer-funded programs.The following is from a recent survey of load factors.The transparency and level of detail regarding methods evaluation practices in the United States: and assumptions also ranges from substantial to minimal. We found that nearly all states(36 states,86%)use some type Thus,as would be expected when using any assumptions or stipulated of deemed values in the evaluation framework.In terms of values in an analysis,caution should be used to understand the what types of values are"deemed,"we found 35 states(97%of sources of such values and ensure that the assumptions that went those responding to this question)deem savings amounts for into determining a value are applicable to the situation(e.g.,mea- particular measures,32 states(89%)deem the"lifetime"over sures,measure delivery mechanism,facility types)being evaluated. which to claim savings for particular measures,and 20 states Deemed values,if used,should be based on reliable,traceable,and (65%)deem free-ridership or net-to-gross factors. documented sources of information,such as the following: We also inquired about the source of the deemed values used • Standard tables from recognized sources that indicate the by the states.It appears that there is a lot of"borrowing" power consumption(wattage)of certain pieces of equipment going on within the industry.Twenty-six states(70%)cite the that are being replaced or installed as part of a project use of sources or databases from other states. In nine states, (e.g.,lighting fixture wattage tables) the utilities develop and file certain key deemed values,and in • Manufacturer's specifications two states,the Commission is responsible for developing the • Building occupancy schedules deemed values.In most states(28 states,80%),the results of . Maintenance logs. their own in-state evaluations are used to modify and update deemed values over time.20 In addition,it is good practice to have an ongoing process in place to assess the validity of deemed savings values,such as an annual Deemed savings values and deemed savings calculations are usually or biennial process to update TRMs. In particular,check to see if the documented in a database in formats from spreadsheets to online assumptions used to determine deemed savings values are valid in searchable databases.A term of art for such databases is technical the years after their initial determination(e.g.,the applicability and reference manuals(TRMs).These are resource documents that validity of assumed code requirements or standard practices). include energy efficiency measure information used in program plan- ning and energy efficiency program reporting. It can include savings When using deemed values,it is important to realize that technologies values for measures,measure life information,hourly load shapes of alone do not save energy;it is how they are used that saves energy. savings,engineering algorithms to calculate savings,impact factors Therefore,a deemed energy savings value depends on how and December 2012 www.seeaction.energy.gov 4-8 where a technology is placed into use.For example,a low-wattage and quasi-experimental methods(QEMs).Both involve the use of lamp's savings are totally dependent on its operating hours.Such control groups,but RCTs provide less biased and typically more a lamp installed in a closet will save much less energy than one precise results than QEMs,although they may require more effort to installed in a kitchen. implement. For efficiency programs,RCTs and QEMs are particularly useful for programs in which there are a relatively large number of The example of the residential lamp raises the issue of"granularity" participants with similar characteristics(e.g.,low-income,single-fam- of the deemed savings values.In that example,if an average house- ily houses in a particular city).Currently,these methods are primarily hold's annual operating hours were used to estimate savings,the used for evaluations of residential behavior-based programs and result would be underestimated savings if lamps were only installed whole-house retrofits and weatherization programs. in high-use areas(kitchens)and overestimated savings if lamps were only installed in low-use areas(closets).Thus,the value of deemed 4.3.1 Randomized Controlled Trial (RCT) Methods savings depends not only on the validity of the value used,but on In an RCT,first a study population is defined(such as homes eligible for whether the value is applied correctly(i.e.,it must be based on the a residential weatherization program or a behavior-based program in use conditions as well as the technology). a particular city).Then the study population is randomly assigned to either the treatment group or control group.23 Energy use data must In summary,sources of stipulated values must be documented in be collected for all facilities in both the treatment group and control the evaluation plan. Even when stipulated values are used in place group in order to estimate energy savings. Measured energy use of measurements,equipment installation and proper operation typically comes from utility meter or billing data,often in hourly or are still verified. Properly used,stipulations can be very useful for monthly increments. program planning purposes and can reduce M&V costs,create certainty,and simplify evaluation procedures. Improperly used, The energy savings estimate is then calculated in one of two ways they can give evaluation results an inappropriate aura of authority. (options):(1)by comparing the difference between the measured Deciding whether parameters could be stipulated requires that users energy use of the treatment group households24 and the energy use understand how they will affect savings,judge their effect on the of the control households during the evaluation study period,or(2) uncertainty of results,and balance the costs,risks,and goals of the by comparing the energy use reduction(i.e.,the change in use from program being evaluated. before the study period to after the study period)between the treat- ment households and the control households.Thus,savings from 4.3 LARGE-SCALE CONSUMPTION DATA the program are essentially the difference in changes in energy use ANALYSIS (RANDOMIZED CONTROLLED (positive or negative)between the treatment group and the control TRIALS AND QUASI-EXPERIMENTAL group. Between the two options listed above,the second is prefera- METHODS) ble so that the differences between the control group and treatment As indicated in Chapter 3,a reliable approach for estimating energy group with respect to their differences in energy use before and after savings from efficiency programs is to measure the difference the program implementation(i.e.,differences in differences) between the energy use of those facilities(e.g.,houses)participating are determined. in a program(the"treatment group")and the energy use of a comparison group of nonparticipating facilities(the"control group') Random assignment is a defining aspect of RCTs. It means that each that are similar to those in the participant group during the same facility(e.g.,household)in the potential program participant population period of time.The difference between the energy use ofthe treatment is randomly assigned to either the control or treatment group(i.e., group and the control group facilities can be attributed to those that will be in the program)based on a random probability, three sources: as opposed to being assigned to one group or the other based 1. The true impact of the program on some characteristic of the facility or participant(e.g.,location, energy use,or willingness to sign up for the program).Randomization 2. Preexisting differences between households in the treatment thus eliminates preexisting differences that are observable(e.g., and control group,which is called "bias"or"selection bias" energy use or household floor area)as well as differences that are 3. Inherent randomness.22 typically unobservable(e.g.,attitudes regarding energy conservation, number of occupants,expected future energy use,and occupant There are two generic types of large-scale consumption data age)unless surveyed.Thus, because of this random assignment, analyses discussed in this guide: randomized controlled trials(RCTs) an RCT control group is an ideal comparison group:it is statistically 4-9 www.seeaction.energy.gov December 2012 identical to the treatment group in that there are no pre-existing 4.3.2 Quasi-Experimental Methods (QEMs) differences between the two groups,which means that selection Other evaluation design methods that use nonrandomized control bias is eliminated. Randomized controlled trials can also be used groups are called quasi-experimental methods.With these methods, for various program enrollment options,including opt-in,opt-out, the control group is not randomly assigned.Thus,quasi-experimental and a randomized encouragement design that does not restrict methods often suffer from selection bias that may produce biased program participation.25 estimates of energy savings—sometimes very biased results that result in unreliable savings estimates. However,in some specific It is worth pointing out two net savings aspects that RCTs address: cases in which RCTs are not feasible,quasi-experimental approaches free riders and participant spillover.This is one of the main benefits can still provide reliable results(especially compared with deemed of an RCT over traditional evaluation methods.Thus,if net savings savings values and certain M&V approaches). are defined for a program evaluation as the gross savings plus consideration of free riders and participant spillover, RCTs deliver The sections that follow provide brief discussions of some common estimates of net savings. QEMs. RCTs address the free-rider concern because the treatment and Pre-Post Energy Use Method control groups each contain the same number of free riders through Probably the most common quasi-experimental method is to the process of random assignment to the treatment group or the compare the energy use of participants in the treatment group control groups.When the two groups are compared,the energy after they were enrolled in the program to the same participants'his- savings from the free riders in the control group cancel out the torical energy use prior to program enrollment. In effect,this means energy savings from the free riders in the treatment group,and the that each participant in the treatment group is its own nonrandom resulting estimate of program energy savings is an unbiased estimate control group.This is called a pre-post within subjects,or interrupted of the savings caused by the program(the true program savings). time series design analysis. Participant spillover is also automatically captured by an RCT design. The challenge in using this method is that there are many other An RCT design also addresses rebound effects or take-back during factors(independent variables)that may influence energy use before, the study period,which can occur if consumers increase energy use during,and after the program that are not captured with this (e.g.,as a result of a new device's improved efficiency).Rebound method, resulting in biased savings estimates.Some of these factors, is sometimes,although not often,a consideration in determining such as differences in weather or number of occupants,can be savings persistence. Rebound effects after the study period can be measured and reliably accounted for in the analysis. However,other accounted for with an RCT if the data collection is continued for the factors are less easily observed and/or accounted for. For example, time under consideration. the economy could worsen,leading households and businesses to decrease energy use(even if there were no program);prices for However,free riders and participant spillover are not the only factors energy can change,affecting energy use;and cultural norms could differentiating gross and net savings.The following are some other change—perhaps,say a pop culture icon could suddenly decide to net-to-gross considerations that are not addressed by either RCTs or advocate for energy efficiency.27 the QEMs about to be discussed: • Nonparticipant spillover issues,in which a program influences To minimize bias when using the pre-post calculation method,it is the energy use of non-program participants(although there are necessary to include a regression analysis that discerns and controls some specialized techniques that can determine nonparticipant for the impact of other influences(e.g.,economic recession)that spillover)26 may affect energy use over time compared with the impact of the • Natural gas-related changes in energy use if only electricity efficiency program.Simple comparison before and after energy use consumption is measured is not acceptable. • Long-term market effects such as changes in efficiency product Matched Control Group Method pricing and availability(unless control and treatment groups If it is not possible to create a randomized control group,then savings are maintained over long periods)or the influence of a pro- estimates could be calculated by constructing a nonrandom control gram on establishing an energy efficiency measure as common group made up of participants that are as similar to the treatment practice or part of a code or standard. group as possible.The challenge with a matched control group December 2012 www.seeaction.energy.gov 4-10 method is that participants and their facilities have both observable and other statistical and econometric methods,to analyze characteristics(e.g.,level of energy use,ZIP code,presence of measured energy use data and to control for variations in central air conditioning)that could potentially be matched,as well independent variables. as unobservable characteristics(e.g.,energy attitudes or propensity to opt in to an energy efficiency program)that are harder or With regression analyses,an equation or group of equations that impossible to match. model the relationship between the dependent variable and one or more important independent variables is defined. Dependent Variation in Adoption variables are those that are modeled to be influenced by the inde- The variation in adoption approach takes advantage of variation in pendent variables. Independent variables are the variables that are the timing of program adoption in order to compare the energy use not influenced by other variables in the model,and are assumed to of participating facilities that optin to the program with the energy affect or determine the dependent variables,and are thus the inputs use of facilities that have not yet opted in but will ultimately opt in to an analysis.Independent variables include both the variable(s)of at a later point.It relies on the assumption that in any given month, interest(e.g.,a variable that indicates which customers experience participants that have already opted in and participants that will opt a critical peak pricing event during which hours)as well as control in soon share the same observable and nonobservable characteris- variables that seek to observe and account for other factors that may tics.Thus,this method creates a control group that is very similar to influence the dependent variable(e.g.,the average temperature). the treatment group over time in both observable and unobservable Which independent variables are relevant to calculating energy sav- characteristics,and therefore is likely to result in less bias than ings?Often,this is decided by common sense,experience,program matched control or pre-post methods. characteristics,or budget considerations(with respect to how many variables can be measured and tracked),but it also can be determined Regression Discontinuity Method through field experiments and statistical tests. For weather data,the Among the quasi-experimental methods,regression discontinu- most common independent control variable,there is a wide range of ity typically yields the most unbiased estimate of energy savings. public and private data sources. However,it is also the most complicated method,as it requires knowledge of econometric models and often requires field condi- In the case of energy efficiency analyses,the output of a regression tions that allow the evaluator to use this analytic technique. analysis is a coefficient that estimates the effect of independent Therefore,it is neither common nor always practical.This method variables(e.g.,a program)on the dependent variable(energy or works if the eligibility requirement for households to participate in demand consumption and/or savings).The analysis itself is done a program is a cutoff value of a characteristic that varies within the with a computer model,which can be anything from a spreadsheet population. For example,households at or above a cutoff energy tool to sophisticated proprietary statistical modeling software. consumption value of 900 kWh per month might be eligible to par- ticipate in a behavior-based efficiency program,while those below 4.4 SELECTING AN ENERGY SAVINGS 900 kWh are ineligible. In this case,the households that are just EVALUATION APPROACH below 900 kWh per month are probably very similar to those that Selecting an evaluation approach is tied to objectives of the program are just above 900 kWh per month.Thus,the idea is to use a group being evaluated,the scale of the program,evaluation budget and of households right below the usage cutoff level as the control group resources,and specific aspects of the measures and participants and compare changes in their energy use to households right above in the program.The following subsections describe situations in the usage cutoff level as the treatment group.This method assumes which each of the three generic impact approaches discussed in the that the program impact is constant over all ranges of the eligibility proceeding sections are applicable. More information on planning requirement variable that are used in the estimation(e.g.,that the evaluation activities is included in Chapter 8. impact is the same for households at all levels of energy use). In addition,regression discontinuity relies on the eligibility requirement One criterion that is applicable across all of the approaches is evaluator being strictly enforced. experience and expertise.Thus,a common requirement for selecting an approach is that the evaluator has expertise with the approach 4.3.3 Analysis Techniques selected.A related requirement is that the resources required for All of the methods described above(RCTs and QEMs)use one of that approach—such as the data,time,and budget—are available. a number of different analysis techniques,including regressions 4-11 www.seeaction.energy.gov December 2012 QUALITY The M&V approach is used for almost any type of program that involves retrofits or new construction projects.While a census of The impact evaluation approaches described in this guide projects can be used with the M&V approach,it is generally applied are based on new and unique analysis of energy and demand to only a sample of projects in a program.This is because the M&V savings.Sometimes,however,there is documentation approach tends to be more expensive on a per-project basis than that indicates energy and demand savings were calculated the other two approaches. In general,the M&V approach is applied independently of the subject impact evaluation,such as when the other approaches are not applicable,such as when there information prepared by implementers and/or administra- are no deemed savings values available that are applicable to the tors.Although such documentation was not necessarily given measure or combination of measures,or when per-project prepared per pre-determined evaluation requirements,it results(savings)are needed.An example is a performance-contract- may be sufficient for meeting the evaluation objectives. ing program with multiple contractors. Using existing documentation in combination with quality assurance guidelines(QAGs)can save significant costs for the Because the selection of the M&V approach is contingent on which of program sponsor—and perhaps encourage participation in the four M&V options is selected,Table 4.2 summarizes some selec- the program if a portion of evaluation costs is borne by the tion criteria for each M&V option.Table 4.3 indicates factors that participants.Essentially,a QAG can help determine whether affect the cost of implementing each option.Table 4.4 and Figure 4.1, indicated savings,and the assumptions and rigor used to both from the 2010 version of the IPMVP,indicate key project char- prepare the documentation,can be used in place of a new acteristics that are better for the different options and a flowchart evaluation effort. summarizing the selection of M&V options,respectively.These tables and the figure are included as the last pages of this chapter. 4.4.2 Deemed Savings Approach • The uncertainty associated with savings estimates is low and/or The deemed savings approach is most commonly used for programs the risk of under-(or over-)estimating savings is low.That is, that involve simple new construction,or for retrofit energy efficiency the project's likelihood of success is high. measures with well-defined applications and savings calculations • Documented,reliable,and applicable per-measure stipulated that have been verified with data. Examples might be a residential values are available and applicable to the measure installation second-refrigerator recycling program or a CFL giveaway for residen- circumstances. tial utility customers. In each of these two examples,an assumption • The primary goal of the evaluation is to conduct field inspections would be made about the baseline and energy savings,as well as for all or a sample of projects to ensure they are properly installed perhaps the life of the measures(e.g., 15 years for a refrigerator and have the potential to generate savings(rather than having and 10,000 hours for a CFL).The deemed savings values would have rigorously determined energy savings). to be defined for specific applications(e.g.,second refrigerators in single-family housing)and program delivery mechanisms(point of Assessing a few key aspects of the project can help in making decisions sale or direct install).The deemed savings values might also have a about whether to use deemed savings or deemed calculations. "vintage"in that they are valid for certain years of a program or for Uncertainty in predicted savings,and the degree to which individual certain vintage baseline equipment. parameters contribute to overall uncertainty,should be carefully considered in deciding whether to use stipulations.The"rules of In general,the deemed savings approach is most applicable when all thumb"are as follows: (or at least most)of the following are true: The most certain, predictable parameters can be estimated • There are limited evaluation resources. and stipulated without significantly reducing the quality of the • The projects involve simple energy efficiency measures with evaluation results. well-understood savings mechanisms that have been verified • Stipulating parameters that represent a small degree of uncer- with data,and are not subject to significant variation in savings tainty in the predicted result and a small amount of savings will due to changes in independent variables.The stipulated values do not produce significant uncertainty concerns. not significantly increase the uncertainty of the evaluation metrics. December 2012 www.seeaction.energy.gov 4-12 • Parameters should be measured when savings and prediction uncertainty are both large. • Even if savings are high, but uncertainty of predicted savings is low,full measurement may not be necessary for M&V purposes. 4.4.3 Large-Scale Consumption Data Analysis Approach These approaches are used for programs that have many participants that share many common characteristics,such as single-family detached homes in a particular community with residents of similar economic demographics.These can be equipment retrofits,new construction,or behavior-based programs. Because of the require- ment for a large amount of data,this approach is almost always used with residential programs,such as a weatherization program with thousands of homes being retrofitted with a variety of measures (e.g.,insulation,weather stripping,low-flow showerheads,and CFLs). In general,the large-scale consumption data analysis approach is most applicable to programs that meet most(if not all)of the following criteria: • Participation is well defined(i.e.,the specific consumers or facilities that participate in the program are known). • The program has a relatively large number of participants (i.e.,probably more than 100). • At least one year's worth of baseline and reporting period energy consumption data are available for both the treatment group and the control group. If an RCT method is used, a shorter data period may be adequate. • If an RCT method is not used,then there are observable similarities between participants,or relatively homogenous subgroups of participants can be formed with similar facility and energy efficiency measure characteristics. • Either the program design is such that the target population for the program can be randomly divided into participants and nonparticipants(for randomized controlled trial methods),or sufficient data about the characteristics of the participants are available for selecting an appropriate control group for quasi- experimental methods.This is a particularly important criterion with respect to control group members,as they typically do not have an incentive to provide data to the evaluator. • Independent,third-party evaluators select the treatment and control group,rather than program implementers,to minimize the potential for"gaming." 4-13 www.seeaction.energy.gov December 2012 TABLE 4.2: Applications for Each IPMVP M&V Option OPTION OPTION OPTIONRetrofit Isolation- Key Retrofit Isolation- Whole Facility Calibrated Simulation Parameter Measurement All Parameters is best . . applied is best applied where: Measurement is best applied • The magnitude of savings The project involves simple • The project is complex • New construction projects is low for the entire project equipment replacements . predicted savings are large are involved or for the portion of the Energy savings values (typically greater than 10%) • Energy savings values per project to which Option A per individual measure compared to the recorded measure are desired is applied are desired energy use • Option C tools cannot • The project is simple . Interactive effects are to • Energy savings values cost-effectively evaluate with limited independent be ignored or are stipulated per individual measure particular measures variables and unknowns using estimating methods are not needed • Complex baseline adjust- • The risk of not achieving . Independent variables • Interactive effects are to ments are anticipated savings is low are not complex be included • Baseline measurement • Interactive effects are to • Independent variables that data do not exist or are be ignored or are stipulated affect energy use are not prohibitively expensive using estimating methods complex or excessively to collect difficult to monitor Source:Efficiency Valuation Organization(EVO).(2010).International Performance Measurement and Verification Protocol.IPMVP,EVO 10000-1:2010.www.evo-world.org. TABLE 4.3: Factors Affecting the Cost of Each M&V Option OPTION A OPTION B Retrofit Isolation- Key Retrofit Isolation- OP • OPTION Parameter Measurement All Parameters Whole Facility Calibrated Simulation Measurement i • Number of measurement • Number of points • Number of meters to be • Number of meters to be points and independent installed and/or analyzed installed and/or analyzed • Complexity of deriving variables measured . Number of independent • Number of independent the stipulation • Complexity of measure- variables used in models variables used in models • Frequency of postretrofit ment system • Effort required for inspections • Length of time measure- calibration of models ment system maintained • Frequency of postretrofit inspections December 2012 www.seeaction.energy.gov 4-14 TABLE 4.4: Key Project Characteristics and Applicability for Different M&V Options ECIVI PROJECT CHARACTERISTIC SUGGESTED OPTION Need to assess ECMs individually X X X Need to assess only total facility performance X X Expected savings less than 10%of utility meter X X X Multiple ECMs X X X Significance of some energy driving variable is unclear X X X Interactive effects of ECM are significant or unmeasurable X X Many future changes expected with measurement boundary X X Long-term performance assessment needed X X Baseline data not available X Nontechnical persons must understand reports X X X Metering skill available X X Computer simulation skill available X Experience reading utility bills and performing regression analysis available X Source:Efficiency Valuation Organization(EVO).(2010).International Performance Measurement and Verification Protocol.IPMVP,EVO 10000-1:2010.www.evo-world.org.In this table,ECM stands for energy conservation measure,which is equivalent to energy efficiency measure as used in this guide. 4-15 www.seeaction.energy.gov December 2012 FIGURE 4.1: M&V option selection flowchart START ECM FACILITY PERFORMANCE PERFORMANCE Measure Facility or ECM Performance? Able to isolate NO ECM with meter(s)? YES Expected NO Savings>10%? YES Need to separately assess each ECM? Need full NO YES performance demonstration? NO YES Analysis of If main meter Install isolation Install isolation meters data Simulate meters for all for key parameters, system or parameters and assess interactive facility assess interactive effects, and estimate effects well known parameters Obtain calibration data Missing baseline YES or reporting Calibrate period data? simulation Missing baseline or NO reporting period data? Simulate with and without YES ECM(s) NO OPTION B OPTION WholeRetrofit Isolation: Retrofit Isolation: All Parameter Key Parameter . -. Measurement Measurement Simulation Source:Efficiency Valuation Organization(EVO).(2010).International Performance Measurement and Verification Protocol.IPMVP,EVO 10000-1:2010.www.evo-world.org.In this table,ECM stands for energy conservation measure,which is equivalent to energy efficiency measure as used in this guide. December 2012 www.seeaction.energy.gov 4-16 Chapter 4: Notes 13"Uniform Methods Project."(2012). U.S.Department of Energy. 21 State and Local Energy Efficiency Action Network.(June 2011). www.eere.energy.gov/deployment/ump.html. Scoping Study to Evaluate Feasibility of National Databases for 14 NAESCO.www.naesco.org. EM&V Documents and Measure Savings. Prepared by Jayaweera,1; Haeri, H.; Lee,A.; Bergen,S.; Kan,C.;Velonis,A.;Gurin,C.;Visser,M.; 15 Efficiency Valuation Organization(EVO). International Performance Grant,A.; Buckman,A.;The Cadmus Group Inc.www.eere.energy. Measurement and Verification Protocol(IPMVP).(multiple dates). ov seeaction dfs emvsco in databasefeasibilit g / /p / p g_ Y•pdf. www.evo-world.org.The Efficiency Valuation Organization(EVO)is a non-profit organization that also offers a Certified M&V Professional 22 Randomness can be a factor in these analyses. For example,it Certificate Program. could be that for the chosen population,during the time interval the energy use was monitored,the households in the treatment 16 Federal Energy Management Program(FEMP).(April 2008). group randomly happened to lower their energy use at around the M&V Guidelines:Measurement and Verification for Federal Energy same time that the program started.The precision of an estimate of Projects.wwwl.eere.energy.gov/femp/pdfs/mv_guidelines.pdf. energy savings quantifies the effect of this inherent randomness and Other related M&V documents can be found at http://mnv.lbl.gov/ allows us to decide whether it is a problem or not. Bias and preci- keyMnVDocs/femp. sion,as well as the general subject of certainty of savings estimates, 17 American Society of Heating, Refrigerating and Air-Conditioning is discussed in Section 7.4. Engineers(ASHRAE).(2002).Guideline 14:Measurement of Demand 23The control and treatment groups could contain equal sizes of and Energy Savings.www.ashrae.org.Also ASHRAE.(2010). households,or the control group could be bigger or smaller than Performance Measurement Protocols for Commercial Buildings. the treatment group. It is only necessary to keep a control group www.ashrae.org/standards-research--technology/special--project- that is sufficiently large to yield statistical significance of the savings activities.These documents provide a standardized,consistent set of estimate. protocols for facilitating the comparison of measured energy,water, and indoor quality performance of commercial buildings. 24 Because households are the most common example,the rest of this section uses households as the example unit of analysis. 18 Spot measurements are one-time measurements,such as the power draw of a motor.Short-time measurements might take place 25 State and Local Energy Efficiency Action Network.(May for a week or two,such as to determine the operating hours of lights 2012).Evaluation, Measurement and Verification(EM&V)of in an office.Continuous measurements,as the name implies,involve Residential Behavior-Based Energy Efficiency Programs:Issues and measuring key factors such as power consumption or outdoor Recommendations. Prepared by Todd,A.;Stuart, E.;Schiller,S.; temperature throughout the term of the evaluation,which may be Goldman,C.; Lawrence Berkeley National Laboratory.wwwl.eere. years. Utility billing meters provide continuous measurements and energy.gov/seeaction/pdfs/emv_behaviorbased_eeprograms.pdf. are the typical measurements used with IPMVP Option C.The use A complete source of information on use of RCTs(and QEMs)for of longer-term measurements can help identify under-performing evaluation of energy efficiency programs,particularly behavior-based energy efficiency projects,which in turn can lead to improvements in residential programs. Much of the text in this section is from that their performance. report. 19 DOE-2 information can be found at http://appsi.eere.energy.gov/ 26 it is possible to explicitly conduct an experiment to determine build ings/tools_directory/alpha_list.cfm and EnergyPlus information the spillover effects. For example,an experiment could observe the can be found at http://appsl.eere.energy.gov/buildings/energyplus/ impacts of the intervention for households in the experiment as well energyplus_about.cfm.An extensive list of building energy simula- as others and compare the impacts in two or more communities. tion programs can be found at:http://appsl.eere.energy.gov This type of experiment is often used in medicine and epidemiology. build ings/tools_directory/alpha_list.cfm. 27 However,in some programs,such as critical peak pricing or critical 20 Kushler, M.; Nowak,S.;Witte, P.(February 2012).A National peak rebates,a pre-post method may be less biased.This is because Survey of State Policies and Practices for the Evaluation of Ratepayer- with such programs the experimental factor(e.g.,the call to curtail Funded Energy Efficiency Programs.American Council for an Energy- electricity use or a price signal)can be presented repeatedly so that Efficient Economy(ACEEE). Report Number U122.www.aceee.org/ the difference between the energy use when it is present and when research-report/u122. it is not present is observable. 4-17 www.seeaction.energy.gov December 2012 Chapter 5 Determining Net Energy Savings This section describes an important, and perhaps the most controversial of, impact evaluation topics—the determination and use of net savings. Net (energy and/or demand) savings are the changes in energy consumption or demand that are attributable to an energy efficiency program.The net-to-gross (NTG) ratio is the portion (it can be less than or greater than 1.0) of gross savings (those that occur irrespective of whether they are caused by the program or not)that are attributed to the program being evaluated. Determining net savings involves separating out the impacts that To help understand these terms,the following are some of the are a result of influences other than the program being evaluated, related definitions: such as consumer self-motivation or effects of other programs.The Free rider:a program participant who would have implemented controversy of net savings can be summarized as follows: the program's measure(s)or practice(s)in the absence of the Given the range of influences on consumers'energy consumption—and program.Free riders can be(1)total,in which the participant's the complexity in separating out both short-term and long-term activity would have completely replicated the program's market effects caused by the subject programs(and other pro- intended actions; (2)partial,in which the participant's activity grams)—attributing changes to one cause(i.e.,a particular program) would have partially replicated the program's actions;or(3) or another can be quite complex. deferred, in which the participant's activity would have partially or completely replicated the program's actions,but at a future This controversy is compounded by a lack of consensus by energy time beyond the program's time frame. efficiency policymakers and regulators as to which short-term and Spillover(participant and non-participant):the reduction in long-term market influences and effects should be considered when energy consumption and/or demand caused by the presence determining net savings and the role of net savings in program of an energy efficiency program,beyond the program-related design,implementation,and "crediting"of savings to program gross savings of the participants and without financial or administrators. technical assistance from the program.There can be participant and/or non-participant spillover.Participant spillover is the The following subsections start with classic energy efficiency industry additional energy savings that occur when a program participant definitions of net savings,including the commonly used factors that independently installs incremental energy efficiency measures differentiate net from gross savings,and a discussion of the uses of or applies energy-saving practices after having participated in net savings determinations.Next is a description of the approaches the efficiency program as a result of the program's influence. used to determine net savings.The final subsection discusses some Non-participant spillover refers to energy savings that occur of the issues and nuances of net savings. when a program non-participant installs energy efficiency 5.1 DEFINITIONS AND USES OF NET SAVINGS measures or applies energy savings practices as a result of a program's influence.Sometimes the term free drivers is used 5.1.1 Defining Net Savings for those who have spillover effects. The energy efficiency community agrees on the basic definition of Market effect:a change in the structure of a market or the net savings:the total change in energy use(and/or demand)that is behavior of participants in a market that is reflective of an attributable to an energy efficiency program.However,as noted in increase(or decrease)in the adoption of energy-efficient a recent regional scoping paper on net savings,the operational products,services,or practices and is causally related to market definition of net savings—essentially what factors are considered intervention(s)(e.g.,programs). Examples of market effects when determining net savings—is not unanimously agreed to,with include increased levels of awareness of energy-efficient tech- different factors being applied in different programs and jurisdictions.28 nologies among customers and suppliers,increased availability Factors that cause the difference between net and gross savings, of efficient technologies through retail channels,reduced prices implicitly or explicitly,include free ridership,participant and non- for efficient models,build-out of efficient model lines,and—the participant spillover,and induced market effects.These factors may end goal—increased market share for efficient goods,services, be considered in how a baseline is defined(e.g.,common practice) and design practices. and/or in adjustments to gross savings values. December 2012 www.seeaction.energy.gov 5-1 In terms of how different jurisdictions define net savings,and which FIGURE 5.1: Single program evaluation components of the above factors are included,a 2012 American Council for an Energy-Efficient Economy study indicated the following:29 "I'lllql We asked states what they used when they report their energy savings results,and found that 21 states(50%)said they reported net savings,12 states(29%)said gross savings,and 9 states(21%) said they report both(or use one or the other for different FREE purposes).We explored the net savings issue in a little more RIDERS detail,and asked whether states made specific adjustments XXAOS for free riders and spillover.Interestingly,while 28 states(67%) indicated they make an adjustment for free riders,only 17 states SPILLOVER (44%)make an adjustment for free drivers/spillover. MARKET EFFECTS It is important to recognize that the study survey did not specify any particular definition of what qualifies as net or gross savings.Rather, they allowed states to categorize their own approach. From a summary paper of evaluation practices,the traditional eval- FIGURE 5.2: Multiple program evaluation overlaps uation approach to net savings assumes a measurement boundary around a program,as shown in Figure 5.1.30 The participant savings, which are the documented savings from the efficiency measures installed through the program's transactions with the customer, are shown in the shaded circle.Any of those savings attributableFREI to free riders are subtracted out(inner circle),leaving the shaded donut of net participant savings. If there are spillover savings,either Or I& FREE F RIDERS from additional measures installed by participants without program incentives or from measures installed by non-participants who were indirectly influenced by participants'actions,these may be added to the total. Finally,as the programs begin to affect the market(e.g., RIDERS by inducing retailers to sell only efficient equipment in response to market demand),there may be additional savings(i.e.,the market Fr FREE SPILLOVER Ai RIDERS effects).These additional savings would expand the shaded area of countable program savings. SPILLOVER FREE MARKET EFFEC_[S _A RIDERS However,as also pointed out in the above-referenced paper,with the evolution of integrated portfolios of programs targeted at broad and deep savings goals,the situation looks more like Figure 5.2.Note that the overlaps may,in practice,be even greater than this simpli- fied diagram would suggest.Often,the participant savings overlap FREE V between program offerings,and one program's participants may i RIDERS be another program's free riders.The overlaps can also occur over time,with one program's influence extending to the next generation SPILLOVER of programs.The point of the diagram is to suggest that the simple MARKET EFFECTS evaluation strategy of drawing a boundary around each program, over a single program cycle,will encounter the problems of multiple program influences that can occur before,during,and after the Source:Mahone,D.;Hall,N.(August 2010)."Pursuit of Aggressive Energy Savings Targets—Aligning Program Design and Evaluation."2010ACEEESummer Study Pro- subject program's implementation.This can lead to biased estimates ceedings;August 15-20,2010,Pacific Grove,California.Washington,D.C.:American Of net savings,either higher or lower than actual. Councilforan Energy-Efficient Economy.http://0eep.org/uploads/EMV%20Foruml Steering%20Committee%2ONotesIDMahone_ACEEE Paper for SC Notes.pdf. 5-2 www.seeaction.energy.gov December 2012 Implications of these complexities are discussed after the following The following considerations about net savings are important to subsections on the uses of net savings and the methods used to define for an evaluation effort,but also can be a potential source of determine net savings. controversy due to the different perspectives and objectives of those who will use the evaluation results: 5.1.2 Importance and Uses of Net Savings 1. Which factors should be included in the definition of net Generally speaking,net savings are of most interest for regulated savings(e.g.,free ridership,spillover,and/or long-term market government and utility programs.In these cases,the responsible art e. effects/market transformation) party( g.,a city council or utility regulator)wants to know if the use of public or energy consumer-funded programs are actually having 2. Whether common methods of determining net savings do or an influence,and thus their efficiency investments are"wise! That is, do not properly account for these factors "Are the programs of interest providing incremental benefits,or do 3. Whether and how net savings can be used in a retrospective the benefits result from some other influences?"For example,in the manner,in particular for determining attribution and if an case of programs funded by utility customers(the most common administrator has or has not met its savings goals. situation where the issue of free riders comes up),why would one group of consumers subsidize the efficiency actions of other consum- These first two issues are discussed after the following subsection on ers that would have taken the action anyway?Or,as another example, methods used to determine net savings.The last issue above is simi- the environmental benefits of energy efficiency programs are usually lar to the issue of how savings estimates are applied,as discussed in considered valid only if they are additional to naturally occurring Section 8.3.5. efficiency activities(i.e.,based on some version of net savings). 5.2 APPROACHES FOR DETERMINING In contrast,there are other situations where gross savings are of the NET SAVINGS most interest.For example,private sector energy efficiency performance Before describing the methods used to determine net savings,it contracts are a case where gross energy savings are the primary is important to understand that beyond just defining what factors concern.Also,from a"bottom line"resource and environmental are considered when defining net savings,the actual calculation perspective,it may not be relevant what exactly caused a change of net energy and demand savings can be more of an art than a in energy consumption,only that it is occurring or will continue to science.Essentially,one is attempting to separate out the influence occur.Table 5.1 summarizes different"public"audiences and their of a particular energy efficiency program(or portfolio)from all the current uses of net savings estimates. other influences—such as self-motivation,energy prices,and other From the net savings scoping paper referenced earlier in this section, efficiency programs—that determine participant and non-participant as well as other sources,there are at least four uses of net savings behavior and decisions.With the increasing"push"for energy that do not appear to be controversial,paraphrased below: efficiency by utilities and governments at the local,state,and • Using net savings, prospectively,for program planning and national level and by private groups and large companies,it can be design (e.g.,for setting consumer incentive levels). quite difficult to separate out how one particular program among all this activity influences consumer decisions about whether,when, • Assessing the degree to which programs cause a reduction in and to what degree to adopt efficiency actions. energy use and demand—with net savings as one of numerous measures that should be given serious consideration in the As indicated in Chapter 3,net savings evaluation methods can be assessment of program success. categorized as follows: • Obtaining insight into how the market is changing and trans- Stipulated net-to-gross(NTG)ratios.These ratios are multiplied forming over time by tracking net savings across program by the gross savings to obtain an estimate of net savings and years and determining the extent to which free ridership and are based on historical studies of similar programs.Sources of spillover rates have changed during the period—and potentially stipulations can cover a wide range,from simply using"negoti- using this insight for defining when,and how,to implement ated guesses",to historical values(perhaps based on prior year a program exit strategy. NTG studies),to structured expert judgment panels. • Gaining a better understanding of how the market responds Self-reporting surveys.Information is reported by participants to the program and how to use the information to inform and non-participants without independent verification or modifications to program design, including how to define review. Respondents are simply asked if they would have eligibility and target marketing. December 2012 www.seeaction.energy.gov 5-3 TABLE 5.1: Audiences for Net Energy Savings ESTABLISHED AUDIENCE EMERGING USE OF NET SAVINGS ESTIMATES AUDIENCE Energy efficiency Established(in some areas, • Assess if program achieved savings goals program administrators legislators and advocacy • Identify strong and weak areas of program design and redesign and planners;energy groups are emerging program accordingly regulators;legislators; audiences) • Apply strong program designs for other products,in other advocacy groups jurisdictions • Adjust payments to/funding of programs based on goal achievement • Determine if the ratepayer/taxpayer funds are being spent cost-effectively and wisely • Define program administrator financial incentives and/or cost recovery levels Air regulators Emerging • Will apply emission factors to energy savings to estimate greenhouse gas and other avoided emission reductions • Assess degree to which efficiency programs have achieved greenhouse gas reduction and other avoided emissions targets Source:Modified version of information provided in NMR Group,Inc.;Research Into Action,Inc.(November 2010)."Net Savings Scoping Paper."Submitted to Northeast Energy Efficiency Partnerships:Evaluation,Measurement and Verification Forum.http://neep.org/uploads/EMV%20Forum/EMV9lo20Products/ FINAL%20Net%20Savings%20Scoping%20Paper%2011-13-10.pdf. undertaken the action promoted by the program on their own monetary compensation for information on their recent(e.g., without the incentive(free ridership).Then,they are asked last 50)projects.Details requested would include manufacturer, whether they had undertaken additional energy efficiency efficiency levels,size,price,installation date,installation location actions(purchased products or made a behavioral change) (ZIP code),whether the project was eligible for energy program as a result of their participation in the program(participant incentives,whether energy incentives were received,whether spillover).Through non-participant surveys, respondents are any other incentives were received(e.g.,tax credits or manu- asked if they had recently undertaken energy efficiency actions facturer rebates),and an assessment of the program's impact and if those actions were undertaken as a result of the utility on incented and non-incented efficiency actions.Trade allies program(s)(non-participant spillover). would include both program participants and non-participants. • Enhanced self-reporting surveys.The self-reporting surveys This approach,while not currently common,can yield reliable are combined with interviews and documentation review and information on standard market practices,and—through an analysis.The survey instruments themselves include more ongoing annual update—provide context for tracking ongoing enhanced batteries of questions(e.g.,would you have taken program impacts or market effects. exactly the same action as promoted by the program or would large-scale consumption data analysis approaches(randomized you have undertaken the action at the same time?). controlled trial methods and quasi-experimental methods). • Panel of trade allies.A significant number of trade allies(e.g, Statistical models are used to compare energy and demand contractors,retailers,builders,and installers)are offered patterns of a group of participants and a control group.Where 5-4 www.seeaction.energy.gov December 2012 a control group of non-participants is used,the savings indicated Each of these methods comes with its own data collection are"net"of free riders and participant spillover. and measurement challenges. Particularly when dealing • Cross-sectional studies.These studies consist of comparisons of with subjective indicators,such as why someone chose to market share of targeted technologies or behaviors between a implement an energy efficiency measure,care should be taken baseline area not served by the program and the area served by in not only obtaining the needed data but also in reporting it the program.The main disadvantage of this type of study is diffi- with appropriate indications of the quality and reliability of culty in obtaining quality data.Also,as energy efficiency programs results obtained from the data. become more prevalent,finding control areas(areas without similar program activities)is becoming exceedingly difficult. In terms of accuracy requirements,the challenge in surveying comes from the nature of collecting both qualitative and quantitative • Top-down evaluations(or macro-economic models).Top-down data from various participants and non-participants involved in the evaluations use state, regional,or national data at the sector decision to install energy efficiency measures.Another uncertainty level to assess the extent to which markets for energy-efficient challenge in surveying is the subjective nature of assigning NTG ratios products and services have been affected by programs(See to each participant—their free ridership and participant spillover Appendix B). "score"A participant is a"total"free rider if he or she would have The following are some general notes on these methods: absolutely installed the exact same project at the exact same time, at the same price,even if the program did not exist—and they • The most commonly used methods are the survey-based self- know that.Assigning NTG ratios to individual participants is more reporting,stipulation,and the large-scale consumption data complicated,however,in cases where the participant had multiple analysis approaches. reasons for making the project decision,might have installed a • Net savings values can be the output of(most)large-scale data different project,or would have installed it in two years if not for the analysis and top-down evaluations.With the other impact program—or all of the above.Table 5.2 shows an approach that one analysis approaches(M&V and deemed savings),the net evaluator used to define full,partial,and deferred free riders. savings correction is calculated independently. • All of these methods can address,to one degree of reliability The following subsections discuss the more common NTG determi- or another,free ridership. However,the ability of the methods nation methods:surveys and stipulation of the NTG ratios. to address spillover is limited in terms of how the method is 5.2.1 Self-Reporting Surveys applied and whether non-participants as well as participants Surveys can be complex to design and administer. Respondents' are included in the analyses.Market effects can only be analyzed perception and understanding of the questions is absolutely critical if the studies are conducted over a long period of time,and if to the success of the inquiries.Surveying approaches have become such effects are actually"looked for."When selecting a method, somewhat standard practice with the guidelines developed for it is very important to define what factors are included in the Massachusetts and the self-report guidelines developed by the definition of net savings and whether a selected method can California Public Utilities Commission's Energy Division.31 actually and reliably address all of the factors. • With respect to program size and scale,the survey methods, The survey approach is the most straightforward way to estimate stipulation,and cross-sectional studies can be used with any free ridership and spillover. It is also the lowest cost approach.It program,regardless of the number of participants. However, does,however,have its disadvantages, regarding potential bias the top-down and large-scale consumption data analysis and overall accuracy. For example,typical responses such as"don't approaches can only be used with programs that have a large know,"missing data,and inconsistent answers are very hard to numbers of participants.This is because the models need large address without additional data collection.While there are ways to amounts of data to provide reliable results. improve survey quality(e.g.,using techniques like adding consistency • In terms of timing of net savings analyses,it is preferable to check questions and adjusting the individual's estimate accordingly), analyze a long period of time to address spillover effects. the accuracy of simple self-reports is typically marginal. Conversely,the free ridership reporting accuracy is probably One of the elements that should be addressed in surveys is self- highest when the inquiry is made as close as possible to the selection bias.Self-selection bias is possible whenever the group actual energy efficiency action. being studied has any form of control over whether to participate December 2012 www.seeaction.energy.gov 5-5 TABLE 5.2: Example of Assigning Free Ridership Percentages WOULDFEE- ALREAD • IP Y INSTALLED SAME INSTALLED AD WITHOUTRIDERSH ORDERED OR SCORE • • PROGRAM MEASURES 100% Yes Yes — — — — 0% No No — — — — 0% No Yes No — — — 50% No Yes Yes Yes Yes Yes 25% No Yes Yes Yes No Yes 25% No Yes Yes Yes Yes No 0% No Yes Yes Yes No No 25% No Yes Yes No Yes Yes F12.5% No Yes Yes No No Yes 12.5% No Yes Yes No Yes No 0% No Yes Yes No No No Source:Courtesy of The Cadmus Group,Inc.It is also safe to assign a score of 100%free ridership to those that had already installed the measure prior to receipt of rebate. in the survey;for example, people who have strong opinions or equipment?"For a response,participants might choose between substantial knowledge may be more willing to spend time answer- "definitely would have "probably would have "probably would ing a survey than those who do not.Self-selection bias is related to not have,"and"definitely would not have."This use of a scale,rather sample selection bias and can skew the results of a NTG analysis that than a yes/no response,is thought to allow greater confidence and is not very well planned,funded,and/or executed. precision in the estimate.Based on the responses to the various questions,each response is assigned a free-ridership score.These Another form of survey bias is response bias:the tendency of respon- estimates are then combined(additively or multiplicatively)into an dents to gauge their responses to conform to socially acceptable values. individual participant free-rider estimate.The participant estimates This issue is well recognized in social sciences and is discussed in a are subsequently averaged(or assigned a weighted average based on vast body of academic and professional literature.Another aspect of expected savings)to calculate the overall free-ridership estimate. response bias is construct validity,which raises questions about what the survey results actually measure.The problem stems from the 5.2.2 Enhanced Self-Reporting Surveys fact that while survey respondents,by virtue of their participation in To improve the quality of NTG ratios drawn from self-reported survey the program,are predisposed to efficiency,it is not clear to what extent responses,the evaluation can rely on multiple data sources for the their responses are conditioned by the effects of the program itself. decision to install or adopt energy efficiency measures or practices. Some common additional data sources and techniques include Generally,the best means for implementing self-reporting surveys the following: have involved asking a series of questions,with each question allow- personal surveys.Conducting in-person surveys is probably the ing a scale of responses.A typical initial question asked of participants best way to qualitatively improve the quality of self-surveys. is,"If the program had not existed,would you have installed the same 5-6 www.seeaction.energy.gov December 2012 Key participants in the decision to install efficiency measures sets the value. It is the simplest approach,but one with a high can help determine the level of influence of the program on potential for inaccuracy relative to other approaches that involve participants and non-participants. For commercial and gov- some level of data collection and analyses. ernment facilities, potential interviewees include managers, engineers,and facilities staff.Contractors,design engineers, Sources of stipulated NTG ratios include evaluations of similar programs, and product manufacturers,distributors,and retailers can also hopefully applied to similar populations with a similar level of effi- provide information on the influences and motivations that ciency adoption and during a time period similar to that of the program determine the role of energy efficiency programs in the being reviewed.Other sources use historical or other information decision-making process.When working with professionals from a wide range of sources to develop a"weight of evidence" involved in the efficiency measure installation,individuals conclusion regarding the program's influence. For example,in a familiar with the program and projects should conduct the three-year portfolio cycle,a stipulated NTG ratio may be used for interviews.The interviewer should attempt to eliminate or at the second two years based on a NTG ratio determined with other least minimize any bias they may have. approaches in the first year of the portfolio.One common approach • Project analysis.This consists of two general types of reviews. for developing a stipulated value is to use a panel of experts that have The first is an analysis of the barriers to project installation and relevant technology,infrastructure systems,and market experience.33 how the project addresses these barriers.A common barrier These experts are asked to estimate a baseline market share for a is financial (project costs),so an analysis is done of a project's particular energy efficiency measure or behavior and,in some cases, simple payback. For example,if the project has a very short forecast market share with and without the program in place. payback period without any program-provided benefits,then it 5.2.4 Selecting a Net Savings Evaluation may be considered as more likely to have been installed with or Approach and Timing without the program.32 The other type of analysis is to review As mentioned in Chapter 4,selection of an evaluation approach is any documentation the participant may have of the decision tied to the objectives of the program being evaluated(e.g.,to help to proceed with the project.Such documentation may include understand/improve program design or to adjust savings estimates), internal memos or feasibility studies,and can indicate the basis the scale of the program,the evaluation budget and resources,and of the decision to proceed. specific aspects of the measures and participants in the program. • Market data collection.Through the review of other informa- Another criterion—probably the most important one cited for these tion resources prepared for similar programs,the survey data studies—is the cost of the net savings analysis.The lowest-budget can be adjusted.Such resources might include analyses of approach is to use stipulated NTG ratios,followed by self-reporting market sales and shipping patterns,studies of decisions by par- surveys and enhanced surveys,and then various cross-cutting and ticipants and non-participants in similar programs,and market modeling approaches;although,if the data are available,the top- assessment, potential,or effects studies. Market sales methods down evaluation approach can be quite inexpensive.One option for rely on aggregate data on total sales of a particular technology keeping costs down while using the more sophisticated approaches in a given jurisdiction.They compare this sales volume with a is to conduct an NTG ratio analysis every few years and stipulate NTG baseline estimate of the volume that would have been sold in ratios for the intervening years as long as the market influences and the absence of the program.The accuracy of these methods participants'behavior are relatively consistent. depends on the completeness and accuracy of the sales data as well as the validity of the baseline estimate. 5.3 ISSUES AND NUANCES OF NET SAVINGS As noted above,within the energy efficiency industry,it is agreed All or some of these three data sources can be combined with that net savings determinations are well used for assessing certain written or Web-based participant and non-participant self-surveys to programmatic features such as focusing program designs to maximize triangulate on an estimate of the free ridership and spillover. their efficacy.The net savings controversy is over(1)what factors 5.2.3 Stipulated Net-to-Gross Ratio should be included in the definition of net savings;(2)whether This approach,although not a calculation approach,is often used. common methods of determining net savings do or do not properly NTG ratios are stipulated in some jurisdictions when the net savings account for free ridership,participant and non-participant spillover, value is not considered critical,or if the expense of conducting NTG and long-term market effects(i.e.,market transformation);and(3) analyses and/or the uncertainty of the potential results are whether administrator goal achievement should be based on net considered significant barriers. In such a situation,a regulatory body savings or gross savings. December 2012 www.seeaction.energy.gov 5-7 As summarized in the EM&V Forum"Net Savings Scoping Paper" However,in certain regulatory environments,when free-ridership referenced in Table 5.1,the issues can be summarized in one levels are deemed excessive,program administrators are penalized, word—attribution: claimed savings are discredited,and programs are cancelled.34 Attribution assessment has always involved isolating the effects Beyond the application of free-ridership results,as noted in the of the program from other influences. Increasingly,however, section above on spillover and market effects,the actual determina- when the energy efficiency community mentions the"challenge of tion of free ridership can be difficult if the large-scale consumption attribution"or"sorting out attribution,"it refers to the fact that data analysis approaches of randomized controlled trials and certain reductions in end users'energy consumption can be affected quasi-experimental methods are not used(although these methods not only by myriad efficiency programs offered by a broad range do not separate out participant spillover and free ridership from one of sponsors,but also by economic ups and downs,changes in another and do not include non-participant spillover and long-term energy prices,concerns about climate change,and ongoing market effects).And,as noted in the section on methods,the most advances in technology,among other influences.This situation common method for determination of free ridership is participant has significantly exacerbated the difficulty of establishing causa- surveys—the"self-reports."Self-reports can have problems of tion,and,therefore,of estimating net savings,and it is likely response bias beyond the issue of whether people can state their that this situation will persist.Because of the increased difficulty reasons for undertaking efficiency activities(or,for that matter,any of establishing causation,some commentators in the energy behavior that humans undertake). efficiency community believe that the net savings estimates developed recently are less accurate than those developed in In addition,in areas with long histories of efficiency programs and the past when there were fewer programs and messages activities and many programs operating at the same time,it may not promoting efficiency and"being green." be possible to parse out who is a free rider and who was influenced by the program. In effect,it may be that,in the case of transformed The following are some brief discussions covering some specific net markets or markets being transformed,what is being measured in savings issues that are subsets of this attribution question.These are free-ridership surveys is in fact spillover from other programs.35 presented to not necessarily provide complete answers and recom- mendations,as the solutions tend to be jurisdiction-specific,but to 5.3.2 Spillover and Market Effects point out these issues and suggest that these issues be considered as For many energy efficiency programs,the ultimate goal is market part of the evaluation planning process(see Chapter 8). transformation,where"interventions"are no longer required for the specific measure or practice because market share is such that 5.3.1 Free Ridership adoption is outstripping standard practice.Therefore,it can be a Free-ridership issues are by no means peculiar to energy efficiency; primary evaluation objective for metrics associated with spillover they arise in many policy areas,whenever economic agents are paid and market effects to be assessed. Unfortunately,many net savings an incentive to do what they might have done anyway.However,few determinations only consider free-ridership levels and/or assess issues bring about more discussion in the energy efficiency industry performance for the first year of a program's implementation—too than free ridership. Even the use of the term itself is controversial,as short to assess market effects or long-term spillover benefits. In the way in which it is used in the efficiency industry diverges from its addition,when assessing free ridership,a survey-based method may classic economic definition. be asking the questions of the wrong people.Those identified as free riders might actually be exactly the type of participants that policy- The basic concept speaks to the prudent use of energy efficiency makers would want for a market transformation program,those who dollars:they should be spent to encourage customers to take energy will take the action and continue to do so once the intervention is efficiency actions that they would not otherwise take on their own. over,when the market is transformed.36 It may be that free ridership If program dollars are spent on people who would have taken the is only low in programs(and efficiency actions)that can never result actions anyway,without program support,then those people are in transformed markets.This may be creating a counter-incentive for free riders,and those dollars were perhaps misspent. Evaluators are administrators to implement programs that cannot result in market tasked with studying how much of a program's resources were spent transformation—by not counting the savings from their free riders on free riders,and what the program savings were,net of free riders. and/or spillover—and giving high marks for programs with low free The consequences of free-ridership measurements vary. In some ridership(i.e.,those programs where only an intervention will get cases,the information is used to refine program plans to better target people to implement the efficiency activity). customers and to assess progress toward market transformation. 5-8 www.seeaction.energy.gov December 2012 If overly conservative free-ridership measurements are taken,and 5.3.3 Baselines and Net Savings if free ridership is used to penalize programs,then some program Related to determining free-ridership levels and spillover is the point efforts may be killed prematurely before market transformation that free-ridership savings essentially raise a question about what or ambitious levels of savings are achieved.This also relates to the the correct baseline is—what would have occurred in the absence importance of estimating program spillover and overall market of the program or project. In short,would someone have done this effects.As an efficiency measure is moving up the market transfor- anyway?Some baselines,such as a codes and standards baseline mation"hump,"spillover should increase,as there are more satisfied or existing conditions baseline,may require a free-rider analysis consumers implementing more of the efficiency actions,even after the program(project)is implemented to get participant-by- without the interventions. Indeed,without the snowball effect of participant indications of free ridership.However,if market-based spillover,programs have a large burden to push the measure up and "common practice"was used to define the baseline for calculating over the hump entirely on their own.37 energy savings,the resulting estimates could require no further adjustments for free riders,as the estimates include consideration of Most current approaches to defining net savings do not address what typically would have been done in the absence of the efficiency the benefits that programs have on establishing infrastructures that action.See chapter 7 for more information on baseline selection. allow future efficiency actions(equivalent to a bus passenger being able to be a free rider only because prior programs resulted in the This"common practice"approach to baselines can be used to road and the buses being built and the drivers trained and hired), include consideration of free riders.What"common practice" because program impacts are generally measured over 1-3 years, baselines do not account for is spillover or long-term market effects rather than 5+years. Net-to-gross ratio adjustments,including (i.e.,delayed participation/spillover)created by prior programs. those that account for spillover,typically presume that the measure However,neither does the use of other definitions of baselines. would have been equally available at the same price had no prior Therefore,if common practice baselines(i.e.,net savings)are to programs existed. include spillover and other market effects,additional adjustments are required. The following is an approach to this dilemma,paraphrased from a paper on this subject:38 5.4 SUMMARY An approach for policymakers would be to set market targets The above discussion on the factors associated with net savings (e.g.,a percentage of market share)and perform market studies determination is intended to show the complexities of net savings that track the progress toward increased market-share.There determination.It is not intended to indicate that net savings should is no known"tipping point"percent at which any specific not be determined or that net savings metrics do not have value. technology will be likely to flip to the point of majority adoption As noted previously,there are many valuable uses of net savings (market transformation). But,it is possible to make estimates determinations. of the rate of adoption to determine whether the rate of adop- What is suggested is that the limitations of net savings determina- tion is occurring in a manner that justifies public support,and lion be acknowledged in the policy setting and evaluation planning at what level.This recommendation leaves open the possibility processes,and that,when net savings are determined,there be that incentives may be even more important for later adopters clear indications of what factors are and are not included in the than for early ones.This approach allows program design determination,and over what time frame. Building on the policy decisions about how to adjust the program over time to be and program implementation strategy recommendation mentioned supported by real-time data on market progress and a clear above,with respect to market effects and spillover,is this overall sense of the desired direction for the market through realistic recommendation:39 goal setting and adjusting. With the above points in mind,another conclusion can be that Regulators could establish a series of hurdles,or tests,that program administrators should budget for tracking of market data, a program has to pass to avoid high free ridership.The exact such as sales volumes for specific energy-efficient equipment,price nature of the tests would vary depending on the program,but the amount of the incentive relative to the cost of the measure information,and market saturation indicators.This data collection is a good general gauge.When very low incentives appear fits under the category of market effects studies.These market to attract a large number of participants,or net benefits to effects studies are briefly discussed in Appendix B. December 2012 www.seeaction.energy.gov 5-9 participants are very high,chances are the majority of partici- pants will be free riders. Programs administrators must avoid offering incentives for projects with very short paybacks to participants who most likely would—or should—undertake the project on the project's own financial merits. [With respect to evaluation activities] program administrators would have to monitor energy-efficient product markets closely to see if a transformation has occurred and exit the market when it has. Expected savings and costs of conservation mea- sures should be revised periodically based on actual saturation of energy-efficient products through well-designed and detailed market effects studies. Baselines can also be adapted over time, based on market baseline studies. 5-10 www.seeaction.energy.gov December 2012 Chapter 5: Notes 28 NMR Group,Inc.;Research Into Action, Inc.(November 2010). 35 Haeri,H.;Khawaja,M.S.(March 2012)."The Trouble With "Net Savings Scoping Paper."Submitted to Northeast Energy Freeriders."Public Utilities Fortnightly.www.fortnightly.com/ Efficiency Partnerships:Evaluation, Measurement,and Verification fortnightly/2012/03/trouble-freeriders. Forum. http://neep.org/uploads/EMV%20Forum/EMV%20Products/ FINAL%20Net%20Savings%20Scoping%20Paper%2011-13-10.pdf. 36 Haeri,H.;Khawaja,M.S.(March 2012)."The Trouble With Freeriders.' Public Utilities Fortnightly.www.fortnightly.com/ 29 Kushler, M.;Nowak,S.;Witte,R(February 2012).A National fortnightly/2012/03/trouble-freeriders. Survey of State Policies and Practices for the Evaluation of Ratepayer- Funded Energy Efficiency Programs.American Council for an Energy- 37 Mahone,D.(August 2011)."Free-Ridership as a Way to Kill Efficient Economy(ACEEE).Report Number U122.www.aceee.org/ Programs—How Evaluation Policies Can Frustrate Efficiency Goals." research-report/u122. Boston:International Energy Program Evaluation Conference. www.iepec.org/20llPapersTOC/papers/054.pdf. 30 Mahone,D.; Hall, N. (August 2010)."Pursuit of Aggressive Energy Savings Targets—Aligning Program Design and Evaluation." 38 Peters,J.;McRae,M.(August.2008)."Free-Ridership Measurement 2010 ACEEE Summer Study Proceedings;August 15-20,2010, Is Out of Sync With Program Logic...or,We've Got the Structure Built, Pacific Grove,California.Washington, D.C.:American Council but What's Its Foundation?"2008 ACEEE Summer5tudy Proceedings. for an Energy-Efficient Economy. http:Hneep.org/uploads/ August 17-22,2008,Pacific Grove,California.Washington, DC: EMV%20Forum/Steering%20Committee%20Notes/DMahone_ American Council for an Energy-Efficient Economy(ACEEE). ACEEE_Paper_for_SC_Notes.pdf. www.aceee.org/files/proceedings/2008/data/papers/5_491.pdf. 31 Rathbun,P.;Sabo,C.;Zent,B.(2003).Standardized Methods for 39 Haeri,H.;Khawaja,M.S.(March 2012)."The Trouble With Free-Ridership and Spillover Evaluation—Task 5 Final Report(Revised). Freeriders."Public Utilities Fortnightly.www.fortnightly.com/ Prepared for National Grid,NSTAR Electric,Northeast Utilities,Unitil, fortnightly/2012/03/trouble-freeriders. and Cape Light Compact.www.ceel.org/eval/db_pdf/297.pdf. California Public Utilities Commission(CPUC).(2007).Guidelines for Estimating Net-To-Gross Ratios Using the Self-Report Approaches. CPUC Energy Division Master Evaluation Contractor Team. ftp://ftp.cpuc.ca.gov/puc/energy/electric/energy+efficiency/ ee+workshops/se Ifreportguideli nesd eta iled_v20.pdf. 32 Simple payback is a common metric but it is complicated to apply. For example,what one entity might consider a short payback another might see as a long payback.Also,because the payback is usually calculated based on energy cost savings,it ignores other benefits to participants,such as increase in value of property,more reliability, comfort,and others,and thus can overestimate the payback period from the participant's perspective. 33This is called a Delphi process,which is an interactive forecasting method using a panel of experts. 34 Mahone,D.(August 2011)."Free-Ridership as a Way to Kill Programs—How Evaluation Policies Can Frustrate Efficiency Goals." Boston:International Energy Program Evaluation Conference. www.iepec.org/20llPapersTOC/papers/054.pdf. December 2012 www.seeaction.energy.gov 5-11 Chapter 6 Calculating Avoided Air Emissions State and federal policymakers and utility regulators are broadening the scope of efficiency programs, and thus their evaluation, by focusing on objectives beyond energy savings. Examples of these broader objectives include reducing the need for on-peak generation (demand response), promoting economic development through job creation, reducing greenhouse gas emissions, and achieving a wide range of air quality and health benefits.40 Because avoided air emissions in particular are seen as a significant benefit of energy efficiency actions throughout the United States,this chapter has been included to provide guidance for those interested in documenting these benefits. This chapter first features a summary of the topic of avoided emissions Historically,emission reductions from efficiency activities were usually and end-use energy efficiency,which is followed by a discussion of only described subjectively in program evaluations as a non-quantified special issues associated with this topic.The remainder of the chapter (non-monetized)benefit.This is changing with increasing interest in focuses on ways in which emission factors can be calculated,with quantifying these benefits for at least two purposes: one section covering calculation of emission factors associated with avoided onsite fuel use and the following section covering avoided • Determining the cost-effectiveness of efficiency programs emissions calculations for grid-connected electricity approaches.The (to possibly justify more investment in efficiency)by second-to-last section provides brief summary comments on selecting monetizing the environmental benefits of efficiency a calculation approach,and the last section provides further references • Supporting state claim of emissions benefits in state air on this topic.Also included in this chapter is a discussion of issues pollution plans(e.g.,State Implementation Plans—SIPS—for related to avoided emissions calculations,including additionality, criteria pollutants)or GHG reduction requirements(e.g., boundary area definitions,and the design of cap-and-trade programs. California's Assembly Bill 32:Global Warming Solutions Act42). 6.1 ENERGY EFFICIENCY AND AVOIDED EMISSIONS SUMMARY EFFICIENCY AS . COST-EFFECTIVE Energy efficiency can reduce air emissions associated with the EMISSIONS REDUCTION STRATEGY production of electricity and thermal energy from fossil fuels.Air emissions that can be reduced by efficiency include the six commonly Energy efficiency policies and programs offer the potential to found air pollutants(also known as the U.S.Environmental Protection achieve emissions reductions at a cost that can be lower than Agency[EPA] "criteria pollutants"): particle pollution(often referred traditional control measures.The EPA is gaining experience to as particulate matter or PM),ground-level ozone,carbon monox- with these potentially cost-effective strategies in rulemak- ide,sulfur dioxide,nitrogen oxides,and lead.All of these pollutants, ings affecting the utility and other sectors. For example, as well as others such as mercury,have geographic-specific impacts; a recent EPA modeling scenario for EPA's Mercury and Air therefore,where the emission reduction occurs is critical to deter- Toxics Standard rule predicts that moderate levels of energy mining the benefits of efficiency-induced avoided emissions. demand reduction—equivalent to the continuation of current The other major potential avoided emissions are greenhouse policies—could lower total compliance costs,reduce rate- gases(GHGs)—primarily carbon dioxide(CO2)—from fossil fuel payer bills over the long term,and,in some cases,delay or combustion. Energy efficiency is very important for reducing avoid the need for equipment upgrades or new construction GHGs because there are few options or"controls"for reducing of generating facilities and emissions controls.This energy COz emissions from combustion once the COz is formed,and demand reduction is also likely to reduce emissions of air because,unlike the pollutants mentioned above,the impact of GHG pollutants on high electricity demand days when air quality reductions is not location dependent.Therefore,energy efficiency can be especially harmful. can be the lowest-cost option for reducing GHG emissions.The Source:U.S.Environmental Protection Agency.(2012).The Roadmap for importance of efficiency also becomes clear in light of the fact that Incorporating Energy Efficiency/Renewable Energy Policies and Programs into approximately 60%of all human-induced (anthropogenic)GHG State and Tribal Implementation Plans.wwwepa.gov/airquality/eere. emissions come from energy-related activities.41 6-1 www.seeaction.energy.gov December 2012 The federal government(including EPA and DOE)and many states 6.2 SPECIAL CONSIDERATIONS FOR have recognized the value of incorporating end-use energy efficiency CALCULATING AVOIDED EMISSIONS FROM into air regulatory programs.Several programs—including EPA's EFFICIENCY PROGRAMS Acid Rain Program,EPA's NOx Budget Trading Program,and the This section describes some critical considerations for efficiency Regional Greenhouse Gas Initiative(an effort of nine states from the program evaluators to consider when determining avoided emis- Northeast and Mid-Atlantic regions)—have provided mechanisms for sions.One important consideration for both the avoided emissions encouraging energy efficiency.43 The EPA has also provided guidance calculation approaches listed above is that the net energy savings on incorporating energy efficiency into SIPs44 and has approved the calculated for the purposes of an energy resource program may be inclusion of energy efficiency measures in individual SIPs.45 different from the net savings that need to be calculated to meet the requirements of an avoided emissions program.The following are For any type of energy efficiency program,the avoided air emissions three potential causes of the difference: are determined by comparing the emissions occurring after the • Different definitions of additionality program is implemented to an estimate of what the emissions would have been in the absence of the program(i.e.,emissions under a • Different definitions of boundary areas baseline scenario).Conceptually,avoided emissions are calculated • Characteristics of emissions control mechanisms/regulations using the net energy savings calculated for a program and one of two that may be in place. different approaches: 1. Emission factor approach.This approach involves multiplying The first two items are discussed in Sections 6.2.1 and 6.2.2,respectively. the program's net energy savings by emission factors(e.g., The"cap-and-trade"emissions control mechanism and its attributes pounds of CO2 per MWh) representing the characteristics of with respect to energy efficiency are discussed in Section 6.2.3. displaced emission sources to compute hourly,monthly,or Although it is not the only option to achieve widespread emission annual avoided emission values(e.g.,tons of NOx or CO2 per year). reductions,it is addressed here because of its unique characteristics There are several sources of emission factors,as well as approaches and its current application in the United States for controlling both for calculating the factors.The emission factors approach can criteria pollutants and GHG emissions. be used with any project type or energy resource,such as Following these sections is a brief overview,in Section 6.2.4,of the boiler retrofits that save fossil fuels or lighting projects that possible objectives associated with calculating avoided emissions and save electricity. Below is the basic equation for this approach: how they can affect decisions about which calculation approaches avoided emissions=(net energy savings)x(emission factor) should be used and which specific issues should be addressed. 2. Scenario analysis approach.This approach involves calculating a base case of source(e.g.,electricity generating units connected 6.2.1 Additionality to a grid)emissions without the efficiency programs and Additionality is the term used in the emissions mitigation industry comparing that with the emissions of those sources operating for addressing the key question of whether a project will produce with the reduced energy consumption associated with the reductions in emissions that are additional to reductions that would efficiency programs.This is done with a range of approaches, have occurred in the absence of the program activity.Note that from using historical generation and load data or capacity additionality is typically defined as project-based,versus program-or factor data with emission rates in spreadsheet calculations portfolio-based,and is directly related to the efficiency evaluation to the use of sophisticated computer simulation approaches issue of defining proper baseline conditions and in some cases to using"dispatch models."Scenario analysis is typically only used participant free ridership,as described in Chapter 5.As with baselines, with large-scale,electricity-saving programs. Below is the basic additionality cannot be directly measured and must be inferred from equation for this approach: available information. avoided emissions=(base case emissions)—(reporting While the basic concept of additionality may be easy to under- period emissions) stand,there is no common agreement on the procedures for More information about each of these approaches and their applications defining whether individual projects or whole programs are truly is provided later in this chapter. additional(i.e.,different than a baseline scenario).As such,there is no technically correct level of stringency for additionality rules. Evaluators may need to decide,based on their policy objectives, December 2012 www.seeaction.energy.gov 6-2 what tests and level of scrutiny should be applied in additionality Direct and indirect emissions are two categories for consideration testing.For example,program objectives that focus on formally when setting an emissions assessment boundary. Direct emissions claiming avoided emissions benefits as part of a regulatory program are changes in emissions at the site(controlled by the project may necessitate stringent additionality rules.On the other hand, sponsor or owner). For efficiency projects affecting onsite fuel use programs that are primarily concerned with maximizing energy (e.g.,high-efficiency,fossil fuel water heaters or boilers),the avoided efficiency and only need to approximately indicate avoided emissions emissions are direct. Indirect emissions are changes in emissions may establish only moderately stringent rules. that occur at a source away from the project site(e.g.,of an electric generating unit,or EGLI). Indirect emissions are the primary source One area of common consideration for additionality is whether of avoided emissions for electrical efficiency programs. an efficiency program is mandated by existing legislation(e.g.,an energy efficiency resources standard),regulation(e.g.,a building When defining the assessment boundary,one must also consider code),or policy(e.g.,a government agency's or private company's intended and unintended consequences,also called primary and policy to reduce energy use by a certain percentage through efficiency secondary effects: actions).That such legislation,regulation,or policy was influenced by . primary effect:the intended change in emissions caused by anticipation of an entity being required to meet a future greenhouse a program. Efficiency programs generally have only one gas reduction target or simply to improve air quality would argue for primary effect—energy savings at facilities that consume such efforts as being additional. However,that such efforts will occur energy,translating into avoided emissions. irrespective of additional support from an emissions program would • Secondary effect:an unintended change in emissions caused argue against such an additionality finding.This conflict is illustrative by a program.Secondary effects are sometimes called "leak- of the policy decisions to be made in deciding additionality.46 age"Leakage48 and interactive effects49 are similar concepts, One approach for addressing the additionality conflict is to review although leakage is a more"global"issue,whereas interactive what was assumed in terms of baseline energy use and efficiency effects tend to be considered within the facility where a project activity when a greenhouse gas or air emissions target is set.Levels takes place.Two categories of secondary effects are one-time of efficiency activity included in an air emissions baseline might very effects and upstream and downstream effects: well not be considered additional once an air emissions target is — One-time effects:changes in emissions associated with set based on such a baseline;that is,the level of emissions already the construction,installation,and establishment or the assumes that level of efficiency activity.However,if the efficiency decommissioning and termination of the efficiency projects— activity is beyond what is included in the baseline,then the program net of the same level of efficiency activity in the baseline may be considered additional. scenario. Upstream and downstream effects:recurring changes in 6.2.2 Assessment Boundary Issues: Primary and emissions associated with inputs to the project activities Secondary Effects/Direct and Indirect Emissions (upstream)or products from the project activity(down- The emissions assessment boundary is used to define and encom- stream) relative to baseline emissions. For example,one pass all the energy uses and emission sources affected by activities potential upstream effect of possible concern (however in a program.47 For avoided air emissions,the assessment boundary unlikely)for efficiency programs is that if efficiency programs can be much larger than the boundary for calculating energy and displace energy sales and emissions in one area,the same demand savings,including changes to emission rates and volumes amount of energy consumption and related emissions beyond avoided emissions associated with lower energy use at the might be shifted elsewhere.A scenario for this would be if efficiency project sites.This is particularly important for calculating production at the factory that underwent the efficiency avoided emissions associated with electricity efficiency projects, actions is shifted to a factory outside the boundary area when the affected facilities are connected to the grid. Because power or to one in the boundary with less-efficient energy use. plants are interconnected through the electric grid system,to fully understand electricity-related emissions requires an understanding Secondary effects,outside the facility where the efficiency project of the operation of power plants throughout the system,which takes place,are typically minor relative to the primary effects of usually involves regional considerations.See Section 6.4. energy efficiency programs—particularly when compared to baseline secondary effects. For example,the manufacturing,maintenance, 6-3 www.seeaction.energy.gov December 2012 and installation of energy-efficient motors have no meaningfully The level of the cap is an important aspect of a cap-and-trade different associated emissions than the emissions associated with program. In general,emissions may not exceed the cap,and they are standard efficiency motors. In some cases,however,it is possible that also unlikely to be below the cap during any substantial period of secondary effects could undermine the primary effect;therefore, time.The reason for this is that a unit that emits fewer allowances the emissions assessment boundary should be defined and leakage than it has available may sell those allowances to another unit,which possibilities considered,even if only documented that there are no will then use them to emit.54 significant,identifiable,secondary effects.One way to document such secondary effects would be to conduct a generic study of the relative The fact that capped emissions tend to remain at the cap level is upstream(and other)emissions associated with energy-efficient very relevant to the effect of energy efficiency,which reduces the products versus conventional products and apply the results as a output of electricity generators or perhaps very large industrial boilers "factor"to the calculation of avoided emissions(if such a factor were that could be subject to a cap.When emissions are not capped, reliable and relevant). energy efficiency reduces emissions.As noted,this is not typically true for emissions from sources subject to caps(e.g.,large boilers, In summary,when evaluating the avoided reductions associated with power plants).This is because reductions in the output of electricity efficiency programs,it is important to properly define the assessment generators do not alter the overall cap on emissions from electricity boundary and to account for all primary effects(the intended savings), generators,and any reductions in emissions(and demand for allow- as well as all direct emissions(at the project site)and indirect emissions ances)at a particular generator as a result of energy efficiency make (at other sites).In addition,ideally,secondary effects should be at least extra allowances available for other entities to use.This means that qualitatively investigated. freed-up allowances can be sold in the market and used elsewhere or banked for use in a later year,such that total emissions will remain 6.2.3 Special Issues for Capped Pollutants roughly equal to the cap level. under Cap-and-Trade Programs There are several regulatory mechanisms for controlling pollutants The goal of the cap-and-trade program is typically not to go below the and greenhouse gas emissions,and cap-and-trade is one of them. cap but to achieve the cap at the lowest possible cost to society,so Under a cap-and-trade program,an overall emission tonnage cap is energy efficiency contributes to the primary goal of the cap-and-trade set for an affected sector or set of facilities.Allowances are created to program by helping to achieve the emissions target while minimizing represent the emission of each unit(e.g.,1 ton)of pollution under the compliance costs. In addition,efficiency programs may reduce emis- allowable cap.The primary compliance requirement is that each plant sions from non-capped emission sources and non-capped pollutants must hold allowances equal to its actual emissions at the end of each and directly claim avoided emissions if properly calculated. compliance period.However,there is no fixed emissions cap or limit on an individual facility,and each facility's emissions are not limited to There are,however,mechanisms by which efficiency programs under the allowances that it initially receives or buys at auction(depending a cap-and-trade system can claim avoided emissions for capped on how allowances are allocated). It may purchase additional allow- pollutants.The primary mechanism is that allowances are retired,or ances from another facility or sell allowances if it has a surplus. removed from the market.In states that have established an energy efficiency set-aside program(a pool of allowances from within the There are several examples of U.S.cap-and-trade programs: cap that is set aside and made available for energy efficiency project • The Title IV acid rain sulfur dioxide(SO2)trading program sets and program implementers)as part of their cap-and-trade program a cap on annual SO2 emissions for U.S. power plants.so design,efficiency program implementers may acquire allowances to • The Clean Air Interstate Rule(CAIR)includes a cap-and-trade retire by submitting a claim for allowances to the set-aside adminis- mechanism for power plants in the eastern United States to trator.Alternatively,efficiency programs may acquire allowances to achieve reductions of emissions that cross state lines and retire by purchasing them from other market participants. contribute to ground-level ozone and fine particle pollution For example,some states have created special set-aside allocations in other states.51 of allowances in their NOx trading programs for energy efficiency • CO2 emissions from power plants are capped in the nine states projects.55 Qualified project sponsors that obtain these allowances of the Northeastern Regional Greenhouse Gas Initiative52; in can choose to retire them to make emissions-reduction claims and 2011,California enacted a multi-sector cap-and-trade program avoid the expense of an allowance purchase that would otherwise be to limit GHG emissions.53 necessary to make such claims. However,sponsors may also sell the December 2012 www.seeaction.energy.gov 6-4 allowances to finance the efficiency project,in which case they may Calculating avoided emissions for regulatory purposes or a not claim the reduction. Under CAIR,states have an opportunity to primary program objective.Rigorous analyses are appropriate develop energy efficiency(or renewable energy)set-aside programs. when avoided emissions are a primary goal of an efficiency program—typically,when the efficiency program is part of a Lastly,it should be noted that while efficiency generally does not regulatory scheme or is intended to generate creditable emission reduce aggregate emissions for capped pollutants,efficiency may reductions or offsets with a significant monetary value or to affect the temporal or geographic distribution of emissions in ways comply with a regulatory mandate. In these situations,docu- that bring environmental benefits,as well as possibly reduce the mentation should be provided(either on a project-by-project on-site emissions footprint of the entity undertaking the efficiency basis or,preferably,on a program or policy level)56 that the actions. For example,efficiency programs that target cooling loads energy savings(probably net energy savings)and avoided emis- may reduce electric generating unit NOx emissions on hot summer sions are additional or surplus as defined by the air regulator. days with high ozone concentrations in ways that reduce pollution A boundary definition is also desirable to document that there levels,even though aggregate NOx emissions for the ozone season is no anticipated emissions"leakage."That boundary definition would be unchanged.For more information on the air-pollution properly defines the locations of the avoided emission sources— impacts of energy efficiency under capped systems,visit from the efficiency project sites to within a multi-state electric www.epa.gov/airquality/eere.html. grid. In the case of regulatory mandated air emissions control 6.2.4 Avoided Emissions Calculations for programs,the methods for calculating avoided emissions may Different Objectives be defined. In terms of the actual emissions-reduction calcula- Avoided emissions calculations have a wide range of specific applica- tions,one of the more complex approaches described below tions,such as voluntary and mandatory GHG offset programs and will most likely be appropriate. NOx cap-and-trade programs with energy efficiency allowance set- 6.3 DIRECT ONSITE AVOIDED EMISSIONS asides.These programs have varying requirements for documenting CALCULATIONS legitimate avoided emissions.Those interested in creating tradable offsets,allowances,or other program-specific credits should consult Direct onsite avoided emissions can result when efficiency programs the regulations of the specific program they are interested in with save self-generated electricity that would have been produced at respect to additionality, boundary area definitions,and other issues a project site or,more typically,when efficiency reduces the need specific to the program. for onsite heat or mechanical energy,reducing onsite combustion of natural gas,fuel oil,or other fuels.Identifying the appropriate However,the following are some rule-of-thumb recommendations, emission factor is fairly straightforward for onsite emissions,such organized by objective,for calculating the avoided emissions: as those from residential or commercial combustion equipment, Calculating avoided emissions primarily for informational industrial processes,or onsite distributed generation.The emission • purposes.When the primary goal of an efficiency program is factors are commonly calculated in one of two ways: to save energy or demand,the avoided emissions are often Default emission factors. Default emission factors are based reported only subjectively or with minimal analysis,to indicate on the fuel and emission source being avoided.This is the most a co-benefit of the program. In this situation,the expectations common approach and a wide variety of resources provide for the certainty of the avoided emission values are not high, emission factors per-unit of fuel consumption,including and the avoided emission estimates are not used in a regula- manufacturer's equipment performance data,state-certified tory or market scheme where a monetary value is ascribed to performance data,emission permit data,and generic emission the avoided emissions.Thus,one of the simpler approaches data compiled by regulators or industry groups. described below is probably appropriate. It is typical that(1) A standard resource is the EPA's AP-42,Compilation of Air additionality is simply assumed, (2)emissions boundary area Pollutant Emission Factors.57 It is the primary compilation of the issues are not necessarily rigorously addressed,and(3)the EPA's emission factor information and contains emission factors energy savings are simply those reported for the program, and process information for more than 200 air pollution source whether net or gross.These savings are then multiplied by categories.A standard resource for GHG emissions is the 2006 appropriate, preferably time-dependent,emission factors to IPCC Guidelines for National Greenhouse Gas Inventories.58 calculate avoided emissions. 6-5 www.seeaction.energy.gov December 2012 • Source testing.Source testing can determine the emission Finally,annual avoided emissions is calculated: factors for a specific device(e.g.,large-scale industrial boil- - 125,000 therms/year x 0.005 metric tons CO2/therm- ers). Protocols for testing are available,but given the time and 625 metric tons COz equivalent/year. cost of such testing,this approach is usually only taken when required by environmental regulation.This may change if the This example is provided,in part,to demonstrate that even some- value of avoided emissions makes source testing costeffective thing as simple as an emission factor calculation can be complex as a part of a certification process. and involve more than just reading an emission factor off a chart. It is important to understand the assumptions and applicability of The program evaluator must select onsite emission factors that the factor,such as in the above case for calculating COz emissions provide sufficient accuracy to meet the goals of the evaluation.This (equivalent)from a carbon emissions value.Therefore,it is always requires selecting different emission factors for different time periods, suggested that people familiar with the use of such emission factors places,and technologies. In addition,emission factors based on be consulted when doing this type of analysis. historical emission rates may need to be adjusted to account for new, more stringent regulations.Accounting for changing environmental 6.4 EMISSION FACTORS FOR GRID-CONNECTED regulation is an important consideration in calculating emissions. ELECTRIC GENERATING UNITS The following is an example of an avoided COz emissions calculation As with the direct onsite emissions cases,emission reductions for a project that reduces natural gas consumption from a large from reduced electricity consumption occur because less fuel is industrial boiler. combusted. However,calculating avoided electrical grid emission reductions is more complex because the fuel combustion in question • First,avoided natural gas use is calculated, in units of therms would most likely have occurred at many different existing or per year(1 therm=100,000 Btu): future electric generating units(ECUs)connected to the grid.59Thus, - Displaced steam use due to efficiency project=10,000 emissions from displaced electricity usually depend on the dynamic million Btu(MMBtu)/year interaction of the electrical grid,emission characteristics - Steam boiler higher heating value(HHV)efficiency=80% of grid-connected power plants,electrical loads,market factors, - Displaced natural gas use=10,000 MMBtu/year-0.80-_ fuel and electricity economics,and a variety of regional and environ- 12,500 MMbtu/year=125,000 therms/year. mental regulatory factors—all of which can change over time. • Next,an emission factor is calculated: When using electricity savings values for calculating avoided emis- - The average carbon coefficient of natural gas is 14.47 sions,a critical consideration is to convert the electricity savings at the site when the efficiency measures are implemented to the kilograms(kg)of carbon per MMBtu.The fraction oxidized to COz is 100%(from the Intergovernmental Panel on electricity savings at the EGU.This means taking into account trans- Climate Change[IPCC]). mission and distribution(T&D)losses between the end use(e.g., home,office,factory)and the generator.The difference between - COz emissions per therm are determined by multiplying site savings and savings at the generator(sometimes called source the following: natural gas heat content;the carbon(C) or busbar savings)varies with a wide range of factors,with typical coefficient;the fraction oxidized;and the ratio of the values of 5%-20%. molecular weight ratio of carbon dioxide to carbon(44/12). When using this equivalency,keep in mind that it represents This section of Chapter 6 first provides some basic information about the COz equivalency for natural gas burned as a fuel, not the electric grid and the relationship of efficiency-displaced existing natural gas released to the atmosphere. Direct methane (and perhaps future)electricity generation and avoided emissions. emissions released to the atmosphere(without burning)are This is followed by descriptions of several approaches,from simple about 21 times more powerful than COz in terms of their to complex,for calculating avoided emissions. warming effect on the atmosphere. - MMBtu/1 therm x 14.47 kg C/MMBtu x 44 grams(g) 6.4.1 The Electricity Generation Mix CO2/12 g C x 1 metric ton/1,000 kg=0.005 metric tons The electric grid is composed of a T&D system,often covering multiple CO2-equivalent/therm. states,connecting a mix of generating plants with different emissions characteristics,which operate at different times to meet electricity December 2012 www.seeaction.energy.gov 6-6 demand.The mix of plants operating varies by region and overtime FIGURE 6.1: Example electricity load duration curve within regions—both as the demand changes from one hour to the next and as old plants are retired and new plants are built.A common 18,000 -------------------------------------------- way of looking at this varying generation mix is a load duration curve. 16,000 ------------------------------------ The load duration curve shows the electricity demand in MW for a Peaking Turbines 14,000 ------------------------------------------- regionfor each of the 8,760 hours in the year.The hourly demand values are sorted from highest to lowest.Figure 6.1 shows an example 12,000 --- ---------------------------------------- from a typical east coast electric utility. Cycling Coal,Oil,Gas The figure shows that the highest hourly electric demand was16,000 85000 ------- --------------- ------------- MW and the lowest was 5,500 MW.It also shows that the peaking Gas Combined Cycle 6,000 --------- ---- turbines and reciprocating engines operated for only about 200 Base Load Coal hours per year(in this case,during very hot hours of the summer), 45000 -------------------------------------------- while the baseload coal and nuclear plants operated throughout 25000 -------------------------------------------- the year.The total area under the curve is the generation needed to Nuclear meet load plus line losses(in this case,about 80 million MWh).The — varying electric load is met with a large number of different types 0 15000 2,000 3,000 4,000 5,000 6,000 7,000 85000 and sizes of generating units. HOURS Figure 6.1 also indicates a typical mix of generating technologies. The generating units are dispatched based on a number of factors, The load duration curve in Figure 6.1 depicts an existing generation the most important usually being the unit's variable cost—the mix. However,efficiency could also prevent the need for future cost of fuel along with operation and maintenance directly related power plant construction. For most energy efficiency program activity to production and perhaps regulation-based loading orders that in the United States,it is safe to assume that only existing generator prioritize certain resource types,such as renewables. Baseload units emissions are avoided in the short term of one to five years.However, are operated as much as possible,unless there is an environmental if the analysis is estimating impacts over a longer period of time and/ regulation limitation,because they are the least expensive.On the or the scale of the programs being evaluated is large enough,then other hand, peaking and intermediate(cycling)units are used only new units could be considered as well. when needed because of their higher costs.The type of units—such as baseload or peaking—that are the most"polluting"can vary from The emission factor from a generating unit that would not be run due one region to another. to energy efficiency is called the operating margin(OM).The emis- sion factor from a generating unit that would not be built is called Compared to the base case,energy efficiency displaces a certain the build margin(BM).In general terms,avoided emissions can be amount of generation during each hour that it operates. Efficiency estimated by determining the extent to which an efficiency program essentially takes a"slice"off the top of the load curve for the hours or portfolio affects the BM and OM and either(1)determining that it occurs,displacing the last unit of generation in each of these appropriate emission factors for the BM and OM using the emission hours.The displaced emissions can be estimated by multiplying the factor approach,or(2)accounting for new and existing generating displaced generation by the specific emission rate of that unit or by units when using the scenario approach.This is discussed further in preparing scenario analyses. the following subsections. Depending on the hour of the day or year and the geographical The general formula for calculating emission rates for determining location of the avoided electricity use,the displaced unit could be avoided emissions is: a cycling coal,oil,or steam unit;a combined-cycle unit;a central ER=(w)x(BM)+(1—w)x(OM) station peaking turbine;or a reciprocating engine unit—or even a zero-emissions unit.The first challenge in calculating the avoided Where: emissions for electricity generation is defining the mix of technologies ER is the average emission rate(e.g.,tons of displaced by the efficiency programs for the specific program location CO2-equivalent/MWh) and during specific times of the year. 6-7 www.seeaction.energy.gov December 2012 • w is the ratio(between 0 and 1)assigned to the build margin If the BM is included in the analyses,it must be explicitly specified, • BM is the build margin emission factor(e.g.,tons of including the basis for its calculation and justification for its use as the CO2-equivalent/MWh) most likely scenario. In recent years,estimates for BM emission rates • OM is the operating margin emission factor(e.g.,tons of have been based on advanced-technology coal plants or gas-fired, CO2-equivalent/MWh). combined-cycle power plants,as most new thermal plants adopt this technology. However,with new technologies being developed and Time is explicit in the above equation.That is,the emission reduction renewable portfolio standards becoming more prevalent,changes in can vary from year to year(or in theory from hour to hour)as the market conditions should be tracked and accounted for when using a variables w,BM,and OM change over time. In this equation,w BM emission factor. indicates where the generation produced(or reduced)by the project 6.4.2 Using Avoided Emission Factors and Scenario activity would have come from in the baseline scenario.A ratio(w) Analyses for Electric Grid Analyses of 1 means that all generation produced or saved by the project The methods for determining avoided emissions values for displaced activity would have come from an alternative type of new capacity generation range from fairly straightforward to highly complex.They built in place of the project activity(the BM).A ratio between 0 and include both spreadsheet-based calculations and dynamic modeling 1 means that some of the generation would have come from new approaches with varying degrees of transparency,rigor,and cost. capacity(BM)and the remainder from existing capacity(the OM).A Evaluators can decide which method best meets their needs,given ratio of 0 means that all of the generation would have been provided evaluation objectives,available resources,data quality requirements, by existing power plants,and no new capacity would have been built and evaluation framework requirements.Designers of programs or in place of the project activity. regulations that use these estimates may also wish to specify one One approach to determining OM and BM can be found in the or more methods at the outset,and a process for periodic review of World Resources Institute/World Business Council for Sustainable those methods. Development(WRI/WBCSD)Protocol Guidelines for Quantifying The emission rates of the electric grid will vary over time.Thus,the GHG Reductions from Grid-Connected Electricity Projects.60 In this emissions analyses are typically conducted annually for each year WRI/WBCSD approach,there are three options for selecting the BM emission factor:61 of the evaluation reporting period for electricity-saving programs. Emissions rates can also vary hour by hour as the mix of electricity • Option 1.Use a project-specific analysis to identify the type plants operating changes to meet changing loads.The decision to of capacity displaced.Under this option,the BM emission use an annual average analysis,an hourly analysis,or some time factor is representative of a single type of power plant.This type of period of analysis in between is up to the evaluator to decide,based power plant will be either(1)the baseline candidate(i.e.,baseline on evaluation objectives and available resources,as well as evalua- power plant)with the lowest barriers or greatest tion framework requirements. net benefits,or(2)the most conservative,lowest-emitting baseline candidate. The following are descriptions of four emissions quantification • Option 2.Use a conservative"proxy plant"to estimate BM approaches.These four approaches are listed in order of increasing emissions. Under this option,the BM emission factor is determined complexity of assumptions and sophistication of analysis: by the least-emitting type of capacity that might reasonably be 1. Regional non-baseload emission rates(using ERAVs built.In some cases,this baseline candidate could have an emis- eGRID database) sion rate of zero(e.g.,renewables).Another way to determine 2. Regional marginal baseload emission rates(using capacity a proxy is to look at the plants that have recently been built and factors or equivalent) connected to the grid. 3. Regional historical hourly emission rates • Option 3.Develop a performance standard to estimate the BM 4. Energy scenario modeling. emission factor. Under this option,the BM emission factor will reflect a blended emission rate of viable new capacity options. Table 6.1 provides some guidance on selection of an approach for a given set of objectives and metrics. December 2012 www.seeaction.energy.gov 6-8 TABLE 6.1: Choosing an Avoided Grid Electricity Emissions Quantification Approach APPROACHEMISSIONS ENERGY DATA EMISSIONS ANALYTICAL QUESTIONS QUANTIFICATION NEEDS OUTPUTS What is the relative magnitude of emission Load emission rates Annual or seasonal Regional reductions of the energy efficiency portfolio? (using EPA's eGRID energy impacts(MWh) non-baseload database) avoided emissions Which EGUs in my region are on the margin, Regional marginal Annual or seasonal Regional marginal and how much seasonal or annual emissions will baseload emission energy impacts(MWh) unit avoided be avoided? rates(using capacity emissions factors or equivalent) How can I quantify hourly emission reductions? Regional historical Hourly energy impacts Regional hourly How much are emissions reduced during peak hourly emission rates (MW and/or MWh) avoided emissions electricity demand? How will EGU emissions change in future years? Energy scenario Hourly,seasonal,or Regional avoided How can I compare baseline and forecast emissions modeling annual energy impacts, hourly,seasonal, that can result from efficiency portfolios? which depend on the or annual emissions model used(MW and/ based on dispatch How can I estimate avoided emissions in a cap-and- or MWh) order trade program? 6.4.2.1 Regional Non-Baseload Emission Rates Approach Figure 6.2 shows the eGRID subregions,which are identified and (Using EPA's eGRID Database) defined by EPA,using the North American Reliability Corporation This approach entails a simple calculation—multiply the amount of (NERC)regions and power control areas as a guide.An eGRID subregion generation or electricity sales displaced by the efficiency action by is often,but not always,equivalent to an Integrated Planning Model the"non-baseload"emission rate indicated for a specific pollutant (IPM)63 subregion.The 26 eGRID subregions in eGRID2010 are in a region.The non-baseload emission rate represents an average subsets of the NERC regions,as shown in Figure 6.2. emission rate for the EGUs that are likely to be displaced by end-use The eGRID subregion non-baseload output emission rates are used efficiency actions. to estimate emission reductions of end-use efficiency programs, A standard source of emission rates for the United States is the EPA's portfolios,or policies that reduce consumption of grid-supplied eGRID database.The eGRID database includes operational data such electricity. Non-baseload output emission rates are associated with as total annual emissions and emission rates(for GHG, NCx,S02,and emissions from plants that combust fuel and are the EGUs most mercury),generation,resource mix,capacity factors,and heat input. likely to back down when energy efficiency and renewable energy, The eGRID emissions are associated with the generation of electric- policies,and programs are implemented.These emissions data are ity,and thus the values do not account for T&D losses,imports and derived from plant-level data and are aggregated up to the eGRID exports among eGRID subregions(or any other geographic area), subregion level. transmission constraints within any geographic area,or lifecycle emissions at EGUs(e.g.,emissions from the extraction,processing, and transportation of fuels). 6-9 www.seeaction.energy.gov December 2012 FIGURE 6.2: U.S. EPA eGrid subregions MROE NEWE MRO NYUP RFCM NWPP NYU RFCE NYCW RFCW RMPA CAMX SRMW SPNO SRTV SRVC SPSO AZNM SRMV SRSO AKMS ERCT H OA ♦ �� FRCC AKGD �HIMS Source:U.S.Environmental Protection Agency.The Emissions&Generation Resource Integrated Database(eGRID).www.epa.gov/egrid/ The main advantages of the eGRID subregion non-baseload emission The bottom line is that this is an easy approach to apply, but the rates approach are that it is a straightforward,simple calculation and tradeoff can be relatively high uncertainty in the estimate of the that it can be used for communicating to the public that emission emissions-related impacts. In summary,the advantages of this reductions can result from the implementation of efficiency programs. approach are as follows: A possible shortcoming of this approach is that energy efficiency . It provides an easy"back of the envelope"calculation savings tend to vary over time,such as savings from an office lighting • Non-baseload output emission rates provide a basic retrofit that only occurs during the workday or an exterior lighting understanding of how much EGU emissions could likely retrofit with savings only at night;thus,using an annual average emis be avoided or displaced. sion factor that lumps daytime,nighttime,weekday,and weekend values together may skew the actual emissions benefits calculation. And,in summary,this approach also has some limitations: A system-average emission rate may also be purely historical,and • Future-looking EGU representation is missing(i.e.,there is no thus fail to account for changing emissions regulations and new "build margin"analysis) plant additions. Historical system averages will tend to overestimate • Some EGUs in the base year may have already shut down or emissions impacts if emissions limits become more stringent over will do so in future years time.Alternatively,a system-average emission rate could be estimated • The eGRID approach uses averages and does not show where for a hypothetical future system,based on assumptions about emissions or which EGUs will be displaced from new plants and future regulatory effects on existing plants. 0 Information is generally on a 3-year time lag. December 2012 www.seeaction.energy.gov 6-10 • The eGRID approach only accounts for generation within a seasonal weekday and weekend and nighttime and daytime values specific area and does not include information about imports/ (i.e.,six emission factors)to match up the net efficiency savings for exports of electricity(except for state-level net imports) the equivalent time period will significantly improve estimates over • The approach assumes that efficiency activities will affect all the other emission factor methods previously described above. non-baseload plants proportionally to each plant's non-baseload generation. One of these marginal baseline approaches is to use EGU capacity factor data.An EGU's capacity factor is the ratio,in a given period of 6.4.2.2 Regional Marginal Baseload Emission Rates time,of the actual electricity produced by a generating unit to the Approaches (Using Capacity Factors or Equivalent electricity that could have been produced at continuous full-power Approaches) operation.The capacity factor of an EGU can be used as a proxy for These approaches have been developed to provide a reasonably how likely the EGU is to be displaced by end-use efficiency actions. accurate estimate of displaced emissions at a lower cost than The approach helps users understand the relative dispatch order of analyses using the next two approaches,historical hourly data or the EGUs within a state or group of states.After doing this analysis, modeling,but with more accuracy than the simple,non-baseload one has an estimate of which EGUs are on the margin(essentially approach discussed above.Using regional marginal baseload emission the load duration curve)and how much emissions could be displaced rates typically involves using spreadsheets and compiling publicly from each EGU on a seasonal or annual basis. available data to approximate which marginal generating units For example,this approach assumes that EGUs with low capacity will be supplying power at the time that efficiency resources are factors(e.g.,operating at equal to or less than 20%of capacity)are reducing consumption. most likely to be displaced by the efficiency portfolio,and EGUs with The two major steps in a spreadsheet-based analysis are to(1)determine high capacity factors(e.g.,operating at equal to or greater than 80% the relevant set of generating units(i.e.,account for the location of of maximum capacity)would not be displaced by the efficiency activ- the efficiency program's projects as well as transfers between the ity.When available,seasonal capacity factors should be used instead geographic region of interest and other power areas),and(2)estimate of annual capacity factors,which ignore seasonal weather variations. the displaced emissions from those units. For example,many combustion turbines only operate during summer daytime hours in a typical year. Using an annual capacity factor As discussed above,generating units are typically dispatched in a would incorrectly allocate displaced emissions to these units during predictable order,based on cost and other operational characteristics. seasons when they are not operating. This means it is possible,in principle,to predict which unit types will be"on the margin"at a given load level,and thereby predict the In summary,the advantages of these approaches are as follows: marginal emission rates. Data on regional power plants may be used • Emissions can be assigned to each EGU to develop supply curves representing different seasons and times of • The calculation is relatively easy if the analysis infrastructure is day.These curves are then used to match regional electricity loads to set up characteristic emission rates.Although this method can use readily • It is a simple way to get a relative sense of the marginal unit in available public data,it is based on a simplified view of the dispatch the area of analysis. process that does not account for transmission congestion. These approaches also have Imitations: As with the system-average approach,this method does not provide Annual capacity factors assume that the EGUs operate the a way to determine how large a geographic region should be same throughout the year considered or how interregional transfer is estimated. However,this approach improves upon the system-average approach with respect • Emissions estimates are approximate,based on annual or to identification of marginal generators. In either case,the analysis seasonal capacity factors,and they do not account for must include the effect of changing environmental regulation,as maintenance or outages discussed above. • Imported and exported power is not considered • They assume that EGU generation characteristics are the A significant advantage of using time-varying emission rates is same in the base year and future years that they can match up to the time-varying savings from efficiency programs.Even if an hour-by-hour load shape is not used,having 6-11 www.seeaction.energy.gov December 2012 • They assume that all energy savings or generation affect Other examples of data are available as well: all peaking units first,which is not always true(e.g.,street • The Mid-Atlantic Regional Air Management Association lighting programs). (MARAMA)completed an hourly emissions analysis of the As an example of this approach,since 1993, ISO New England Inc. states in the Northeast and Mid-Atlantic region.65 (ISO-NE)has annually analyzed the marginal emission rates of the • The Metropolitan Washington Council of Governments used New England electric generation system.This is motivated by the a time-matched marginal emissions approach that matches need to determine the emission reductions that demand-side man- certain efficiency(and renewable)technologies or measures agement(DSM)programs have had upon New England's aggregate with historical hourly emissions information from the EPA NOx,SOz,and COz generating unit air emissions.The use of these hourly database.This emissions tool can be used for the emission rates was subsequently broadened to include the benefits Virginia/Maryland/Washington, D.C.area.66 of renewable resource projects in the region.The 2010 report is To use this approach,one needs to identify whether the efficiency available at http://iso-ne.com/genrtion_resres/reports/emission/ affects peak hours and/or baseload energy use.The evaluator should fina I_2010_em issions_report_v2.pdf. add the programs together in a"bottom up"approach to obtain an 6.4.2.3 Regional Historical Hourly Emission Rates Approach aggregate level of energy savings and generation on an hourly basis, and then apply their impacts to the predicted displaced EGUs.The This approach requires technical manipulation of historical generation, steps associated with this approach are as follows: load,and emission rates to determine EGU dispatch order and marginal Collect EGU hourly emissions data from EPA or other resources. emissions rates.By applying this approach,one can determine where each EGU fits within the dispatch order for each hour,day, 2. Determine EGU dispatch order(and thus hourly emissions)and and month of a historical year—and thus determine which EGU's prepare hourly"bins"of emission rates. emissions are avoided by less energy demand,due to efficiency,in a 3. Apply hourly end-use energy efficiency savings,with a T&D given hour.This approach does not account for electricity imports or factor,to the"bins"of emission rates to quantify hourly exports into or out of the grid being studied,nor does it specifically avoided emissions as a result of the efficiency activities. address transmission constraints. The advantages of this approach are as follows: With this approach,one can understand (for every hour or segment . Reported data are easy to find on EPA's website on an hourly, of hours of a historical year)which EGUs are baseload(operating daily,and quarterly basis all hours of the day),which EGUs are load-following(EGUs that • Emission rates from any group of hours can be derived from the ramp up or down depending upon demand),and which EGUs are hourly data. peaking units(EGUs that only operate at high demand periods).This approach is most appropriate to answer questions such as:"How This approach also has limitations: much emissions are reduced in blocks of hours,or during periods of • Setting up an hourly emissions database can be resource peak electricity demand?"or"How much emissions are reduced for intensive if the infrastructure is not established demand-response policies?" • Representation of future EGU emissions is missing The EPA has information that can be used for this approach.It col- . Only EGUs subject to EPA's national reporting requirements lects generation,emissions,and heat input data in hourly intervals are represented in its hourly database from continuous emissions monitoring systems(CEMS)for all Energy import and export exchanges and transmission large EGUs subject to EPA trading programs.For example,the EPA constraints are not captured. implements the emissions cap-and-trade program for the Acid Rain Control Program,the NOx Budget Trading Program,and the Clean 6.4.2.4 Energy Scenario Modeling Approach Air Interstate Rule.These programs require an hourly accounting At the other end of the complexity spectrum from calculating of emissions from each affected unit.Affected units are sources simple average emission factors are the energy scenario modeling participating in CEMS that provide hourly emissions data unless the approaches that use dynamic simulation models of the grid.These EGU qualifies to use one of the alternative monitoring methodolo- are generically called dispatch models.Varieties or different names gies specified in EPA rules.64 for dispatch models include capacity expansion models,production December 2012 www.seeaction.energy.gov 6-12 cost models,and system planning models.Dispatch models fore- an anticipated future system based on detailed assumptions about cast which EGUs will operate at any given time based on inputs additions,retirements,and major grid changes(e.g.,capacity and assumptions in the model,and their algorithms simulate the expansion or system planning models). However,dispatch models complex interactions of the grid with consideration of factors such do not model the competition among different generating technolo- as transmission constraints,import/export dynamics,fuel prices,air gies to provide new generation. In general,the model produces a pollution control equipment,and a wide range of energy policies and deterministic, least-cost-system dispatch based on a highly detailed environmental regulations. Dispatch models specifically replicate representation of generating units—including some representation least-cost system dispatch,with the lowest-cost resources dispatched of transmission constraints,forced outages,and energy transfers first and the highest-cost last.All of these models can capture a high among different regions—in the geographic area of interest. level of detail on the specific ECUs displaced by energy efficiency projects or programs. Dispatch models using only data for existing units are generally good to use for projecting avoided emissions 1-5 years into the future, The models are used to generate scenarios of the electric grid's especially when the future EGU fleet is not changing substantially. operation and emissions. If the power system is altered through load Capacity expansion models forecasting future generation and retire- reduction or the introduction of an efficiency program,the model ments,as well as the dynamic fluctuation within the electric grid,are calculates how this would affect dispatch and then calculates the generally useful for analysis 5-30 years into the future.This approach resulting emissions and prices.The basis for this scenario approach is most appropriate to use when a portfolio is expected to be large is that a dispatch model is run with and without the efficiency enough to substantively change electric system operations. actions,and the resulting difference in emissions is calculated.The models can also be used to provide hourly,monthly,or annual The advantages of this approach are as follows: emission factors. • It is the most sophisticated way to capture how the electrical grid will react to implementation of efficiency actions Dispatch modeling can be the most precise means of quantifying • Electricity transfers are well represented avoided emissions(assuming good input assumptions and qualified modelers)because it can model effects of load reductions that are • It can provide very detailed estimations about specific plant substantial enough to change dispatch(as well as future changes and plant-type effects such as new generating units or new transmission corridors)on an • It can provide highly detailed,geographically specific hourly hourly basis,taking into account changes throughout the intercon- avoided emissions data at the EGU level. nected grid.As such,it is a preferred approach where feasible. • Future EGU generation and retirements can be represented On the downside,dispatch modeling typically involves the use of • It can be relatively inexpensive to use for an efficiency portfolio proprietary,commercial programs; requires extensive underlying analysis if the models are already developed,and appropriate data;and can be labor intensive and difficult for non-experts to data populated for other purposes,such as supply-side evaluate.These models can also be expensive,although the costs generation analyses. have been reduced over recent years and—particularly if the This approach also has limitations: results can be applied to a large program or several programs— the improved estimate can be well worth the incremental cost. The models are only as good as the assumptions used Accordingly,they are probably most appropriate for portfolios that • Hourly emissions data are not always available as the seek to achieve significant quantities of electrical energy efficiency or required input long-term effects.For large statewide programs,the modeling costs • Energy models are proprietary,require significant resources to may be relatively small compared to the program and evaluation run compared to other approaches,and can be data intensive costs;the California Public Utilities Commission,for example,used . Input assumptions can be difficult to discern due to the dispatch modeling to determine the avoided greenhouse gases from proprietary nature of the models and the amount and various efficiency portfolios.67 complexity of assumptions and data used An hourly dispatch model simulates hourly power dispatch to explic- • Expertise in energy modeling is normally recommended. itly estimate emissions from each unit in a system.That system can represent the current grid and generating units,or it can represent 6-13 www.seeaction.energy.gov December 2012 6.5 SELECTING AN APPROACH FOR For criteria pollutants,the following document from the EPA provides CALCULATING AVOIDED EMISSIONS guidance on calculating avoided emissions for both energy efficiency The choice of evaluation approach is tied to the objectives of the and renewables programs:The Roadmap for Incorporating Energy program being evaluated,the scale of the program,the evaluation Efficiency/Renewable Energy Policies and Programs into State and budget and resources,and the specific emissions the program is Tribal Implementation Plans,available at www.epa.gov/airquality/ avoiding.For direct onsite fuel savings and the resulting avoided emis- eere(see Appendix I of the roadmap).Additional EPA references are sions,per common practice,standard emission factors can be used. provided in Appendix C.2.3. For electricity savings programs,system average emission values can be used,but they should be avoided except in the simplest estimates. There are also medium-effort approaches(using capacity factors or historical emissions rates)that can fairly accurately quantify the effects of electricity energy efficiency programs.However,the most sophisticated approaches involve dispatch modeling and the result- ing detailed calculation of hourly emissions.While the costs and complexity of these models has limited their use in the past,this is beginning to change. Dispatch models are potentially cost-effective evaluation tools that should be considered for evaluations of large- scale programs. 6.6 ADDITIONAL RESOURCES ON AVOIDED EMISSIONS CALCULATIONS The following documents provide some guidance with respect to greenhouse gas programs.Each is a product of the World Business Council for Sustainable Development(WBCSD)and/or the World Resources Institute(WRI)and is available at www.wri.org/climate. • Guidelines for Quantifying GHG Reductions from Grid- Connected Electricity Projects,www.ghgprotocol.org/files/ghgp/ electricity_final.pdf, published in August 2007 • GHG Protocol Corporate Accounting and Reporting Standard (Corporate Standard),www.ghgprotocol.org/files/ghgp/public/ ghg-protocol-revised.pdf,revised edition,published in March 2004 • GHG Protocol for Project Accounting(Project Protocol), www.ghgprotocol.org/files/ghgp/ghg_project_protocol.pdf, published in December 2005. Examples of energy efficiency projects implemented for their greenhouse gas emissions benefits can be found at the Climate Trust website:www.climatetrust.org. December 2012 www.seeaction.energy.gov 6-14 CALCULATIONOF • • •NSIN'S FOCUS ON • Evaluators for Wisconsin's Focus on Energy(Focus)public benefits To identify marginal plants,the evaluators calculated the average energy efficiency program estimated emission factors for the plants length of time,in hours,that a generating unit remains on once it serving Wisconsin.They used these data to estimate environmental has been brought online. Peaking units,which are brought on for impacts,in the form of displaced power plant emissions,associ- only a short time,have a short average time on;baseload plants ated with Focus energy savings.The evaluation team developed that remain on for hundreds of hours or more have a long average a model to estimate the generation emission rates for NOX,SOX, time on.The evaluators divided the population of generating CO2,and mercury using hourly measured emissions data from the units into five groups:those averaging less than 6 hours on, EPA for the power plants supplying Wisconsin(EPA's`Acid Rain 6-12 hours on,12-24 hours on,24-96 hours on,and more than Hourly Emissions"data series).The evaluation team aligned its 96 hours on for each time they are dispatched. Marginal emis- method for estimating emission rates with recommendations of sions in each hour were defined as those produced by the set of the Greenhouse Gas Protocol initiative(GHG Protocols)developed generating units in the group with the shortest average time on. by the World Resources Institute(WRI)and the World Business Because the EPA data allow an 8,760-hour accounting of pollutants, Council for Sustainable Development(WBCSD). Emission factors insofar as energy savings can be assigned to hours of the day and from reduced use of natural gas at the customer sites were also days of the year,a more accurate emission rate can be estimated taken from EPA data. by matching the amount of energy savings in a given hour to the emission rate for that hour. Focus evaluators call this approach Using the emission rates and evaluation-verified gross electricity time of savings(TOS)emission factors. savings estimates,the Focus programs together potentially avoided 8.7 million pounds of NOX; 10.7 million pounds of SOX;6.6 billion It should be noted that Wisconsin's power plants are included in pounds of CO2;and more than 41.5 pounds of mercury from the federal S02 cap-and-trade program(acid rain provisions). In inception to December 31,2010(see Table 2-23 of the Focus on this cap-and-trade system,S02 emissions may not be considered Energy Evaluation Annual Report[20101, Revised June 17,2011). reduced or avoided unless EPA lowers the S02 cap.One can say that the program avoided generation that previously emitted One implication of adherence to the GHG Protocol is that emission S02, but one cannot claim that future S02 emissions will actually factor calculations are based on generation data specific to the be reduced due to the effect of the trading program.Starting in geography of efficiency programs.The relevant set of plants from 2009,the plants were also subject to a cap-and-trade program which emissions are displaced are those that serve the electric for NOX(the Clean Air Interstate Rule),which has the same effect. grid in the areas where the efficiency programs are implemented. A second implication of following the GHG Protocol is that emis- Provided by David Sumi of The Cadmus Group sion factors are estimated only for plants that are operating on the margin,that is,the plants most likely to have remained off- line as a result of a reduction in demand/consumption resulting from energy efficiency programs. 6-15 www.seeaction.energy.gov December 2012 Chapter 6: Notes 40 U.S. Environmental Protection Agency.(February 2010).Assessing 48 Leakage: In its broadest terms,leakage is the concept that an the Multiple Benefits of Clean Energy:A Resource for States. activity or outcome expected to occur and remain within a defined www.epa.gov/statelocalclimate/resources/benefits.html. boundary flows outside the boundary,leading to unintended results. In energy efficiency programs,an example of leakage is when a 41 Intergovernmental Panel on Climate Change(IPCC).(2007). measure is incented by a program(with the associated costs and Climate Change 2007:Synthesis Report.www.ipcc.ch/pdf/ assumed savings)but is installed outside of the program's jurisdic- assessment-report/ar4/syr/ar4_syr.pdf.Contribution of Working tion. In the context of air regulation,such as a cap-and-trade pro- Groups I,II,and III to the Fourth Assessment Report of the IPCC. gram,an example of leakage is a shift of electricity generation from Share of different sectors in total anthropogenic GHG emissions in sources subject to the cap-and-trade program to higher-emitting 2004 in terms of CO2-eq. sources not subject to the program. 42"Global Warming Solutions Act:Assembly Bill 32."(2006). 49 Interactive Effects:The influence of one technology's application pp California Air Resources Board(CARE).www.arb.ca.gov/cc/ab32/ on the energy required to operate another application.An example ab32.htm. is the reduced heat in a facility as a result of replacing incandescent 43 U.S. Environmental Protection Agency.(August 2011).Federal lights with CFLs,and the resulting need to increase space heating Implementation Plans:Interstate Transport of Fine Particulate Matter from another source(usually fossil fuel-fired)or to decrease cooling and Ozone and Correction of SIP Approvals.www.gao.gov/products/ from another source(usually powered with electricity). C00127.40 CFR Parts 51,52,72,78,and 97[EPA—HQ—OAR-2009— 50 Acid Rain Program."(2012). U.S. Environmental Protection 0491;FRL-9436-81 RIN 2060—AP50.ACTION:Final rule. Agency.www.epa.gov/airmarkets/progsregs/arp. 44 The Clean Air Act requires state and local air pollution control 51"Clean Air Interstate Rule."(2005). U.S. Environmental Protection agencies to adopt federally approved control strategies to minimize Agency.www.epa.gov/cair. air pollution.The resulting body of regulations is known as a State Implementation Plan(SIP). 52"Regional Greenhouse Gas Initiative."(2012). http://rggi.org/. 45 U.S. Environmental Protection Agency.(2004).Guidance on State 53"California Greenhouse Gas(GHG)Cap-and-Trade Program." Implementation Plan(SIP)Credits for Emission Reductions from (2012).California Air Resources Board(CARB).www.arb.ca.gov/cc/ Electric-Sector Energy Efficiency and Renewable Energy Measures. capandtrade/capandtrade.htm. www.epa.gov/ttn/oarpg/t1/memoranda/ereseerem_gd.pdf. For example,the Metropolitan Washington Council of Governments 54 There are some cap-and-trade program design features that may developed a regional air quality plan for the eight-hour ozone stan- lead to exceptions to this general rule,including banking and borrow- dard for the D.C.Region nonattainment area that included an energy ing of allowances,a safety valve(where additional allowances above efficiency measure.The plan was adopted by Virginia, Maryland, the cap level are made available by the government at a known and the District of Columbia,and the respective ozone SIPS were price),and establishing the cap at levels above business-as-usual approved by the U.S. EPA regions in 2007. emissions(so the emission constraint is not binding on emitters). However,as a first approximation,covered facilities will emit 46 While not directly comparable,another term that is used in the approximately at the cap level. context of the Clean Air Act is"surplus."The definition of surplus depends on how the emission reduction will be used,but basically it 55 U.S.Environmental Protection Agency.(September 2005).(Draft). implies that one cannot double count a program's emission reduc- State Set-Aside Programs for Energy Efficiency and Renewable tions.As a simple example,if a program's impacts are assumed in the Energy Projects Under the NOx Budget Trading Program:A Review baseline then one cannot claim credit for the impacts in a SIP. of Programs in Indiana,Maryland,Massachusetts,Missouri, New Jersey,New York,and Ohio. http://epa.gov/statelocalclimate/ 47 World Resources Institute(WRI);World Business Council for documents/pdf/eere_rpt.pdf. Sustainable Development(WBCSD).(December 2005).GHG Protocol for Project Accounting(Project Protocol).www.ghgprotocol.org/files/ 56 An example of such a policy is the inclusion of an energy efficiency ghgp/ghg_project_protocol.pdf.Source of the"assessment bound- resource standard in a State Implementation Plan(SIP). ary"and"primary/secondary"terminology. December 2012 www.seeaction.energy.gov 6-16 57"Compilation of Air Pollutant Emission Factors."(2011). U.S. Environmental Protection Agency.EPA AP-42.www.epa.gov/ttn/ chief/a p42. 58 Intergovernmental Panel on Climate Change(IPCC).(2006). 2006 Guidelines for National Greenhouse Gas Inventories. www.ipcc-nggip.iges.orjp/public/2006gl/index.html. 59 The exception would be if all of the electricity serving the site of the energy efficiency activity is generated in a non-grid-connected, stand-alone generation system. 60 World Resources Institute(WRI);World Business Council for Sustainable Development(WBCSD).(August 2007).Protocol Guidelines for Quantifying GHG Reductions from Grid-Connected Electricity Projects.www.wri.org/publication/guidelines-quantifying- ghg-red uctions-grid-connected-electricity-projects. 61 The WRI/WBSCD Protocol calls for"conservative"estimates,which tend to result in underestimating of benefits,versus using a"most- likely"scenario,which can lead to more accurate estimates. 62 "eGRID."(2012). U.S. Environmental Protection Agency. www.epa.gov/cleanenergy/energy-resources/egrid/index.html. eGRID provides a comprehensive source of data on the environ- mental characteristics of almost all electric power generated in the United States. 63 The IPM is a multi-regional,dynamic,deterministic linear program- ming model of the U.S.electric power sector. It provides forecasts of least-cost capacity expansion,electricity dispatch,and emissions control strategies while meeting energy demand and environmental, transmission,dispatch,and reliability constraints. 64'Air Markets Program Data."(2012). U.S. Environmental Protection Agency. http://ampd.epa.gov/ampd/.See hourly emissions data from continuous emissions monitoring systems. 65 Mid-Atlantic Regional Air Management Association(MARAMA). www.marama.org. 66"Promoting Air Quality and Climate Protection with Clean Energy and Energy Efficiency in the Metropolitan Washington Area." Metropolitan Washington Council of Governments(MWCOG). www.mwcog.org/environment/air/EERE/default.asp. 67 California Public Utility Commission(CPUC)selected a team led by Energy and Environmental Economics, Inc.to model the electric- ity sector's compliance with AB32,California's Global Warming Solutions Act.www.ethree.com/public_projects/cpuc2.php 6-17 www.seeaction.energy.gov December 2012 Chapter 7 Impact Evaluation Considerations Chapter 7 has two major portions.The first several sections cover issues that often arise in impact evaluations: determining baselines, determining demand savings, calculating persistence of savings, addressing uncertainty of savings estimates, setting evaluation budgets, and establishing evaluation principles and ethics. Because most of this document focuses on using evaluation to determine energy (and demand) savings as well as avoided emissions from conventional efficiency programs,the second portion of this chapter provides brief overviews of the other evaluation objectives: program feedback, resource planning, and calculating non-energy benefits as well as evaluating some relatively unique program types.These other program types focus on residential behaviors, education and training, market transformation, codes and standards, demand response, or greenhouse gas (GHG) mitigation strategies. 7.1 SELECTING BASELINES approaches,the existing conditions are determined generically at A major impact evaluation decision is how to define the baseline. the time of program design(and as perhaps defined in a technical Baselines are the conditions,including energy consumption and reference manual [TRM]). demand,which would have occurred without implementation of the For M&V approaches,the existing condition is typically determined subject energy efficiency activity.Baseline conditions are sometimes at the time of measure/project installation.The common means for referred to as business-as-usual conditions and are used to calculate determining such existing conditions baselines,assuming sufficient project-and program-related savings. Baselines can also include information is available,are through(1)an inventory of pre-retrofit definitions of non-energy metrics that are being evaluated,such as equipment site-specific characteristics,including nameplate ratings, air emissions and jobs. and/or(2)an assessment of the existing equipment's or system's Baselines represent the counterfactual condition that is estimated to energy consumption rates,based on measurements or historical complete an evaluation.With large-scale consumption data analysis data.Site-specific characteristics include how and when the affected approaches,the baseline is defined by the characteristics and energy equipment or systems are operated.For example,for an energy use of the control group(s)used in the analyses. However,for the efficient lighting retrofit,the baseline decisions include the type of noncontrol group-based impact evaluation approaches,such as lighting equipment that was replaced,the power consumption(watts deemed savings and measurement and verification(M&V),baseline per fixture)of the replaced equipment,and how many hours the definitions are determined by the type of project being implemented, lights would have operated. site-specific issues,and broader,policy-oriented considerations.These considerations usually result in one of three different types of base DEFINING BASELINE DATA COLLECTION - lines:existing conditions,common practice,or codes and standards. REQUIREMENTS The following is a discussion about selecting appropriate baselines for the deemed savings and M&V impact evaluation approaches. Assessing baseline and baseline adjustment issues in the This discussion first defines the three different types of baselines and planning stage is important for determine data collection and then illustrates how baselines are typically set for common efficiency budgeting requirements.The goal is to avoid reaching the project classifications. analysis stage of an evaluation and discovering that critical pieces of baseline information have either not been collected 7.1.1 Existing Conditions Baselines or have been collected with an unreliable level of quality. As the title implies,existing conditions baselines are what is in place This situation can be guarded against by providing specific (e.g.,equipment,controls,procedures)at the project site before instructions to program administrators prior to program the energy efficiency measures are implemented. For example, for a motor replacement program,it is the energy consumption of implementation. Planning for data collection is necessary to give administrators notice and justification for collecting data the motor that was replaced.As another example,for residential they would not ordinarily collect for managing and tracking behavior-based programs,it is the energy consumption associ- ated with pre-program behaviors.Typically,for deemed savings program progress. December 2012 www.seeaction.energy.gov 7-1 7.1.2 Codes and Standards Baseline BASELINES: THE CONSUMER Energy codes and standards set minimum requirements for energy PERSPECTIVE VERSUS THE PUBLIC efficient design;as such,they affect energy use and emissions for POLICY the life of a piece of equipment or building.Codes typically refer to energy requirements associated with construction of new build- As discussed in this section on baselines,it is typical for ings and major renovations of existing buildings(e.g., maximum replacement of failed equipment and projects for the baseline energy use per square feet or minimum insulation requirements). to be defined as common practice or a codes or standards Standards can refer to buildings or efficiency requirements for requirement.Even projects that are early replacement specific pieces of equipment such air conditioners and motors(e.g., projects may in subsequent years be evaluated as having a minimum motor efficiency). a common practice or codes and standards baseline.This Thus,codes and standards(C&5)baselines are the energy consump- makes sense from a public policy perspective,to not incent tion associated with buildings or specific pieces of equipment that consumers to buy what they would have normally purchased meet the legal requirements in place,in the location where a project or what they would be required to purchase,to avoid to over- is implemented.For example,for a motor replacement program, claiming lifetime savings. However,from a consumer perspec- the C&S baseline standard might consist of a motor that meets the tive,they are looking at savings from a baseline of what they minimum requirements of the federal Energy Independence and had before the project was implemented,not what"some Security Act of 2007(EISA)for new motors.Three nuances associated evaluator says they would have,or should have,done! They with C&S baselines are as follows: in effect want to see the savings as compared to past energy • Not all efficiency actions and not all measures are subject to bills,not hypothetical bills.This difference in perspectives codes or standards;thus,the application of this baseline defini- is important to acknowledge when different baselines are tion is limited. selected for determining savings from the"policy"perspective • A variation on a C&S baseline is when a code or standard exists versus the"consumer"perspective,as the savings for each but is not well complied with;in these cases,one of the other perspective may need to be evaluated differently. baseline definitions might be more applicable(or a combina- tion of definitions). In a situation where an applicable C&S exists,the common • Codes and standards tend to change,and thus C&S baselines practice baseline and the C&S baseline might be the same need to be regularly updated. value—or not,if the C&S are not complied with or are 7.1.3 Common Practice Baseline too"easy." Common practice baselines are estimates of what a typical con- • As with C&S,common practices tend to change over time; sumer would have done at the time of the project implementa- thus,these baselines also need to be regularly updated. tion.Essentially,what is"commonly done"becomes the basis for 7.1.4 Defining Baselines for Specific Program Types baseline energy consumption. For example,if the program involves For the purposes of defining baselines,most programs can be incenting consumers to buy high-efficiency refrigerators that use categorized as follows: 20%less energy than the minimum requirements for ENERGY STAR® refrigerators,the common practice baseline would be refrigerators • Early replacement or retrofit of functional equipment still that consumers typically buy.This might be non-ENERGY STAR refrig- within its current useful life erators,or ENERGY STAR refrigerators,or,on average,something in • Improvement of existing processes(e.g., reducing the energy between.Common practice is determined by surveys of participants, consumption per-unit of production in a factory) non-participants,or analysis of market data.The following are three • Replacement of functional equipment that is beyond its rated nuances associated with common practice baselines: useful life • If a common practice baseline is selected,the resulting • Unplanned replacement of failed equipment savings determination probably excludes,on average,savings • New construction and substantial existing building improve- from free riders. ments(tenant improvements) • Non-equipment based programs(e.g., behavior-based and training programs). 7-2 www.seeaction.energy.gov December 2012 TABLE 7.1: Standard Practices for Selection of Baselines for Common Program Categories PROGRAM • FOR PURPOSES • • CONDITIONS STANDARDS OF DETERMINATION Early replacement or retrofit X X X of functional equipment still Existing conditions baseline C&S baseline for the time Common practice baseline within its current useful life for the remaining life of period after the remain- for the time period after Process improvements the replaced equipment or ing life of the replaced the remaining life of the process equipment equipment Replacement of functional equipment beyond its rated X X useful life Unplanned replacement for X X (of)failed equipment New construction and substantial existing building X X improvements Non-equipment based X programs(e.g.,behavior- What people in a control based and training programs) group would be doing in the absence of the program Table 7.1 summarizes standard industry practice for defining base- • The difference between energy use for the old equipment/process lines for each of these categories of programs. Note that these are replaced(existing conditions baseline)and the new efficiency not mandates;each jurisdiction and each program should establish measure prior to the time the old equipment/process would its own baseline scenarios. typically have failed or ceased to be used("initial savings rate"). • The difference between energy use associated with a C&S or As shown in Table 7.1,the early replacement and process improve- common practice baseline and the new efficiency measure ment programs present their own challenges in determining the after the time the old equipment or process would have had to appropriate baselines.This is because the issue of likely remaining have been replaced("long-term savings rate"). life of the replaced equipment(or systems or process)becomes critically important.In contrast to the other program categories Thus,ideally,early replacement/process improvement programs listed,early replacement/process improvement programs are geared would be able to take credit for initial savings during the remain- toward replacing existing(lower-efficiency)equipment with energy ing useful lifetime of the replaced equipment/process.This is the efficient equipment before the old equipment ceases to function approach that evaluators typically refer to as dynamic baselines or the or before it would otherwise be replaced.In these early replace- "stair step"approach because of the two levels of savings estimated ment/process improvement programs,an approach for defining the over the two different periods of time.A key nuance of this dynamic baseline can actually result in savings being determined based on baseline is that it assumes data on age of the existing equipment can two different baselines,for two different periods of time: December 2012 www.seeaction.energy.gov 7-3 be gathered and that the time that a process or piece of equipment reduction is to be reported as part of an evaluation,the term could be expected to fail and/or simply be ready for replacement must be clearly defined with respect to which time period is can be reliably estimated.Even so,it is a generally recommended associated with the reduction. approach for early replacement/process improvement programs. Coincident peak demand savings:the demand reductions that 7.2 DETERMINING DEMAND SAVINGS occur when the servicing utility is at its peak demand from all (or segments)of its customers.This indicates how much of a For efficiency programs,determining energy savings is almost utility's peak demand is reduced by the efficiency program. always a goal of impact evaluations.Energy use and savings are Calculating coincident peak demand requires knowing when expressed in terms of consumption over a set time period and are the utility has its peak(which is not absolutely known until the fairly straightforward to define(e.g.,therms of natural gas consumed peak season is over).A term used to describe the relationship per month or megawatt-hours of electricity consumed over a year). of facility electrical loads to coincident peak demand is diversity Energy savings results may also be reported by time-of-use period, factor—the ratio of the sum of the demands of an energy which breaks the year into several periods coinciding with a utility user,or a group of energy users,to their coincident maximum rate schedule. Examples include peak and off-peak periods of the demand;it is always equal to or greater than 1.0. summer and winter seasons. DR peak demand savings:the demand reduction associated In addition,a program's electrical demand savings are also often of • with a DR program. DR programs reduce a utility customer's interest and,for some programs,are a primary goal,such as with electricity demand in response to dispatch instructions or price demand response(DR)programs.Electrical demand savings are signals sent to the program participants—a"call"for reduc- expressed in terms of kilowatts(kW)or megawatts(MW),which tions.Thus,the DR peak demand savings are determined for indicate rates of consumption—during a specific period of time. when there is a"call"for program participants to reduce their Historically,demand savings(specifically,peak demand savings, energy consumption rate. rather than simple annual average demand savings)have been Forward capacity market demand savings:(1)the demand harder to define and determine than energy savings.This is because reduction proposed ("bid")to an electricity system operator to determining demand savings requires data collecting and analysis meet the level of resource commitments the electricity system for specific time periods(e.g.,data might be required for summer operator estimates will be needed to meet future peak demand weekdays between noon and 6 p.m.)compared with just need- on the system,and(2)the actual demand reduction that occurs ing aggregated monthly energy data.However,with technology once a commitment is made.68(See the sidebar on page 70, advances lowering the cost of meters,the spread of"smart meters," the Independent System Operator of New England [ISO-NE] sophisticated wireless sensors,and related software,it is becom- program.) ing easier to cost-effectively collect the data needed to calculate demand savings. The calculation for demand savings isstraightforward—whether for a year,a day,or a specific 15-minute period of time: The first step in determining demand savings is defining the specific metric of interest(i.e.,how demand savings are defined).The follow— ing are common demand savings definitions that indicate the specific Each of the impact evaluation approaches described in Chapter 4,to time period of interest for determining demand savings. varying degrees of accuracy and with varying degrees of effort,can • Annual average demand savings:total annual energy savings be used to determine demand savings using the above equation.The divided by the hours in the year(8,760). In the Northwest "trick,"as mentioned above,is to collect the energy savings data for United States,this is termed average MW,or MWa.Similarly, the intervals of interest(i.e.,the time period in the above equation). average monthly demand savings or average daily demand If annual average demand savings is the only metric of interest, savings may be determined. then only annual energy savings data are necessary. However,if • Peak demand savings:several definitions are used;all involve peak demand reduction,coincident demand reduction,or demand determining the maximum amount of demand reduction response peak demand reduction values are desired,then hourly during a "peak" period of time,whether that is annual, or 15-minute(a typical period of demand recording)energy sav- seasonal,or a specific period such as during summer weekday ings data,or estimates,are required for at least those specific time afternoons or winter peak billing period hours. If peak demand periods of interest. 7-4 www.seeaction.energy.gov December 2012 ISO-NE M&V MANUAL FOR WHOLESALE FORWARD CAPACITY In 2007,the Independent System Operator of New England Baseline conditions.The manual specifies baseline condi- (ISO-NE)developed an M&V manual that describes the mini- tion requirements for failed equipment(codes/standards mum requirements the sponsor of a demand resource project or standard practice,whichever is more stringent),early must satisfy to qualify as a capacity resource in New England's retirement(codes/standards or measured baseline),and new wholesale electricity forward capacity market(FCM). Demand construction(codes/standards or standard practice).Where resources eligible to participate in FCM include demand response, standard practice is used,baseline conditions must be docu- emergency generation,distributed generation,load management, mented and meet the confidence and precision requirements. and energy efficiency.They are eligible to receive a capacity For distributed generation and emergency generation,the payment($/kW per month)based on the measured and verified baseline is zero.The baseline for real-time demand response is electrical reductions during ISO-specified performance hours. calculated using a modified rolling average of the host facility The manual was developed with input from key stakeholders in load on non-event weekdays during the same hours as the the region,including members of the New England Power Pool, called event. ISO-NE,the New England state regulatory staff,electric utility Measurement equipment specifications.The project sponsor program administrators,Northeast Energy Efficiency Partnerships, must describe measurement,monitoring,and data recording and energy service,consulting,and technology providers.The device type that will be used(and how it will be installed)for manual specifies the minimum requirements a project sponsor's each parameter and variable.Any measurement or monitoring M&V plan must address,including the following: equipment that directly measures electrical demand(or proxy • M&V methods.The sponsor must choose from options variables such as voltage,current,temperature,flow rates, largely based on the IPMVP options A through D(or and operating hours)must be a true root-mean square(RMS) equivalent). It should be noted that ISO-NE deviates from measurement device with an accuracy of at least±2%. IPMVP guidance in particular for Option A,in an attempt to Monitoring parameters and variables.The project sponsor incorporate use of deemed savings values as an acceptable must describe variables that will be measured,monitored, approach for efficiency programs,given that the M&V counted,recorded,collected,and maintained,and meet manual does not explicitly allow for use of deemed savings minimum requirements for data to be collected by end-use as a"methodology,"as provided in this guidance document). and monitoring frequency. The ISO-NE manual also allows for other M&V techniques to be used in combination with one or more of these,including The PJM Interconnection subsequently developed and adopted engineering estimates supplemented with data collected on an energy efficiency M&V manual in 2009,building largely on the the equipment affected by the measures,and/or verifiable ISO-NE M&V Manual.These two documents have largely served measure hourly load shapes(which must be based on actual to inform the development of wholesale M&V standards and retail metering data,load research,or simulation modeling).All M&V model business practices for energy efficiency by the North demand resources,including distributed generation and American Energy Standards Board(NAESB). emergency generation,must be metered at the generator. Provided by Julie E.Michals,Director,Regional EM&V Forum at Northeast Energy • Confidence and precision.The project sponsor must Efficiency Partnerships,Inc.For more information,see the ISO-NE and PJM EE M&V describe a method for controlling bias(e.g.,calibration of manuals,respectively,at www.iso-ne.com/rules proceds/isone mnls/index.html measurement tools,measurement error,engineering model) and www.pjm.com/committees-and-groups/closed-groupsleetf..ospx. and achieving a precision of±10%,with an 80%confidence level(two-tailed test)around the total demand reduction value.This requirement also applies to precision level for statistical sampling. December 2012 www.seeaction.energy.gov 7-5 Ideally,evaluation results would indicate 8,760 hours(or about rates,such as those for a lighting controls/daylighting program or 35,000 15-minute periods)per year of energy savings data that could a program that involves installation of variable-speed motor drives be easily translated into demand savings. However,in practice,both on building ventilation fans.In the latter case,there might be zero primary and secondary methods for determining demand savings (or even negative)savings when the fan is at full speed(which often are used.Primary methods involve the actual collecting of hourly, occurs during peak period hours). 15-minute,or even"continuous"demand data during the periods of interest—for example,during the peak hours of the summer months 7.3 PERSISTENCE OF SAVINGS (peak season)of each year. One important evaluation issue is how long energy savings are expected to last(persist)once an energy efficiency activity has taken Sources of hourly or 15-minute data include facility interval-metered place.While energy and demand savings are often reported with data,time-of-use consumption billing data,monthly billing demand respect to just first-year savings,the longer a measure,project,or data,and field-measured data.When interval or time-of-use program provides savings,the more valuable it is. In addition,a proj- consumption data are available,they can be used in regression ect with a 5-year payback will not be of much value if the efficiency analyses to account for the effects of weather,day type,occupancy, measures only last three years.Thus,estimates of total or lifetime and other pertinent change variables to determine the demand savings are important for determining the cost effectiveness of an savings caused by the subject program.Of course,hourly demand efficiency action,as well as their value as a long-term resource. data can require hourly independent variable data(e.g.,weather)for proper regression analysis. 7.3.1 Definitions Secondary methods apply load shapes to collected energy A persistence study measures changes in program impacts over time. consumption data that are only available as averaged or total values Related to these studies are measure-retention studies that assess on a daily,monthly,or annual basis. Load shapes indicate energy (1)the length of time the measure(s)installed during the program consumption per hour.These load shapes,for whole facilities or by year is maintained in operating condition,and(2)the extent to end use(e.g.,lighting,cooling,or heating),may be available from which there has been a significant reduction in the effectiveness of studies of related programs in similar markets.One source for the the measure(s).Two related terms and their definitions are helpful data is the energy savings load shapes,by measure,included in the to know: California Database for Energy Efficiency Resources(DEER).69 Measure persistence:the duration of an energy consum- ing measure,taking into account business turnover,early For example, load shapes might exist for typical weekdays and retirement of installed equipment,technical degradation weekend days in the form of hourly percent of monthly energy use factors(s),and other reasons measures might be removed or for an office building. In this example,if the lighting system retrofit in discontinued. a building saved 3,000 kWh in a month,and the load shapes indicate Effective useful life(EUL):an estimate of the duration of that the lighting system energy use in a building during the hour of savings from a measure. It is estimated through various means, interest(say 4 p.m.-5 p.m.on weekdays)is 1%of the monthly total, including the median number of years that the efficiency then it is assumed that 1%of the energy use is saved during that measures installed under a program are still in place and peak hour-30 kWh,or a peak demand reduction of 30 kW.Another operable. It is also sometimes defined as the date at which 50% example of secondary data is if peak-to-average demand ratios for of installed units are still in place and operational. EULs are lighting systems in the building types of interest are available.So,as also defined as"measure life,"which is essentially the length of another example,for this building that saved 3,000 kWh in a month, time that a measure is expected to be functional. or an average of about 4.2 kWh per hour,if the lighting peak-to- average demand ratio is 5:1,then the estimated peak demand Paraphrasing from a recent conference paper,70 more than 100 savings is 21 kW. energy efficiency persistence-type studies have been conducted, examining in-situ median lifetimes for residential and non residential These load shape approaches can be relatively effective for measures measures.A review of results from measure-based EUL studies with constant performance,such as a lighting retrofit that reduces around North America showed that measure lifetimes for the type energy consumption as a percentage of baseline energy use(i.e., of measures normally offered in prescriptive programs(programs change-out of 200 W fixtures for 100 W fixtures). However,these with defined measures and incentives for implementing the mea- approaches do not work as well with projects with variable savings sures)and installed in typical end uses are fairly consistent for many 7-6 www.seeaction.energy.gov December 2012 measure-based programs in commercial,residential,and industrial the pre-project level of service is very low,for example,in low- sectors.71 A key factor affecting the quantity of savings being deliv- income housing where prior to the efficiency actions,heating and ered from program-related installations of equipment is whether the lighting systems were barely functional,but the new systems allowed measures perform at the new efficiencies consistently over time,or the occupants to adequately light their homes and heat them to whether their efficiency performance degrades over time. Decays in comfortable levels. net technical performance can be an important issue,particularly for measures for which savings accrue over long periods of time. The issue for impact evaluation is whether rebound is explicitly or implicitly included in the savings determination.An example of an As discussed throughout this guide,energy savings achieved over explicit consideration is the use of a deemed rebound effect factor, time are a"difference"from a baseline performance.Thus,persistence a form of a non technical degradation factor.As with all deemed looks at degradation patterns that would be realized in standard factors,it should be specific to the applications associated with the efficiency equipment,or typical(non-efficiency)consumer behaviors, subject program and based on actual historical data.Another explicit and compares them with the degradation patterns of the program's approach would be a long-term study of rebound in the participants efficient equipment or behaviors.Therefore,savings are the differ- and a control group of non-participants.Current standard energy ence over time between the energy use of the efficient equipment or efficiency evaluation practices do not use either of these approaches behavior and the standard equipment or behavior it replaced—with or any other explicit approaches for assessing rebound. consideration of both baseline and project equipment/behavior degradation in performance(which may be the same). What is used commonly,though,is an implicit approach,and although it is not often acknowledged as including rebound effects, 7.3.2 Rebound it does.The implicit approach is associated with a savings determina- Although it is not usually included in persistence determinations, tion method that includes savings determination based on both the rebound effect(also called take back,snap back,backfire,and the project period(after the efficiency project is implemented)and a few other terms)resurfaces as an issue periodically,as does its pre-project service levels.This is true with the large-scale consump- potentially negative impact on energy savings proponents.A form of tion data analysis approaches discussed in Chapter 4.If the program this concept was first put forth by British economist William Jevons participants increase their service level and energy use compared in 1865.An interpretation of the"Jevons paradox"is that increases with a control group of non-participants,the calculations will capture in efficiency of coal processes would cause coal consumption to this in the savings determination. increase,to a level that would exceed previous consumption levels. Jevons paradox was based on the industrial sector,where coal costs There is no"correct"answer to whether or not rebound should be where a high percentage of operation,and he did not consider the evaluated,particularly given that the magnitude of the rebound counterfactual,or what would happen if the efficiency of processes effect is not known;however,the evaluation planning efforts and the did not increase.In any case,Jevons,fundamental argument has reported results should indicate whether rebound is addressed or not. been repackaged and repurposed,correctly or incorrectly,to cover micro-level and macro-level impacts;it has been extended to 7.3.3 Determining Persistence questions about energy efficiency(and carbon emissions reduction) Persistence studies can be expensive,and are thus not often efforts in general,and been given the general term"rebound." In undertaken for individual programs. Past experience indicates that this guide,the rebound effect is defined as a change in energy-using long periods of time are needed for persistence studies,and there behavior that yields an increased level of service accompanied by are significant challenges to conducting these studies,such as long an increase in energy use that occurs as a result of taking an energy lifetimes of measures(making it impractical to wait for measure efficiency action. failures or consistent patterns of degradation),incomplete data sets, high cost of data collection,and of course,the need for trained staff. Few,if any,in the efficiency industry argue that the rebound effect never occurs,although few rigorous studies of rebound have been The most accurate manner to report savings persistence is probably completed.72 Clearly,in some situations,an energy efficiency action as an annual percentage(e.g., 100%in year 1,98%in year 2,95% results in an increased level of service and an associated increase in in year 3,and so on).However,the most common way that persistence absolute energy use greater than would be experienced if the level is indicated is via a value for EUL(e.g.,20 years for an energy-efficient of service did not increase.This is most probable in situations where refrigerator). December 2012 www.seeaction.energy.gov 7-7 The basic approaches for assessing persistence are as follows: the difference between(1)actual energy consumption and(2)what • Use of historical and documented persistence data,such energy consumption would have been if the efficiency measures had as manufacturer's studies or studies done by industry not been implemented is an estimate of energy(and demand)sav- organizations such as the American Society of Heating, ings.While program evaluations seek to reliably determine energy Refrigerating,and Air-Conditioning Engineers(ASHRAE) and demand savings with reasonable accuracy,the value of the • Laboratory and field testing of the performance of energy- estimates as a basis for decision making can be called into question efficient and baseline equipment if the sources and level of uncertainty of reported savings estimates are not understood and described.While additional investment in • Field inspections,over multiple years,of efficiency activities the estimation process can reduce uncertainty,trade-offs between that constitute the subject programs evaluation costs and reductions in uncertainty are inevitably • Non-site methods such as telephone surveys and interviews, required;thus,improved accuracy(and associated EM&V costs) analysis of consumption data,or use of other data(e.g.,data should be justified by the value of the improved information. from a facility's energy management system) Uncertainty is a measure of the"goodness"of an estimate.Without The 2006 California Evaluation Protocols contain a complete section some measurement of uncertainty,it is impossible to judge an on persistence analyses and can be used to learn more about this estimate's value as a basis for decision making. Uncertainty refers to subject.73 the amount or range of doubt surrounding a measured or calculated value.Any report of gross or net program savings,for instance,has a There are two specific nuances of persistence analyses: halo of uncertainty surrounding the estimated values relative to the • Savings can live as long as the behavior that enables the true values(which are not known).As defined this way,uncertainty efficiency is continued.This can be well past the life of the first is an overall indicator of how well a calculated or measured value purchase of that type of technology,or the opposite—well represents a true value. less than the technical life of a measure,if it is not being used. Thus,it is advisable to evaluate persistence of measures associ- Uncertainty of savings-level estimates is the result of two types ated behaviors(e.g.,continuing to use a new control system of errors: or always buying the efficient product when replacement is • Systematic errors:errors that are subject to decisions and required or responding to information programs)rather than procedures developed by the evaluator and are not subject to just the effective useful life of a technology. chance,also called bias. • As discussed in Section 7.1 on baselines,one baseline option is • Random errors:errors occurring by"chance."One important to use dynamic baselines for early replacement programs.The type of random error is the changes in energy use that can be classic example is a motor with five years of remaining life being due to unobserved influences(i.e.,unobservable independent replaced with a new efficient motor.For the first five years of the variables such as different preferences among people within new motor's effective useful life,the baseline is the motor that a population).Additionally,whenever a sample is selected to was replaced.However,for the remaining years of the EUL,the represent the population—whether the sample is of appliances, baseline would be a common practice or code/standard compli- meters,accounts,individuals,households,premises,or organiza- ant motor,as of year five—when it is assumed that the original lions—there will be some amount of random sampling error. motor would have been replaced.Some may ask whether this Any selected sample is only one of a large number of possible correction in year five is a persistence correction or a baseline samples of the same size and design that could have been correction.Whichever way this may be classified,baseline or selected from that population.For each sample,values calculated persistence,it is important to transparently report measure will differ from the other potential samples simply because of the savings over the lifetime of the motor(and program)with an element of chance in choosing a particular sample.This variability indication that a dynamic baseline was used if such is the case. due to chance factors(the"luck of the draw")is termed random 7.4 CONTROLLING UNCERTAINTY74 sampling error.Random error can be reduced by increasing the sample size. As discussed in Chapter 3,a significant challenge in evaluating energy efficiency programs is the impossibility of direct measurement of Because uncertainty arises from many different sources,it is usually the primary metric—energy(and demand)savings.Consequently, difficult to identify and quantify the effect of all potential sources. The distinction,described above,between systematic and random 7-8 www.seeaction.energy.gov December 2012 sources of error is important because different procedures are program,are different than the rest of the population.As such, required to identify and mitigate each.Evaluation reports often it may be very difficult to find an appropriate control group to identify only uncertainty arising from random error,because this use;thus,evaluation findings may potentially produce biased source of error is usually the easiest to quantify using confidence results.Other forms of self-selection bias include those who intervals and statistical significance tests that are available to provide respond to surveys versus those who do not. quantitative estimates of uncertainty caused by random error(see 4. Analysis(modeling).Estimates are often created through sidebar on statistical terms). statistical models.Some are fairly simple and straightforward On the other hand,uncertainty attributable to systematic errors does (e.g.,estimating the mean)while others are fairly complicated not have a single comparable,quantitative measure.Rather,sources (e.g.,estimating response to temperature through regression of systematic error/bias are specific to individual studies,depending models).Regardless,errors may occur due to using the on equipment used,research staff,or research and data collec- wrong model,creating the wrong counterfactual,assuming tion procedures employed.To assess uncertainty from systematic inappropriate functional forms,including irrelevant information, sources/bias,it is necessary to address the rigor of data collection, or excluding relevant information.For example,in modeling measurements,and analysis.Thus,in summary,uncertainty is energy use of air conditioners,the evaluator may only use typically calculated and reported through the objective analysis of cooling degree days as the independent variable if home type random errors and the subjective analysis of systematic errors. or square footage is not available.Thus,the statistical model will attribute all the observed differences in energy use to 7.4.1 Systematic Errors/Bias temperature,while clearly a portion of the use is attributable to Systematic errors are problematic because they cause a savings the home size.This model will introduce systematic error. estimate to be biased. Bias is simply the extent to which a measure- ment or analytic method systematically underestimates or overesti- For the measurement-related sources of bias,the best solutions mates a value.Systematic errors potentially occur from the way data are sound engineering and analysis practices such as calibration of are measured and/or collected,and/or the way in which analysis meters,use of experienced and trained personnel;use of rigorous (and/or modeling)is conducted.Some specific examples of sources data analysis,and development and rigorous application of quality of bias are below. assurance and control procedures.With respect to bias associated 1. Measurements.At times,equipment used to measure with data collection and analysis,implementing the proper random consumption may not be completely accurate. Human errors selection approaches is very important for avoiding bias and mitigat- (e.g.,errors in recording data) may also cause this type of error. ing systematic errors in a savings estimate. Measurement error is reduced by investing in more accurate 7.4.2 Random Errors measurement technology and training and by more accurately There are many possible reasons why energy use might change recording and checking data. In terms of specific instruments, for any program participant's facility in any particular month the potential magnitude of this type of error is provided by (e.g.,a household may take a vacation,a commercial office may manufacturer's specifications. In most applications,this error add employees,a factory might have higher or lower production or source is ignored, particularly when data sources are utility- change product mix,a school might be in a location with an unusually grade electricity or natural gas meters. However,other types of cold month).An evaluation of a program's impact tries to tease measurements,such as flow rates in water or air distribution out only the changes in energy use that are due to the program,as systems,can have significant errors. opposed to the other factors.While some of the other factors may 2. Data collection.If some parts of a population are not included be easily observable(e.g.,occupancy or weather),many factors(e.g., in the sample,non-coverage errors result.This may cause a vacations)are not. problem because the value calculated from the sample will not accurately represent the entire population of interest. Therefore,when change in energy use is estimated,there is some 3. Self selection.Other sources of bias include using a control chance(hopefully,a very high chance)that the actual program impact group that is not comparable with the treatment group of is being measured. However,there is also some chance that other, participants. For example,in programs where participation is unobserved factors happened to cause the change in energy use. voluntary,a potential source of bias is what is referred to as There is therefore some chance that the evaluation erroneously "self-selection." Participants,by simply choosing to be in the estimates that the program caused the change in energy use when in fact it was caused by other factors(or combinations of the program December 2012 www.seeaction.energy.gov 7-9 and other factors).The same is the case when indicators other than convenient shorthand for expressing the interval believed to contain energy use are being determined as a precursor to determining the actual value.The confidence level is the probability that the inter- energy savings(e.g.,changes in operating hours of lighting systems val actually contains the target quantity.For example,if the savings that have had occupancy sensors installed). estimate is 530 kWh,and the relative precision level is 10%,then the confidence interval is 530±53 kWh).75 One specific source of random errors that is often a concern is that evaluators typically do not have access to an entire popula- Put another way,if a project with a 10%(of baseline energy use) tion of interest,either because the population is too large or the savings estimate with a 20%precision at the 90%confidence level measurement process is too expensive or time-consuming to allow means that the true program energy savings is between 8%and 12% more than a small segment of the population to be observed.As a of baseline energy use,with 90%probability.It is very important to result,they must base their decisions about a population on popula- note that this indication of savings certainty is based on analysis of tion sample data,resulting in sampling errors(one of the major random error and does not address systematic errors. sources of random errors). Note that even if the entire population of interest(e.g.,every household in a program)is observed,there can One way to interpret confidence intervals and precision is as a also be random error due to unobserved variables. measure of risk. For example,with an estimate of 10%savings, there is some risk that the true savings are more or less than 10%; The following are examples of energy efficiency program impact indications of confidence and precision can help quantify that evaluation samples: risk. In general,high levels of confidence that the estimated value • Residential efficiency retrofit program.A sample of homes is falls within the predicted interval can be achieved with wider(less selected for analysis(versus all of the homes that were retrofit- precise)intervals,while narrower(more precise)intervals permit ted).The sample may be organized into homes with similar less confidence. In other words,when all else is held constant,there physical characteristics,similar occupants,similar vintages,or is a trade-off between precision and confidence.76 As a result,any other similarities. statement of precision without a corresponding confidence level is • Commercial building lighting retrofit program.A sample of the incomplete. "spaces"(e.g.,offices,hallways,common areas)is selected for For example,suppose the average savings among participants is determining average operating hours of the lights in each space. estimated to be 1,000 kWh per year,and the analyst determines this • Industrial motors retrofit program.A sample of motors that estimate to have 16%relative precision at the 90%confidence level. were installed is selected for metering of power draw during a The same data set and the same formulas may be used to estimate range of operating conditions and time periods. 10%relative precision at the 70%confidence level.Without report- • New construction building incentive program.All of the build- ing the confidence level,the second uncertainty expression seems to ings in a program are selected for analysis,but only within a have less uncertainty when,in reality,they are identical.In reporting certain time period(e.g.,one month per year). estimates from a sample,it is essential to provide both the precision • Net savings analysis of participants in an efficiency program. and its corresponding confidence level(typically 80%to 90%for A sample of participants and a sample of non-participants are energy efficiency evaluations). selected for interviews. Before concluding this section on controlling uncertainty,two other Random error as a whole can be estimated by using the laws of terms are introduced:internal validity and external validity.The probability.In other words,the potential magnitude of the random above discussion is addressing internal validity—that is,an evalu- error for any value calculated from a sample can(usually)be estimated. ation is internally valid if the observed results are known to have Sample size can be a particularly important aspect of an evaluation been caused by the program as opposed to other factors that may design,and decisions about the sample size are one of the key influ- have influenced the outcome(i.e.,the estimate is unbiased and ences on the random error and the overall uncertainty of the evalua- precise).On the other hand,an evaluation is externally valid if the tion.A larger sample size will increase the statistical significance of the observed outcomes can be generalized and applied to the population estimate,which will reduce the chance of the random error. from which the sample was selected,as well as new populations, circumstances,and future years.The external validity of an evalua- The common factors for reporting random error-associated tion is an important consideration if the evaluation results are to be uncertainty are confidence and precision.Precision provides applied to other programs(without their own potential studies or 7-10 www.seeaction.energy.gov December 2012 evaluations)operating in different years,locations,and/or popula- evaluation requires consideration of all of the aspects of the evalua- tions;for example,external validity answers the question of whether tion process,particularly consideration of 13 of the 14 issues raised the results of a 2010 Oregon CFL evaluation can be applied to a 2012 in Chapter 8 on planning(the last issue is evaluation scale and budget version of the Oregon CFL program. itself).This section,however,discusses budgeting in the context of managing the risks of savings uncertainty.It also provides some In conclusion,evaluation of savings uncertainty is an ongoing process information on the budgets that different states have assigned to that can consume time and resources.It also requires the services evaluation activities,as an indication of the range of typical budgets. of evaluation contractors who are familiar with data collection and analysis techniques.And,of course,reducing errors usually increases 7.5.1 Using Impact Evaluations to Manage Risk evaluation cost.Thus,the need for reduced uncertainty should be The evaluation process must find the right balance between interest justified by the value of the improved information.That is,is the for the most accurate estimates possible with the costs for obtaining value worth the extra cost,and are the evaluation activities them- such accurate estimates.This leads to a basic impact evaluation selves cost effective? question:"How good is good enough?"Asking"How good is good enough?"is a short version of asking"How certain does one have 7.5 EVALUATION BUDGETS: BALANCING to be of the energy savings estimate that results from evaluation THE VALUE OF EVALUATION RESULTS AND activities,and is that level of certainty properly balanced against the UNCERTAINTY amount of effort(e.g.,resources,time,money)it takes to obtain This section provides input on establishing a budget for an impact that level of certainty?"The implication is that energy efficiency evaluation.Establishing a budget(i.e.,setting the funding level)for an investments should be cost effective and evaluation investments STATISTICAL TERMS Census(total enumeration)consists of collecting data from each that is the case.Alternatively,one may be 99%confident that the and every unit in the population(e.g.,metering every replaced light true average savings are within±50%of the estimate of 1,000 kWh. bulb).Sampling only chooses a small part of the units from the population for data collection and analysis. If the estimated outcomes are large relative to the variation,they tend to be statistically significant.On the other hand,if the amount For any value calculated from a sample,a set of descriptive of variability is large relative to the estimated outcome,one is statistics(such as the mean,standard deviation,standard error,and unable to discern if observed values are real or simply random.In a confidence interval)can be calculated.Standard deviation is a other words,when variability is large,it may lead to precision levels measure of variability showing the extent of dispersion around the that are too large(e.g.,more than±100%)for observed estimates mean.In normally distributed data,about 68%of observations are (e.g.,estimated savings)to be meaningful.In an extreme example, within one standard deviation of the mean.Based on the amount if the observed average is 1,000 kWh and the associated precision of variability and standard deviation,a confidence interval can is±150%,true average savings are somewhere between a negative be calculated. 500 kWh(which means the measure actually caused consumption to increase)and 1,500 kWh. To communicate evaluation results credibly,outcomes need to be expressed with their associated variability.Confidence refers to To formalize these relationships,evaluators use a test called the the probability that the estimated outcome will fall within some t statistic.The t statistic is a measure of how reliable a statistical level of precision.Statement of precision without a statement of estimate is.A t test produces a critical value;for example,at a confidence proves misleading,as evaluation may yield extremely 90%level of confidence,the critical value of t is 1.645.When high precision with low confidence or vice versa.For example, the parameter estimate,such as the mean kWh savings,is small after metering a sample of affected equipment,one may estimate relative to its associated variability,the t statistic value is low.When average savings as 1,000 kWh.This is an estimate of the true aver- using a 90%confidence level,if the t statistic is less than 1.645,the age savings.Further,one may able to state that the true average is evaluator concludes the estimated value(e.g.,mean kWh savings) within±1%of the estimate(precision),but only be 30%confident is not reliable:it could be negative,positive,or zero. December 2012 www.seeaction.energy.gov 7-11 should consider risk management principles,and thus balance 7.5.2 Budget-Setting Considerations the costs and value of information derived from evaluation(i.e., With respect to setting an evaluation budget,the primary challenges evaluation should also be cost effective).Impact evaluation is about are typically balancing(1)the cost,time,and effort to plan and managing risk. complete the evaluation(s);(2)the uncertainty of various impact evaluation approaches;and(3)the value of the information gener- Conceptual approaches that draw upon risk management techniques ated by the efforts.Conceptually,this is shown in Figure 7.1,where provide a useful structure for addressing evaluation issues identified the goal is to find the balance point between increasing incremental in this guide.Unfortunately for energy efficiency,risk management investments in evaluation(costs)and decreasing incremental value is hampered by the large number ofdifficult-to quantify aspects of in the evaluation information. efficiency and evaluation,although the tools for addressing these difficulties are improving.Energy supply-side resources have uncer- Most of the value of impact evaluation information is tied to the value tainty and risks as well(e.g.,uncertainties associated with future fuel of energy savings and overall program integrity.In general,low-risk costs). However,perhaps the single most identifiable risk of efficiency projects require less evaluation confidence and precision;high-risk is the inability to directly measure savings,which creates uncertainty. projects require more confidence and precision.The acceptable level of uncertainty is often a subjective judgment based on the value of the Tolerance for uncertainty is driven by how much risk is associated energy and demand savings,the risk to the program associated with with getting the wrong answer.For example,with energy efficiency, over or underestimated savings,and a balance between encouraging the risks include crediting too much or too little savings to the efficiency actions and high levels of certainty. actions that have been taken to comply with an energy efficiency program objective or mandate,such as an energy efficiency resource Thus,an important aspect of evaluation planning is deciding what standard.This can lead to expending too many resources on ineffec- level of risk is acceptable and determining the requirements for tive actions(or the opposite),or simply not obtaining the desired accuracy and a corresponding budget.How much risk is acceptable is outcome(i.e.,less energy consumption). However,there is another usually related to many factors,which raise the following questions: counterbalancing risk if policymakers minimize energy efficiency as • How large is the program in terms of budget and savings goals? an eligible resource because of this measurement risk.This could Larger programs tend to have larger evaluations but smaller take the form of spending too much on EM&V beyond the need for evaluation costs as a percentage. lowering uncertainty and, more important,trying to eliminate this What is the level of uncertainty associated with the expected measurement risk by excluding viable(or promising but difficult savings of a program,and what is the risk that the program to"prove")efficiency resources that are replaced with supply-side poses in the context of achieving(or not)portfolio-savings resources that have different,and perhaps greater,risks associated goals( g•,oa e. are the subject program's projected savings 10/° or 90 with their performance and/or lifecycle costs. °%of the portfolio savings goal)? An important aspect of risk management is always the relative risk • Does the savings determination need to indicate how much (the`As compared to what?"question),which for efficiency includes energy was saved,or just that the savings were above a certain comparing the measurement risk with the risks associated with level?The latter is usually easier to determine than the former. supply-side resources,such as future estimates of fuel costs.Thus, • Is it a new program with uncertain savings or an established pro- another risk to be managed with efficiency evaluation is to avoid gram with well-understood savings?Established programs with requirements that over-specify certainty and/or the use of overly a history of well-documented savings may not require the same conservative versus most-likely assumptions,which can result in level of evaluation that a new program with no history requires. unnecessarily excluding or limiting energy efficiency as an energy Related to this consideration is how much confidence exists in resource strategy. pre-program,projected savings estimates. If a fair amount of Perhaps these two quotes attributed to Albert Einstein best capture effort has gone into feasibility studies and perhaps pre-testing the essence of conducting evaluations: (e.g.,pilots),then less of an evaluation effort may be required. • "Everything should be as simple as it is,but not simpler." Is it adequate to simply verify that the individual projects in a program were installed(and perhaps operating correctly)? • "Everything that can be counted does not necessarily count; Or,on the other end of the cost spectrum,are rigorous field everything that counts cannot necessarily be counted." inspections,data collection,and analyses on all,or a large sample of, projects in a program required? 7-12 www.seeaction.energy.gov December 2012 FIGURE 7.1: Incremental value of information versus incremental cost of evaluation I Value of Improved Certainty Value of Improved Certainty 0 > Cost of Improved EM&V < Cost of Improved EM&V Q O LL Z LL O W >J Balance Point of I Diminishing Returns J H Z W Z INCREMENTAL COST • Is the program likely to be expanded or contracted?A program savings needs to be.As discussed earlier in this chapter,there that may be expanded(i.e., increased in goals and budget) are different ways to calculate and report demand savings,with probably deserves more analyses to confirm whether it should very different levels of effort required. be expanded than one that is likely to receive no additional What are the reporting requirements and who must review funding or n be cancelled. (and approve)evaluation results?While all evaluations should • How long has it been since the last evaluation of the subject have well-documented results,the frequency that savings need program,and has the program changed in the interim? to be reported,and to what audience(e.g.,a regulatory body) • Do savings need to be attributed to specific projects within a can influence the scale of the effort and budget. program(e.g.,a custom retrofit program where each project is Are avoided emissions also to be determined,and will the unique)?If savings values for each project are desired,then a avoided emissions benefits be used in a regulatory program? census evaluation is probably required.This is more costly than As discussed in Chapter 6,emissions can be calculated simply evaluating a sample of projects. or with significant effort and accuracy. If avoided emissions • How long,in months or years,does the evaluation need to be values will be used in a regulated program,the analyses may be conducted?Longer evaluation cycles require more funding.On subject to specific requirements and third-party verification. the other hand,evaluations that have started late,and might • Are other non-energy benefits to be evaluated and quantified? be in a start-up mode,will require smaller budgets(and prob- If this is more than an anecdotal exercise,then additional ably deliver less detailed results). resources will be required. • What is the time interval for reporting savings?For example, • Are the savings reported to be used as"information only"or is reporting annual or monthly savings estimates is usually much there a regulatory or statutory requirement associated with the simpler than reporting hourly savings.This is particularly use of the impact evaluation results?For example: important when deciding how accurate an estimate of demand December 2012 www.seeaction.energy.gov 7-13 — Is there a savings goal,perhaps associated with an en- report,shows total EM&V spending for program administrators who ergy efficiency resource standard,for which the impact responded to their survey in the United States and Canada.Looked evaluation will determine compliance? at in total,for natural gas and electricity programs,the average — Are there cost-recovery and/or lost revenue recovery percentage of program budget spent on EM&V in 2011 was about financial impacts for utilities to be decided upon by a regu- 3.6%.This is consistent with the range of EM&V budgets for both latory body using the impact evaluation results? program administrator efforts and independent third-party evaluator — Does the program administrator have a performance-based efforts that are cited in Tables 7.3 and 7.4. incentive,the monetary value of which will be determined Table 7.3,from a 2010 survey report of evaluation practices and on the basis of the impact evaluation reported metrics? issues in the United States,shows total EM&V 2008 and 2009 bud- Budgets may also be influenced(increased)to accommodate gets for a sample of states.Table 7.4,from the same survey report, follow-up studies aimed at assessing and reducing measurement indicates additional information on EM&V budgets allocated among error,or to pay for additional short-term metering,training of staff, three main types of EM&V activities(process,impact evaluations, or testing of questionnaires and recording forms to reduce data and market research). collection errors.The determination of the appropriate sample size Based on information provided by survey respondents,the range of can be a major factor in setting an evaluation budget.To address EM&V budgets varies significantly between states from very little to this,statistical analyses help evaluators determine the sample size about 6%.The 2008 allocation of EM&V budget among different types needed to ensure that the desired level of precision and confidence of evaluation activities also varied substantially across jurisdictions.Six for key metrics,or factors that determine key metrics,will be states allocated 75%or more of their EM&V budget for impact evalu- statistically significant.Additional resources(more samples)might ation,while three states and the Northwest Energy Efficiency Alliance also be allocated to ensure that"hard-to-reach" portions of the (NEEA)allocated between 50%and 60%of their EM&V budget to population are included in the sample(reducing non-coverage error) impact evaluations.It appears that many states allocate between 10% or devoted to follow-up aimed at increasing the number of sample and 20%of their EM&V budget to market research. members for whom data are obtained (reducing non-response bias). DOE's Uniform Methods Project includes information on steps that Note that these budget figures should be considered as only rough can be taken to increase the accuracy of evaluation results with guidance,as they are mostly self-reported,and the definitions that respect to sampling. are used for what is and is not included in the EM&V budgets varies significantly between states and program administrators.This is While it is difficult to generalize,common practice suggests that particularly true when looking at where the line is drawn between a reasonable spending range for evaluation(impact,process,and activities associated with program implementation and those market)is 3%to 6%of a portfolio budget.However,this should associated with independent evaluation. be considered rough guidance,because evaluation needs and the relative EM&V roles of program administrators and independent 7.6 EVALUATION PRINCIPLES third-party evaluators(and thus how the budget is categorized Reliable evaluation of energy efficiency programs requires between program and evaluation expenses)vary significantly transparency and independence.This results in high-quality infor- between different states and different program administrators.In mation on which business,regulatory,and policy decisions can general,on a unit-of-saved-energy basis,costs are inversely propor be made.Therefore,evaluation processes can be defined by the tional to the magnitude of the savings(i.e.,larger projects have lower following principles: per-unit evaluation costs)and are directly proportional to uncertainty of predicted savings(i.e.,projects with greater uncertainty in the Integral to the portfolio cycle.The evaluation process should be predicted savings warrant higher EM&V costs). integral to what is typically a cyclic planning-implementation- evaluation process.Therefore,evaluation planning should be part 7.5.3 State Evaluation Budgets of the program planning process,so that the evaluation effort can To provide some guidance on overall budgets for evaluation,the not only support program implementation,including the align- following three tables provide information on the EM&V budgets ment of implementation and evaluation budgets and schedules, of various states and program administrators.Table 7.2,from the but also provide evaluation results in a timely manner to support Consortium for Energy Efficiency's(CEE's)2011 annual industry existing and future programs.See Chapter 8 for more information on the portfolio cycle and the integral role of evaluation. 7-14 www.seeaction.energy.gov December 2012 TABLE 7.2: Electric and Gas EMM Expenditure and Budget Dollars, United States and Canada (millions USD)* 7 mco 2010 EM&V 2011 EM&V BUDGETS TOTAL 2011 ENERGY EXPENDITURES EFFICIENCY BUDGETS"" United States 58 154 4,239 Canada 11 32 895 Total 69 186 5,134 COUNTRY 2010 EM&V 2011 EM&V BUDGETS TOTAL 2011 ENERGY EXPENDITURES EFFICIENCY BUDGETS" United States 9 27 782 Canada 1 Less than 1 78 Total 10 27 860 Source:Wallace,P.;Forster,H.J.(2012).State of the Efficiency Program Industry Budgets,Expenditures,and Impacts 2011.Consortium for Energy Efficiency. www.ceel.org/files/2011 CEEAnnual Industry Report.pdf. Notes:*The above table includes only those programs that provided an EM&V dollar figure in response to the survey.Those that provided an estimated percentage of their EM&V activities from their total energy efficiency funding are not included.**Dollar figures in the Total 2011 Energy Efficiency Budgets column exclude load management because CEE did not ask for EM&V expenditures and budgets in the load management portion of the survey. December 2012 www.seeaction.energy.gov 7-15 TABLE 7.3: Ratepayer-Funded Energy Efficiency Budgets and EMU Expenditures for Selected States REGION (MILLIONEFFICIENCY (MILLION BUDGET OF •TAL BUDGET OF • (MILLION (MILLION EFFICIENCY EFFICIENCY BUDGET BUDGET CA 1014.2 80.2 7.9% 1376.7 79.3 5.8% I I I I I I CT 113.6 1.7 1.5% 107.4 1.4 1.3% FL 124.3 ? NA 138.9 0.1 0.1% IA 58.9 3 5.1% 90.5 3.2 3.5% ID 19.7 0.9 4.6% 33.1 0.5 1.5% IL 41.0 1.3 3.2% 67.4 2.2 3.3% MA 148.9 5.1 3.4% 208.5 7.8 3.7% ME 16.8 0.2 1.2% 20.8 0.2 1.0% MN 136.5 1 0.7% 73.7 1.3 1.8% N E EA 97.5 1 1.0% 105.2 1.8 1.7% NY 287.9 7.7 2.7% 421.2 7.6 1.8% OR 76.8 1.6 2.1% 105.4 2.2 2.1% PA ? ? NA 8.7 0.1 1.1% TX 106.4 01 0.9% 101.8 0.2 0.2% W I 140 2.4 1.7% 162.4 4.8 3.0% Source:Messenger,M.;Bhorvirkar,R.;Golemboski,B.;Goldman,C.A.;Schiller,S.R.(April2010).Review of Evaluation,Measurement and Verification Approaches Used to Estimate the Load Impacts and Effectiveness of Energy Efficiency Programs.Lawrence Berkeley National Laboratory.Report LBNL-3277E.http://emp.lbLgov/sites/all/files/Ibnl-3277e.pdf..Source of data in table as indicated in the original report:Consortium for Energy Efficiency(2008),Consortium for Energy Efficiency(2009),and U.S.Census Bureau.See source report for notes. 7-16 www.seeaction.energy.gov December 2012 TABLE 7.4: EM&V 2008 Funding Levels and Allocation Useful retrospective analyses.The evaluation process is Among Activities designed to support the policy goals of the energy efficiency programs being evaluated.As such,evaluations should STATE/REGION • • develop retrospective estimates of energy savings attributable OF to a program in a manner that is defensible in proceedings conducted to ensure that energy efficiency funds are properly PROCESS/ and effectively spent. Evaluation activities should go beyond % MARKET RESEARCH) documenting savings to actually improving programs and providing a basis for future savings estimates. CA 75/15/10 • Adequate resources. Evaluation budgets and resources should be adequate to support the evaluation goals and the level of CT NA quality(certainty)expected in the evaluation results over the FL NA entire time frame that program impacts need to be assessed. • Completeness and transparency.Results and calculations IA 50/30/20 should be coherently and completely compiled.Calculations should be well documented and transparent,with reported ID NA levels of uncertainty.The scope of the documentation should take into account the relevant independent variables that IL 75/15/10 determine benefits and include a properly defined baseline. In addition,documentation and reporting should include all MA 75/15/10 relevant information in a coherent and factual manner that allows reviewers to judge data quality and results.Among the ME 30/50/20 key qualities of a good,transparent analysis are the following: MN NA — Project descriptions indicate the approaches and the vari- ables used to determine energy savings. NEEA 60/30/10 — Critical assumptions are stated and documented. — Documentation is presented in a format that allows the NY 80/10/10 reviewer to follow a connected path from assumptions to OR 50/30/20 data collection,data analysis,and results. — Levels and sources of uncertainty are reported. PA 50/30/20 Relevance and balance in risk management,uncertainty,and costs.The data,methods,and assumptions are appropriate L TX 75/15/10 for the evaluated program.The level of effort expended in the evaluation process is balanced with respect to the value of the WI*** 100/0/0 savings,the uncertainty of their magnitude,and the risk of Source:Messenger,M.;Bharvirkar,R.;Golemboski,8.;Goldman,C.A.;Schiller,S.R. overestimated or underestimated savings levels. Benefits are (April2010).Review of Evaluation,Measurement and Verification Approaches Used to calculated at a level of uncertainty such that the savings are Estimate the Load Impacts and Effectiveness of Energy Efficiency Programs.Lawrence neither intentionally overestimated nor underestimated,and Berkeley National Laboratory.ReportLBNL-3277E.http://emp.lbl.gov/sites/all/files/ the quality of the reported information is sufficient for main- Ibnl-3277e.pdf. taining the integrity of the program being evaluated. *The range depicts answers provided by different respondents from California;the funding also includes evaluations octivitiesfor codes and standards.NA=not available. • Consistency. Evaluators working with the same data and using **Annualfunding for the 2009-2013 cycle. the same methods and assumptions will reach the same con- ***Program administrators may conduct some market research during program clusions. In addition,for efficiency programs that are part of implementation phase;market research is not carried out as part of EM&V activities. broad efforts(e.g., utility resource procurement programs or emissions cap-and-trade systems),energy and demand savings December 2012 www.seeaction.energy.gov 7-17 and avoided emissions calculated from one program area are website at www.eval.org,these principles are summarized here: as valid as those generated from any other actions,whether . Systematic inquiry.Evaluators conduct systematic,data-based demand-side or supply-side.This allows for comparison of the inquiries. range of energy resources,including energy efficiency. • Competence. Evaluators provide competent performance With counterfactual baselines,uncertainty is inherent and savings to stakeholders. estimates are prone to a certain degree of subjectivity.Because of • Integrity/honesty. Evaluators display honesty and integrity in this subjectivity,and possibly a lack of relevant information,some their own behavior and attempt to ensure the honesty and believe that"conservativeness"should be added to the list of integrity of the entire evaluation process. principles for the purpose of counteracting what is seen by some as a . Respect for people. Evaluators respect the security,dignity, natural tendency toward savings inflation.There are many real-world and self-worth of respondents,program participants,clients, incentives for people to over-report savings,and fewer incen- and other evaluation stakeholders. tives working the other way.This subjective bias may be difficult Responsibilities for general and public welfare. Evaluators to keep in check without an explicit directive to be conservative. articulate and take into account the diversity of general and However,many in the evaluation field believe that credibility, not public interests and values that may be related to the evaluation. conservativeness,is the desired characteristic,and that underesti- mates can be just as biased and damaging as overestimates.Thus, 7.7 USING IMPACT EVALUATIONS FOR the correct guidance is to develop the"most likely"result and not PROGRAM FEEDBACK one that is biased to be conservative or aggressive. Impact evaluation results are used to make informed decisions Beyond the characteristics of the evaluation itself,evaluations can on program improvements and future program designs and offer- only be effective if those conducting the evaluations perform their ings throughout the efficiency portfolio implementation cycle. tasks fully and completely,and are free of bias without a stake in the The implementation cycle is one in which programs are designed, outcome,with respect to the performance of the programs under then implemented,and then evaluated.Using the results of the consideration.Related to the characteristics of the evaluation itself, evaluation,programs are reexamined for design changes that may the credibility of evaluators is essential for providing credible findings be needed.This cycle provides a continuing process of program on the results from the program and for providing recommendations improvement,so that the programs match available market opportu- for program refinement and investment decisions. nities and continually improve their cost-effectiveness over time.The impact evaluation planning process is discussed more in Chapter 8. The relationship between the evaluator and the implementers and/or administrators—whose work is being evaluated—needs to Impact evaluations tend to be a retrospective process for be cooperative.This allows for information sharing,access to project determining how a program performed over a specific period of time sites,and for the results of the evaluator to be considered valid by (e.g.,month,season,year); nonetheless,evaluations that produce the implementers and administrators and thus considered as useful results while the program is operating can be very useful.When input for program improvement. However,there will always be some possible,evaluations should be done within a program cycle so that stress in the relationship as(1)the evaluator cannot allow itself to feedback is timely and systematic,benefits the existing program,and be unduly influenced by the implementer/administrator,or for that informs the design of future programs and their evaluation. matter,whoever hires the evaluator,including an entity such as a state regulator;and(2)the administrator/implementer will have a For planning a future program,historical evaluation results can help sense that their work is being judged by the evaluator,because the with program design.However,for estimating how a program will evaluator may very well have a significant say in the compensation or perform,potential studies and feasibility studies are the typical penalties applied to the implementers and administrators. analyses performed. Both of these studies look at what levels of savings are possible from technical,economic,and market- Thus,evaluation ethics are a critical foundation for the activities acceptance perspectives. Potential studies are typically conducted on described in this guide.The American Evaluation Association(AEA) a market-sector(e.g.,residential,commercial,industrial)basis and has a set of guiding ethical principles for evaluators. Located on AEA's feasibility studies tend to be focused on specific customers that may be involved in a particular program. 7-18 www.seeaction.energy.gov December 2012 RESOURCE PLANNING: NORTHWEST POWER AND CONSERVATION COUNCIL Since 1980,the Northwest Power and Conservation Council has inputs for its assessment of the remaining potential for been charged with preparing regional integrated resource plans energy efficiency in the region. For example,the forecast- for the Pacific Northwest states of Idaho,Oregon,and Washington, ing model uses the sales-weighted average efficiency and as well as the western portion of Montana.The council plans energy use of new clothes washers derived from market consist of forecasts of electric loads and a least-cost/risk-resource assessments and program impact evaluations to compute portfolio to meet those regional electric demands over a 20-year the regional energy demand created by this appliance.The time frame.All of the council's plans have relied heavily on energy baseline use and efficiency for clothes washers in the coun- efficiency.Therefore,the council's resource-planning process is cil's energy efficiency potentials assessment is the same as critically dependent on data derived from market characterization that used in the load forecast.Thus,for resource-planning research and impact evaluation results.Results of these stud- purposes,the forecast of future demand and the potential ies support load forecast model calibration and assessments of for reducing that demand start with the same value.One energy efficiency potential in the following ways: implication of this approach for impact evaluation is that • The council uses an end-use econometric load forecast the use in the forecast and potentials assessment of a model. In simple terms,this model calculates future "market average efficiency"(even if it is more efficient electricity sales by multiplying the energy use of a given than local code or federal standards)reflects the efficiency device(e.g.,clothes washer, refrigerator)by the expected choices that are occurring absent future program effects. number of those devices in the region. Market characteriza- As a result,at least in theory,the only difference between tion studies,such as residential and commercial appliance gross and net savings is potential spillover impacts. and equipment surveys,are used to estimate the saturation The council's assessment of energy efficiency potential and of each end-use device included in the forecast model. cost(which is also sometimes based on program evaluation In addition,these surveys are also used to establish the data)is used for resource planning. Because the council baseline characteristics of the existing stock,such as the treats energy efficiency as a resource,it competes directly average efficiency of equipment and the lighting power against generating resources in the agency's resource plan- density of residences and commercial buildings. In addition ning model.That is,rather than reduce the load forecast to"savings,"impact evaluation results,especially those that based on estimates of achievable potential,the amount of provide estimates of both pre-and post-measure adoption energy efficiency to be acquired is based on testing its cost energy use,are used to calibrate the forecasting model to and availability against supply-side resources.Thus,energy actual observed use. savings from impact evaluations are used directly to derive • The council forecast model's estimates of end-use energy the future need for supply-side resources. consumption and efficiency are used as the"baseline" Provided by Tom Eckman of the Northwest Power and Conservation Council 7.8 EVALUATION FOR DEMAND FORECASTING Total ratepayer-funded energy efficiency program spending AND ENERGY RESOURCE PLANNING in the United States(80%of which is targeted to electric end Efficiency is an important energy resource that is being recognized as uses)is projected to rise from$5.4 billion in 2010 to$7.0 such in state and regional energy resource forecasting and planning billion in the low scenario by 2025,$10.8 billion in the medium efforts.77 This is driven in part by at least 24 states having enacted scenario,and$16.8 billion in the high scenario for both electric long-term(three-plus ears specific energy-savings tar ets,which and gas efficiency programs. g ( p years) pg can take the form of a statewide energy efficiency resource standard While the West and Northeast combine for more than 70%of (EERS).78These EERS targets include goals calling for efficiency to efficiency program spending in 2010, by 2025,it is projected reduce electricity consumption on the order of 1%to 3%per year that 50%of the spending will be in the South and Midwest. and/or meet 30%to 100%of increases in demand for electricity. At a national level,the analysis suggests that savings from Looking to the future,a 2012 Lawrence Berkeley National Laboratory energy efficiency programs in the medium scenario have the report found the following:79 potential to offset up to 80%of the projected growth in U.S. electric load in 2020. December 2012 www.seeaction.energy.gov 7-19 An effect of these actions is that efficiency is,and will increasingly Impact studies that consider savings attribution can also be used to become,a resource that has to be accounted for in state and transparently indicate the sources of changes in demand forecasts regional energy-demand forecasts and related capacity,transmission, (whether from voluntary efficiency programs,codes and standards, and distribution resource planning efforts. economic conditions,changes in social norms,or other factors). Cost-effectiveness analyses can be used in integrated resource Market and impact evaluations play an important role in supporting planning efforts for comparing efficiency with other energy resources these resource planning and forecasting activities. Market potential such as conventional and renewable power plants,as well as studies can be used to project savings that will be available from location-specific transmission and distribution requirements. future efficiency efforts and impact evaluations,and market effects evaluations can be used to calibrate demand forecast models. For evaluations to support demand forecast and resource plan- ning efforts,it is important for the evaluation planning efforts to RESOURCE PLANNING: consider the metrics and data requirements,as well as time frames ENERGY TRUST OF OREGON of the people conducting the resource forecasting and planning efforts.These people include forecasters working in utilities,state Energy Trust of Oregon(ETO)impact evaluations,process energy offices,and/or utility commissions;independent system evaluations,and estimates of market effects are used directly operators;regional transmission organizations;and regional energy for integration of efficiency into utility resource planning. and national energy planning organizations.The terminology used ETO provides energy efficiency savings under contract,and in demand forecasting and resource planning,baseline scenarios, develops efficiency supply curves and deployment scenarios historic and future planning time frames—as well as the format and in Oregon for PacifiCorp,Portland General Electric,Northwest definition and granularity of data requirements(e.g.,coincident peak Natural,and Cascade Natural Gas. ETO also provides the versus average peak demand reductions and source versus end-use same services for Northwest Natural in their Washington energy savings)—can be different for the resource planner and territory.This occurs through the following steps: forecaster than they are for a regulator of a retail energy provider. 1. ETO develops supply curves—estimates of the cost and 7,9 DETERMINING NON-ENERGY BENEFITS savings available from various conservation measures Virtually all energy efficiency programs have objectives associ- and sectors. For each integrated resource planning ated with reducing energy use and costs.This guide focuses on (IRP)cycle,estimates of costs are updated as needed documentation of these energy-and demand-related benefits and based on ETO program data.The estimates of savings the associated avoided emissions. However,there is a wide range are adjusted based on the most recent available of other benefits that come from energy efficiency activities;for evaluations. 2. ETO develops deployment scenarios that provide a consumers,these non-energy benefits(NEBs)may actually drive recommended rate of acquisition over 20 years.Impact their interest in efficiency investments. Historically,these NEBs have and process evaluations inform that rate to the extent been,mostly,just subjectively noted as"other benefits,"with little that they show where and how ETO's programs are quantification or documentation of their value. Not putting a value reaching customers.These scenarios show"net"savings to these NEBs,assuming they are positive,can result in negative after market effects,based on trends from recent bias in energy efficiency program investment decisions and less than evaluations. fully effective program participation,designs,and marketing.The 3. Utilities integrate efficiency into their linear planning later would happen because the program implementers might not models for resource selection(as part of the integrated focus on the same benefits that participants focus on when they are planning process)in a variety of ways.The common making decisions about implementing efficiency projects. factor is that the measures that are less costly than NEBs can be categorized as those accruing to utilities(energy generation in the long term are selected.This is affected providers),society as a whole,and to individual participants.80 Some by the savings estimates. research indicates that the value of benefits to society as a whole 4. ETO funding to pursue future conservation is determined and individual participants make up the bulk of the value of NEBs. based on goals coming out of this IRP process.Success is then judged by evaluation estimates of net savings. Provided by Fred Gordon of Energy Trust of Oregon 7-20 www.seeaction.energy.gov December 2012 NEW YORK STATE ENERGY RESEARCH AND DEVELOPMENT AUTHORITY • OF • New York State Energy Research and Development Authority All of these approaches have benefits and drawbacks.The industry (NYSERDA)considers that non-energy benefits can be evaluated standard has been CV and DQ approaches.However,NYSERDA has through a range of approaches: pioneered the joint use of DQ and CA survey methods on its prior • Contingent valuation(CV)survey techniques directly ask New York Energy$martSM Program and is mainly employing DE respondents'willingness to pay for a particular good. approaches currently. • Direct query(DQ)approaches ask respondents to value DQ approaches have somewhat consistently produced NEB values NEBs relative to a given parameter,such as the energy around 50%of the value of the program-induced energy savings. savings achieved on their project.To assist respondents, CA approaches have resulted in similar results for some programs, these surveys often use a scale or provide the dollar value but widely varied results for other programs ranging from 4%to of the energy savings. 340%of the value of the program energy savings.DE approaches, • Conjoint analysis(CA)survey techniques provide which only measure the more quantifiable non-energy impacts, respondents with descriptions of different scenarios can result in relatively lower values as compared to DQ and or levels of NEBs,asking them to either rank or choose CA approaches.NYSERDA's recent commercial/industrial NEBs between the different options presented. Econometric research that found DE-based NEBs of$3 to$4 per natural gas techniques are then applied to calculate the"utility"or MMBtu saved and$5 per MWh of electricity saved.The value of relative value of each attribute. these NEBs to participants ranged from 6%to 39%of the retail • Direct estimates(DE)of non-energy impacts can be made value of their energy savings at typical New York utility rates. in conjunction with impact evaluation site visits,and Source:JenniferMeissner,Program Manager,Evaluation,NYSERDA.2012. generally focus on operations(e.g.,water savings),materials, References include:Energy&Resources Solutions,Inc.(ERS);Left Fork Energy;West Hill Energy&Computing;Megdal&Associates,LLC.;Opinion Dynamics Corporation. and labor hours due to reduced maintenance requirements. (June 2010).Report on Impact Assessments of Two NYSERDA Natural Gas Efficiency These estimates are based on equipment specifications, Programs:Con Edison Natural Gas Efficiency Program&National Grid Low Income operating parameters,and other customer-supplied Natural Gas Efficiency Program.Preparedfor the New York State Energy Research and Development Authority(NYSERDA).http://documents.dps.ny.gov/public/ information,and they are monetized at the project level. Common/ViewDoc.aspx?DocRefld=(CFCD7DE1-8868-46CC-8648-83ACD072A4FD). The following are examples of utility NEBs:81 these effects on marginal peak costs are not always included in • Avoided transmission and distribution capital and operating benefit-cost analyses) costs(particularly in areas with high energy use, high demand • Reduced air emissions,water use,water pollution,and the growth,and/or constrained distribution systems) related emission trading values and/or health/hazard benefits • Reduced line losses,voltage support(reliability),and power • National security improvements through reduced reliance on quality improvements certain fuels. • Customer bill collection and service-related savings such as Examples of efficiency program participant NEBs include the following: avoiding notices,shutoffs/reconnects,and carrying costs on arrearages. Indoor air quality improvements,improved comfort(e.g.,quality of light,less noise,fewer drafts,better building temperature con- The following are examples of societal NEBs: trol),higher productivity and lower rates of absenteeism through • Economic development—for example,job development,both better-performing energy using systems(e.g.,ventilation,building direct and indirect shell,lighting) • Energy price effects,such as stability, lower peak costs,and Reduced equipment operations and maintenance(O&M)costs downward pressure on wholesale energy prices(although because of more efficient,robust systems(although more these price effects can also be considered"energy"benefits, complex systems could require more maintenance). December 2012 www.seeaction.energy.gov 7-21 JOB OF • ' POLICIES The following information is from a report presenting the results of the California Workforce Education and Training Needs Assessment for Energy Efficiency,Demand Response,and Distributed Generation.The assessment was conducted throughout calendar year 2010. Findings of the report are summarized in the table below and also include the following: • By 2020,energy efficiency policies will result in about$11.2 billion of public and private investment,resulting in 211,000 direct, indirect,and induced jobs. • Two-thirds of the direct jobs are in traditional construction trades,one-sixth are in building professional occupations. • There are many more jobs are in traditional trades occupations than in new or specialized occupations. • There are many more incumbent workers than new workers,which indicates the need for training in that there are relatively few slots for job seekers per year compared to the stock of current workers who may need skills upgrading. • There are many more displaced construction workers than jobs,indicating that there are more unemployed,experienced con- struction workers due to the recession than there are job openings created by public/ratepayer investment in energy efficiency. EQUIVALENT • INDUSTRY/OCCUPATION GROUP Residential Building Construction 5,072 7,104 Nonresidential Building Construction 5,342 6,924 Electrical Contractors 319 1,649 LPlumbing,Heating,and AC Contractors 4,859 9,407 Drywall and Insulation Contractors 451 840 Manufacturing 51 574 Advertising and Related Services 956 1,794 LEngineering Services,Architectural Services,etc. 2,118 4,026 Management, Public Administration 1,231 2,449 Office Administrative Services 2,021 3,958 LAll Other Industries 108 212 Total 22,528 38,937 The job increases are compared to the base year of 2009.They are in job person years(i.e.,the number of full-time equivalent jobs for one year).They are not permanent jobs in that they depend on the annual investment of energy efficiency dollars. Source:Zobin,C.(2011).California Workforce Education and Training Needs Assessment.Center for Labor Research&Education,Donald Vial Center on Employment in the Green Economy.www.irle.berkeleyedu/vial. 7-22 www.seeaction.energy.gov December 2012 • Water and wastewater savings 82 work that a program is targeting(e.g.,construction,engineering, • Positive personal perceptions(e.g.,"green,"environmental con- architecture). sciousness)and for commercial businesses and public entities, • Indirect. Indirect jobs are jobs in firms supplying goods and improved public perceptions and the ability to market products services to energy efficiency firms(e.g., manufacturing, and tenant leases accounting). • Avoided capital cost for equipment or building component • Induced.Induced jobs are those created by the demand replacements whose capital costs can be paid from savings. generated by wage and business income from energy efficiency investments and by energy bill savings(e.g.,jobs induced by While most NEBs are considered to be positive,they can also be NEBs,such as grocery store or apparel). negative.83 Examples of negative NEBs include aesthetic concerns associated with fluorescent lamps,"hassle"factor associated with The methods for documenting NEBs tend to fall into one or more of implementing projects,and increased maintenance costs due to the following three categories: unfamiliarity with new energy efficient equipment or the need Measurement of benefits.These methods are used with to operate more sophisticated equipment(e.g.,building control benefits that can be directed, measured,or calculated,such as systems)on a continuous basis. water savings.As with energy savings,a variety of"measure- As noted above,a common impact evaluation approach to NEBs is to ment"approaches are possible,including the use of control list them as possible or probable benefits and not to quantify them. groups and M&V. In some cases,the measurements can be This is typically because of program administrators'(and sometimes made of secondary metrics that are indicators. For example, evaluators')unfamiliarity with methods for quantifying these ben- comfort can be defined via monitoring of indoor temperatures efits,the cost of quantifying them,and the sense that the majority of and humidity to document whether participants' homes, economic benefits are associated with saved energy costs.However, compared with those of non-participants,are within defined the methods for documenting NEBs are improving and expertise in comfort conditions.85 this area is increasing.And,perhaps most important,it is becom- • Modeling.These methods include macroeconomic modeling ing increasingly clear that NEBs can have very high value for those and analysis tools that look at broader societal impacts such as making decisions about efficiency projects and to society as a whole. job growth or modeled estimates of emissions impacts. • Surveys.These are used for documenting many different types In fact,for some programs,it appears that these participant NEBs of NEBs.There is a wide range of surveying approaches to can exceed the energy-related benefits.84 For example,in a com- determine these benefits.These include willingness to pay and mercial office building where the cost of the employee salary and willingness to accept contingent valuation(CV)studies,compara- benefits per square foot dwarfs the cost of energy per square foot, tive or relative valuations,and other revealed preference and an increase in employee productivity of a few percent may be signifi- stated preference approaches.Surveys are used specifically for cantly more valuable than decreasing energy costs by 30%. Including determining relatively subjective program participant benefits these non-energy benefits in evaluations can be quite valuable,given when quantification is difficult and/or expensive.86 However, that energy efficiency programs increasingly are emphasizing these surveys can be used for almost all benefit types where partici- types of participant non-energy benefits when marketing energy pants and non-participants can be asked to provide data(e.g., efficiency programs to customers.Those factors arguably play an how many people they hired for determining job impacts, important role in persuading customers to make the significant whether they believe their indoor air quality is better,if there investments necessary to achieve comprehensive energy savings in are distribution projects that were delayed). a home or business. Currently,several states are including NEBs in their evaluations,but In terms of societal NEBs,one benefit that has generated a great deal not many. In particular for cost-effectiveness analyses,the ACEEE of interest is job creation.Jobs are created as a result of efficiency 2012 review of evaluation practices87 indicated the following: programs in three categories(see the California Efficiency Policy ....while 36 states(including all the states with TRC[total sidebar for examples of results ofjobs analyses): resource cost]as their primary[cost-effectiveness]test)treated • Direct. Direct jobs are in firms that are actually receiving the "participant costs"for the energy efficiency measures as a efficiency program dollars and doing the energy efficiency cost,only 12 states treated any type of participant"non-energy December 2012 www.seeaction.energy.gov 7-23 benefits'as a benefit.... most of those"non-energy"participant Obstacles to the widespread adoption of behavior-based programs benefits were confined to"water and other fuel savings."Only include issues relating to whether these programs can be evaluated 2 states quantified a benefit for"participant 0&M savings" in a rigorous way,the savings persist,and the evaluated results and none quantified any benefits for things like"comfort," shown for one program can be applied to another program.Another "heath,""safety,"or"improved productivity"in their primary SEE Action report88 specifically addresses behavior-based programs benefit-cost test. and prepares recommendations with respect to these issues.The following lists the primary recommendations from that report: Given the potential significant value of NEBs,it is expected that For program evaluation design,the use of randomized con- more jurisdictions will analyze NEBs and that more cost-effectiveness trolled trials(RCT)is recommended.These will result in robust, analyses will take them into consideration,such as in the societal unbiased estimates of program energy savings. If this is not cost test. feasible,it is suggested that"quasi-experimental"approaches 7.10 IMPACT EVALUATION FOR UNIQUE be used. ENERGY EFFICIENCY PROGRAM TYPES For a level of precision that is considered acceptable in 7.10.1 Residential Behavior-Based Programs behavioral sciences research,it is recommended that a null hypothesis(e.g.,a required threshold such as the percent sav- Encouraging people and organizations to do things differently ings needed for the benefits of the program to be considered is a tricky proposition. Energy efficiency is not just about new cost-effective)should be established.The program savings esti- technologies;it is about new behaviors and better decisions.For mate should be considered acceptable(i.e.,the null hypothesis more than three decades,program administrators have offered should be rejected)if the estimate is statistically significant at consumers energy efficiency programs that have used strategies the 5%level or lower. such as subsidies,rebates,or other financial incentives to motivate • In order to avoid potential evaluator conflicts of interest,it is consumers to install technologies and high-efficiency measures. In recommended that results are reported to all interested parties many respects,these programs have focused on affecting behaviors by giving people incentives to undertake efficiency actions. and that an independent third-party evaluator transparently defines and implements the following: However,in the last several years,there has been substantial interest — Analysis and evaluation of program impacts in broadening energy efficiency program portfolios(particularly the — Assignment of facilities(e.g.,households)to treatment and residential programs)to include behavior-based programs that use control groups(whether randomly assigned or matched) strategies intended to affect how consumers use energy in order to — Selection of raw utility data to use in the analysis achieve energy and/or demand savings.These programs typically — Identification and treatment of missing values and outliers include outreach,education,competition,rewards,benchmarking, — Normalization of billing cycle days and/or feedback elements. In some cases,this new generation of — Identification and treatment of households that close programs takes advantage of technological advances in Internet and their accounts. wireless communication to find innovative ways to both capture energy data at a higher temporal and spatial resolution than ever For the analyses of savings, it is recommended to use a panel before and to communicate the energy data to households in creative data model89 that compares the change in energy use for the new ways that leverage social science-based motivational techniques. treatment group to the change in energy use for the control group,especially if the evaluation design is quasi-experimental. These programs are unique in that they may rely on changes to consumers'habitual behaviors(e.g.,turning off lights)or"one-time" With respect to external validity of behavior-based program evalua- behaviors(e.g.,changing thermostat settings). In addition,these lion results,it is possible in theory that a predictive model could be may target purchasing behaviors e. created that allows program estimates to be extrapolated to future programs y g p g ( g.,purchases of energy efficient products or services),often in combination with other years and new populations without actually measuring the savings programs such as rebate programs or direct install programs.These estimates in those years. It is also possible that behavior-based pro- grams could move to a deemed or modeled savings approach over programs are also distinguished by normally being evaluated using large-scale data analysis approaches involving randomized controlled time. However,it is generally believed that the industry is not yet trials or quasi-experimental methods versus deemed savings or at this point,due to the small number of behavior-based programs that have been evaluated using rigorous approaches. M&V approaches. 7-24 www.seeaction.energy.gov December 2012 For more details and recommendations on evaluating behavior- this type of evaluation are surveys and focus groups.The following based programs,please see the 2012 SEE Action report referenced are examples of information topics that may be collected as part of above.Another source of information on behavior-based programs surveys and focus groups(paraphrased from the previously cited and their evaluation are the proceedings of the Behavior, Energy,and California Energy Efficiency Evaluation Protocols): Climate Change Conferences(http://beccconference.org/). Information and education program evaluation topics: 7.10.2 Education and Training (E&T) Programs — Number and percent of customers reached or made aware Education and training(E&T)programs are seen as very important — Number and percent of customers reached who take strategies for expanding energy efficiency's reach as a sustained, recommended actions long-term resource. Education and training programs may be — Number and type of actions taken as a result of the program targeted to either end-use customers or other market actors(e.g., — Changes in awareness or knowledge by topic or subject area, trade allies)whose activities influence the energy-related choices of by type of customer targeted end-use customers.These programs can include advertising,public — Customer perception of the value of the information and/or service announcements,education efforts,training activities(e.g., education received for contractors,building operators,and designers),outreach efforts, — Elapsed time between information exposure and action(s) demonstration projects,and other information or communication- taken by type of customer targeted based efforts. — Attribution of cause for actions taken when multiple causes Typically,E&T programs have one or more of the following general may be associated with the actions taken goals or desired outcomes: — Influence of the program on dealers,contractors,and trade allies • Educate energy consumers regarding ways to increase the — Effects of the program on manufacturers and distributors. energy efficiency of their facilities and activities,and thus convince them to take actions that help them manage their Training program evaluation topics: consumption or adopt more energy-efficient practices. — Pre-program level of knowledge to compare with post- • Inform energy consumers and/or other market actors about program levels program participation opportunities in order to increase — The specific knowledge gained through the program enrollment in these programs. — The relevance and usefulness of the training as it relates • Inform energy consumers and/or other market actors about to the participants'specific needs and opportunities to use energy issues,behaviors,or products in an effort to transform the information the normal operations of the market. — Future opportunities and plans for incorporating the • Train contractors,engineers,architects,and building operators knowledge gained into actions or behaviors that provide on skills, best practices,tools,and other issues related to energy impacts energy efficiency.An example program is the Building Operator — Whether participants would recommend the training to a Certification Program(www.theboc.info). friend or colleague — Participant recommendations for improving the program. Almost every energy efficiency program provides some level of educational and/or informational content. However,education-or Note that programs with large training efforts,or programs designed training-specific programs are typically designed to achieve energy or solely for training,should have evaluation designs that are mindful of demand savings indirectly.Thus,while they are important strategies, the rich literature and methods on evaluating training programs that these programs only indirectly result in energy and demand savings are available from the larger(education)evaluation community. and therefore represent unique impact evaluation challenges. 7.10.3 Market Transformation (MT) Programs For E&T programs,evaluations usually focus on documenting the There are many definitions of market transformation(MT),although degree to which the programs are achieving their desired outcomes it is often considered the ultimate goal of publicly and consumer- within the markets targeted by the program,which is educating and funded efficiency programs.In this guide,the definition of market training people on energy efficiency.The primary mechanisms for transformation is:"a reduction in market barriers resulting from a market intervention,as evidenced by a set of market effects(or December 2012 www.seeaction.energy.gov 7-25 perhaps more specifically a set of market progress indicators)that interventions.This role is equally important for resource acquisition are likely to last after the intervention has been withdrawn,reduced, and MT interventions,but arguably more complex for MT programs, or changed."MT denotes a permanent,or at least long-term,change because the interest is in long-term changes in the market versus in the operation of the market for energy efficiency products and more immediate and direct energy savings for resource acquisition services.As such,their"end-point"can be considered to have programs. occurred if the subject energy efficiency measure(s)or practices are either simply part of common practice(i.e.,all consumers,designers, As a final note for MT program evaluations,and as discussed briefly builders,and operators just"do it")or required per a code in Appendix B and the sidebar"Theory-Based Evaluation:A Guiding or standard. Principle for Market Transformation Evaluation,"a program logic model and a program theory are common components of MT programs Market transformation programs attempt to reduce market barriers and their evaluations.A program logic model is a visual representa- through market interventions.During the 1990s,the focus of tion of the program's theory that illustrates a set of interrelated many energy efficiency efforts shifted from resource acquisition program activities that combine to produce a variety of outputs that to market transformation.Subsequently,there was a shift back to lead to key(in these programs)MT outcomes.Thus, MT program more resource acquisition-focused programs that did not necessarily evaluations should entail collecting information that can be used to include market intervention components,or at least did not include refine the underlying program theory(see sidebar). components defined as such. However,current best practices have all publicly or energy consumer-funded programs having at least some 7.10.4 Codes and Standards Programs MT elements,in that they involve changing how energy efficiency As mentioned in the previous section,from a market transformation activities take place in the marketplace.Thus,MT and other program perspective,the end goal of energy efficiency programs is to have types are now often implemented in a complementary manner. effective energy efficiency actions become either common practice or established in energy codes or standards(C&S). Building energy MT evaluation tends to be a combination of impact,process,and codes and standards set minimum requirements for energy efficient market effect evaluation,and can also include cost-effectiveness design and construction for new and renovated buildings.Appliance evaluations.However,given that the ultimate aim of MT programs (equipment)standards set minimum requirements for energy is to increase the adoption of energy-efficient technologies and consumption of appliances and equipment.States and local govern- practices,MT evaluation usually focuses first on energy efficiency ments generally establish building energy codes applicable in their adoption rates by market actors and second on the directly associated jurisdictions and both states(approximately 15)and the federal energy and demand savings. Market actors that influence end-use government establish equipment standards.90 consumer choices include installation and repair contractors, retailer staffs,architects,design engineers,equipment distributors, Local governments,states,and the federal government also establish manufacturers,and of course the consumers themselves.Also,MT energy efficiency programs to support the development of,imple- programs are dynamic,and thus the nature of market effects can mentation of,and compliance with C&S.These programs include be expected to vary over time.Thus,market effects evaluations, efforts such as emerging technology programs,compliance-enhance- conducted at several points in time,are the primary evaluation ment programs,and stretch(or reach)goal programs,as well as activity usually associated with MT programs.See Appendix B for a training on C&S for building code officials,builders,contractors, description of these types of evaluations. and designers.States are interested in C&S and efforts/programs to support them such as increased code enforcement,because these Evaluation of MT interventions also needs to focus on the support energy savings goals as well as air pollution reduction and mechanisms through which changes in adoptions and energy use GHG mitigation goals. are ultimately induced.This means that considerable attention must be focused on indicators of market effects through market tracking. When determining energy(and demand)savings from C&S,there For example,an MT evaluation might first report changes in sales are two aspects to the project scenarios:C&S requirements and patterns and volumes for particular energy efficiency products as an C&S enforcement/compliance(noting that enforcement is not the indication of program progress in meeting program goals. same as compliance,compliance being the actual goal that results in energy savings).Whether a C&S is actually being complied with can Evaluation also plays an important role in providing the kind be determined through surveys of practices in the field.This can be of feedback that can be used to refine the design of market done through review of permits. However,that can be quite difficult, 7-26 www.seeaction.energy.gov December 2012 given the lack of digitized and organized permits in many jurisdictions THEORY-BASED and because permits are often not"pulled"for many projects such as GUIDING PRINCIPLE - MARKET small HVAC and lighting retrofits.Also,"percent compliance"is not TRANSFORMATION EVALUATION a very definable concept,given that compliance is usually not a"yes or no"issue,as partial compliance is often the case.Thus,energy Theory-based evaluation(TBE),an evaluation approach that performance metrics,such as energy use per square foot for the has been widely used in the evaluation of social programs in buildings of interest(over an extended period of time),can be a much other fields,has seen increased use in the energy-efficiency better targeted metric for gauging whether C&S are being complied industry. It involves a relatively detailed and articulated with and to see if there are any energy savings taking place. program theory,established up front,that specifies the sequence of events a program is intended to cause,along The approaches to determining energy savings associated with C&S with the precise causal mechanisms leading to these events. efforts can involve deemed values,M&V approaches,and large- Evaluation then focuses on testing the congruence of scale consumption data analyses approaches.However,the usual observed events with the overall program theory. approach is a combination of calculations and field data collection that involves the following: A TBE can be considered a process of determining whether a • Establishing an efficiency baseline for each energy efficiency program theory is correct or not(i.e.,testing a hypothesis).For measure based on analyses of common practices or naturally example,with an incentive program,the theory is that paying occurring market adoption rate(NOMAD) a certain level of incentives will result in a certain level of • Defining unit energy savings per measure per year based measure installation,resulting in energy and demand savings. on C&S compliant measures'versus baseline measures' Having well-defined program theories helps focus an evaluation efficiency/performance objective on assessing the validity of those theories,primarily • Defining a market volume baseline for how many units per to see whether a program concept is successful and should be year are installed in the marketplace expanded and/or repeated. • Calculating an energy and demand savings value TBE is particularly well adapted to evaluating the effectiveness • Adjusting for compliance rates(which for a specific code or of market transformation initiatives.This is largely because standard may be the compliance rate found in the market or, market transformation tends to take a relatively long time for a program that supports C&S,may be the improved rate to occur,involves a relatively large number of causal steps as a result of the program,compared against a baseline and mechanisms,and encompasses changing the behavior compliance rate). of multiple categories of market actors—all of which makes it particularly important to focus on specifying and testing a Figure 7.2 illustrates the components of C&S program evaluation. detailed and articulated program theory. In terms of programs intended to improve C&S compliance,the ultimate measure for such efforts is of course better compliance. However,intermediate metrics may also be important to analyze, • Are there changes in attitudes in code enforcement offices particularly given the long-term perspective associated with a C&S and/or perceptions of improved enforcement efforts from effort.The following questions can be raised to discover more about code officials? these intermediate metrics: • Are training programs being offered?What are the results When attempting to attribute achieved savings from C&S to a of pre-and post-test knowledge changes as a result of the specific program or group of programs,it is important to review training?How many trained builders and building officials are the entire C&S process.This can include interviewing stakeholders, there as a result of the program? assessing their effectiveness,and documenting the program contribu- tions. In these attribution efforts,evaluators often use a preponder- • Have there been reach/stretch C&S established,and what is ance-of-evidence approach to forming a judgment,recognizing that the participation level as a result of the program? attribution is elusive when so many parties are involved. December 2012 www.seeaction.energy.gov 7-27 7.10.5 Demand Response (DR) Programs In most cases,DR programs incorporate and rely on electric Demand response(DR)is the reduction of a consumer's energy consumption data for the participants being collected and recorded on use at times of peak use in order to help system reliability,reflect at least a 15-minute basis.Therefore,demand is continuously moni- market conditions and pricing,or support infrastructure optimization tored,including during the DR events.The savings are the difference or deferral of additional infrastructure.DR programs may include between the electricity consumption(demand)during an event and contractually obligated or voluntary curtailment,direct load control, the baseline. and pricing strategies.As noted in the demand savings section Thus,the pertinent question about DR program evaluations is a above, DR programs reduce a utility customer's electricity demand in common one:"What is the baseline?" For DR programs,the baseline response to dispatch instructions or price signals sent to the program is defined as part of the program design,and because participants participants—a"call"for reductions.Thus,the DR peak demand are typically paid on performance(i.e.,how much they reduce their savings is determined for when there is a"call"for program partici- demand),the baseline is also defined in the DR participant contract.91 pants to reduce their energy consumption rate.When this call occurs Typical baselines are the demand for energy the hour before the is referred to as the DR"event." event,the hour after an event,the average of hour before and after, Demand response includes an array of programs relying on various or the demand for a day similar to the one when the event occurred incentive mechanisms.Some DR programs simply involve a contract (e.g.,if the event was on a summer Tuesday afternoon,the baseline with a utility customer to reduce their load when asked(or told)to might be the average of all Tuesday afternoons that summer). In do so;this is historically called interruptible electricity contracts or some cases,the baseline might be determined with a control group, rates. Other DR programs include residential air-conditioner cycling, such as if a large-scale consumption data evaluation approach is where demand savings result from a load control device that actively used and the metric of interest is demand reduction for all of the reduces electricity demand when activated by a command from a participants,not just individual participants(this might be the case central control center. Dynamic pricing is another form of DR,where for evaluation of a rate schedule used to induce DR in consumers). a price signal is sent to consumers,with the expectation that they will 7.1 U.6 Greenhouse Gas (GHG) Mitigation Programs respond to a higher price by reducing consumption(and therefore In Chapter 6,the documenting of avoided emissions—including demand).Most DR programs require active and continued partici- greenhouse gases(GHG)—that result from energy efficiency pro- pation because it is assumed that once a participant is no longer grams is described,including issues associated with documenting enrolled in the program,electric loads revert back to normal patterns. avoided emissions in cap-and-trade programs and determining FIGURE 7.2: Components of C&S savings estimate Naturally Market Occurring Jr Baseline Market Adoption Allocation (NOMAD) Energy Pot ntial Gross Net Net C&S 7 Use Standards Standards Standards Program Savings Baseline Energy Energy Energy Savings by L_ Savings Savings Savings L 5Sa Unit nergy Compliance vings Rate Source:California Public Utilities Commission(CPUC).(April 9,2010).Codes&Standards(C&S)Programs Impact Evaluation:California Investor Owned Utilities'Codes and Standards Program Evaluation for Program Years 2006-2008.CALMAC Study ID:CP00030.06 Final Evaluation Report.Prepared by KEMA,Inc.;The Cadmus Group,Inc.;Itron,Inc.;Nexus Market Research,Inc.;and ENRG,Inc.for the CPUC Energy Division.www.calmac.org/publications/Codes—Standards_Vol It/ FinalEvaluationReportUpdated 04122010.pdf. 7-28 www.seeaction.energy.gov December 2012 whether the avoided emissions data are just for informational associated with steam systems,water pumping,energy recovery, purposes or are part of a benefit-cost analysis or regulatory require- boilers,chillers,street lighting,district heating,metals production, ment.While avoided emissions can be part of the benefits of an and agriculture)and households(retrofits and new construction energy efficiency program,end-use energy efficiency activities can projects associated with cook stoves,water pumping,water purifica- also be part of a GHG mitigation program. tion,refrigerators,lighting,and whole-building projects).These methodologies and other tools for documenting GHG reductions End-use energy efficiency represents a significant,cost-effective from energy efficiency projects are available at the CDM website. approach to reducing GHG emissions.This has been well documented in studies by the Intergovernmental Panel on Climate Change and Unlike the CDM,which is regulated and part of a mandatory GHG numerous other government,private,and non-governmental organi- emissions program,the VCS is a greenhouse gas accounting program zations.92 However,as with the implementation of energy efficiency used by projects around the world to verify and issue carbon credits programs solely for the purpose of energy use and demand reduc- in voluntary markets. It is one of several voluntary GHG account- tions,energy efficiency for GHG mitigation suffers from the difficulty ing programs in place throughout the world.Another is the Gold of cost-effectively documenting the benefits of energy efficiency Standard,which is a carbon-offset standard that certifies compliance activities at levels of certainty that can satisfy regulators,investors, credits created through the CDM and voluntary carbon credit pro- and project proponents. grams.The VCS and the Gold Standard also have methodologies that can be used to document emission reductions from energy efficiency Greenhouse gas voluntary and mandatory mitigation programs in activities.These methodologies,as well as the CDM methodologies, the United States and internationally include energy efficiency as a can be useful tools for evaluators of energy efficiency programs seek- mitigation strategy.Examples include the Verified Carbon Standard ing to document GHG and other emission reductions. (VCS,http://v-c-s.org),the Gold Standard(www.cdmgoldstandard. org),the Regional Greenhouse Gas Initiative(RGGI,www.rggi.org), Investing GHG Funds in Energy Efficiency and the Clean Development Mechanism(CDM, http://cdm.unfccc. There are at least two ways to think about the relationship between int/index.html). In these programs,end-use efficiency is either an end-use efficiency and GHG reductions:(1)assume that energy eligible activity for generating emissions offsets(as in VCS,Gold efficiency is an important societal goal and ask whether GHG Standard,and CDM)or a strategy for investing funds generated reduction programs can lend value to its attainment;or(2)focus through the program(as with the use of allowance auction funds on GHG reduction as the goal and ask whether accelerated energy in RGGI). efficiency is a central element in attaining those essential reductions. While both goals—efficient use of energy and climate mitigation— GHG Offsets are important,in the second context,even though energy efficiency The CDM is the world's largest GHG offset program. It is a is essential to GHG attainment,it does not follow that simply mechanism under the Kyoto Protocol,which set targets for indus- monetizing GHG emissions will actually call forth the full level of trialized countries to reduce their domestic emissions as part of energy efficiency that is both attainable and cost-effective,given the the United Nations Framework Convention on Climate Change.The variety of market barriers associated with energy efficiency.94 basic concept of the CDM is that emissions offsets from projects in developing countries without a commitment to reduce emissions However,auctioning cap-and-trade allowances can raise,in aggregate, are sold to entities in developed countries with emissions-reduction large amounts of funds that can be used in complementary programs commitments at a cost less than these entities'costs to reduce their that do address these barriers.95 These funds can be used in a wide own emissions.The money generated from the sale of offsets pays for range of efficiency support activities such as public benefit programs; projects in the developing countries.Greenhouse gas emissions are energy efficiency resource standards;appliance efficiency standards; reduced in an economically efficient manner,and developing coun- building energy codes;and research,development,and demonstration. tries receive funds to invest in sustainable development projects.93 One example of this approach is what was done the Regional As part of the CDM implementation,methodologies(per the Greenhouse Gas Initiative(RGGI),which was the first market-based terminology of this guide,project-specific M&V plans)have been regulatory program in the United States to reduce greenhouse gas created for documenting GHG reductions from a wide range of emissions.It is a cooperative effort among the states of Connecticut, energy efficiency activities.These cover commercial,industrial, Delaware,Maine,Maryland,Massachusetts, New Hampshire,New and municipal facilities(retrofits and new construction projects York, Rhode Island,and Vermont.Together,these states have capped December 2012 www.seeaction.energy.gov 7-29 REGIONAL GREENHOUSE GAS INITIATIVE In 2009,ten Northeastern and Mid-Atlantic states began the strategic energy"purposes,some states contributed a much Regional Greenhouse Gas Initiative(RGGI),the country's first larger amount to those ends. But from a strictly economic market-based program to reduce emissions of carbon dioxide perspective,some uses of proceeds clearly deliver economic (CO2)from power plants.There are now nine states in the returns more readily and substantially than others. For example, program. RGGI-funded expenditures on energy efficiency depress regional electrical demand, power prices,and consumer payments RGGI Produced New Jobs for electricity.This benefits all consumers through downward Taking into account consumer gains,lower producer revenues, pressure on wholesale prices,even as it particularly benefits and net positive macroeconomic impacts,RGGI led to overall job those consumers that actually take advantage of such programs, increases amounting to thousands of new jobs over time.RGGI implement energy efficiency measures,and lower both their job impacts may in some cases be permanent;others may be overall energy use and monthly energy bills.These savings stay part-time or temporary.But according to our analysis,the net in the pockets of electricity users directly. But there are also effect is that the first three years of RGGI led to over 16,000 new positive macroeconomic impacts as well:the lower energy costs "job years,"with each of the original ten states showing net job flow through the economy as collateral reductions in natural gas additions.Jobs related to RGGI activities are located around the and oil in buildings and increased consumer disposable income economy,with examples including engineers who perform effi- (from fewer dollars spent on energy bills),lower payments to ciency audits;workers who install energy efficiency measures in out-of-state energy suppliers,and increased local spending or commercial buildings;staff performing teacher training on energy savings.Consequently,there are multiple ways that investments issues;or workers in state-funded programs that might have been in energy efficiency lead to positive economic impacts;this cut had a state not used RGGI funds to close budget gaps. reinvestment thus stands out as the most economically benefi- cial use of RGGI dollars. Energy Efficiency in RGGI The states have used their RGGI dollars very differently, in ways Source:Hibbard,P.;Tierney,S.;Okie,A.;Darling,P.(November2011). that affect the net benefits within the electric sector and in the The Economic Impacts of the Regional Greenhouse Gas Initiative on Ten Northeast and Mid-Atlantic States,Review of the Use of RGGI Auction Proceeds from the larger state economy.While all states originally committed to First Three-YearCompliance Period.The Analysis Group.www.analysisgroup.com/ using at least 25%of auction proceeds for"public benefit or uploodedfiles/publishing/articles/economic_impact_rggi_report.pdf. and will reduce COz emissions from the power sector 10%by 2018. States sell nearly all emissions allowances through auctions and invest proceeds in consumer benefits such as energy efficiency, renewable energy,and other clean energy technologies. The sidebar"Regional Greenhouse Gas Initiative"summarizes the results of an impact evaluation of the benefits of the RGGI investments in energy efficiency.As of the writing of this guide,the State of California is also reviewing whether to use its new cap-and- trade GHG program allowance proceeds in a similar manner to the RGGI.These programs and their evaluations represent new opportu- nities for energy efficiency and energy efficiency evaluations. 7-30 www.seeaction.energy.gov December 2012 Chapter 7: Notes 68 Electricity system operators are Independent System Operators 75 Note the counterintuitive implication of this standard definition. (ISO)or Regional Transmission Organizations(RTO)that operate Low precision values correspond to narrow intervals and,hence, a regional energy market,capacity market,or both.A summary describe tight estimates.This can lead to confusion when estimates document on the subject of energy efficiency and forward capacity are described as having"low precision." markets is Gottstein, M.;Schwartz,L.(2010). The Role of Forward Capacity Markets in Increasing Demand-Side and Other Low- 76 Although there is a close relationship between confidence and Carbon Resources Experience and Prospects. Regulatory Assistance precision,they are not direct complements of each other. If the Project(RAP).www.raponline.org/docs/RAP_Gottstein_Schwartz_ confidence level is 90%,there is no reason that the precision needs RoleofFCM_ExperienceandProspects2_2010_05_04.pdf. to be 10%.It is just as logical to talk about 90/20 confidence and precision as 90/10. 69 DEER contains information on selected energy-efficient technolo- gies and measures.It provides estimates of the energy-savings 77 National Action Plan for Energy Efficiency.(2007).Guide to potential for these technologies in residential and nonresidential Resource Planning with Energy Efficiency. Prepared by Snuller Price applications.The database contains information on typical measures et al., Energy and Environmental Economics, Inc.(E3).www.epa.gov/ and data on the costs and benefits of more energy-efficient mea- cleanenergy/documents/suca/resource_planning.pdf.A companion sures.www.deeresources.com. document to the original version of this guide,which addresses resource planning and energy efficiency. 70 Vine, E.; Hall,N.; Keating, K.M.; Kushler, M.;Prahl,R.(August 2011)."Emerging Evaluation Issues Revisited! 2011 International 78 American Council for an Energy Efficient Economy(ACEEE). Energy Program Evaluation Conference Proceedings,August 16-18, (October 2011).State Energy Efficiency Resource Standard(EERS) 2011, Boston, Massachusetts. Madison,Wisconsin: International Activity. http://aceee.org/files/pdf/policy-brief/State EERS Summary Energy Program Evaluation Conference(IEPEC).www.iepec.org/ October 2011.pdf.EERS and other energy efficiency policies are 20llPapersTOC/papers/096.pdf#page-1. described in Section 2.2.1.An up-to-date list of states that cur- rently maintain an EERS can also be found in the Database of State 71 Skumatz,L.A.;Khawaya, M.;Colby,J.(2009).Lessons Learned Incentives for Renewable Energy(DSIRE)at www.dsireusa.org. and Next Steps in Energy Efficiency Measurement and Attribution: Energy Savings,Net to Gross,Non-Energy Benefits,and Persistence 79 Barbose,G.;Billingsley;M.;Goldman,C.; Hoffman,I.;Schlegel, of Energy Efficiency Behavior. Prepared for the California Institute J.(August 2012)."On a Rising Tide:The Future of U.S. Utility for Energy and Environment(CIEE)Behavior and Energy Program. Customer-Funded Energy Efficiency Programs."2012 ACEEE Summer http://uc-ciee.org/downloads/EEM_A.pdf. Study Proceedings;August 12-17,2012, Pacific Grove,California. Washington, D.C.:American Council for an Energy-Efficient Economy 72 One such study is Hirst, E.;White, D. (1985).Indoor Temperature (ACEEE). LBNL-5755E.www.aceee.org/files/proceedings/2012/data/ Changes After Retrofit:Inferences Based on Electricity Billing Data papers/0193-000173.pdf. for Nonparticipants and Participants in the BRA Residential Weath- erization Program. Oak Ridge National Laboratory.www.ornl.gov/ 80 Categories based on:Skumatz,L.A.(1997)."Recognizing All info/reports/1985/3445600444954.pdf. Program Benefits: Estimating the Non-Energy Benefits of PG&E's Venture Partners Pilot Program(VPP)."Proceedings of the 1997 73 California Public Utilities Commission(CPUC).(April 2006). Energy Evaluation Conference.August 27-29,1997,Chicago,Illinois. California Energy Efficiency Evaluation Protocols:Technical, Madison,Wisconsin:International Energy Program Evaluation Methodological,and Reporting Requirements for Evaluation Conference(IEPEC).www.iepec.org/1997PapersTOC/papers/033.pdf. Professionals.www.calmac.org/events/EvaluatorsProtocols_Final_ AdoptedviaRuling_06-19-2006.pdf. 81 Skumatz,L.A.;Khawaja, M.S.; Krop,R.(May 2010).Non-Energy Benefits:Status, Findings,Next Steps,and Implications for Low 74"Uniform Methods Project."(2012). U.S.Department of Energy. Income Program Analyses in California. Revised Report. Prepared for wwwl.eere.energy.gov/deployment/ump.html. Includes information Brenda Gettig,Sempra Utilities.www.liob.org/docs/LIEE Non-Energy on steps to increase the accuracy of evaluation results,particularly Benefits Revised Report.pdf. with respect to sampling. December 2012 www.seeaction.energy.gov 7-31 82 Water savings can come from energy efficiency measures,such as 91 Some DR programs might actually be set up with participants low-flow showerheads and cooling tower upgrades that are designed receiving a discounted cost of energy throughout the year,and then to save energy but also reduce water(and wastewater flows). In penalties if they do not reduce their demand when called upon addition,there are water-saving projects(reducing irrigation needs, to do so. process water requirements)that save energy via reduced water pumping,conditioning,and/or heating. 92 Intergovernmental Panel on Climate Change(IPCC).(2007). Pathways to a Low-Carbon Economy:Version 2 of the Global 83 In fact,in a strict sense,when used in association with incremental Greenhouse Gas Abatement Cost Curve. Prepared by McKinsey savings for programs or measures,NEBs should be described as"net" &Company.www.epa.gov/statelocalclimate/documents/pdf/ of what would have occurred absent the energy efficiency programs; mckinsey_summary_11-19-09.pdf. thus,in some cases they might be positive in an absolute sense,but Intergovernmental Panel on Climate Change(IPCC). (2007). Contri- negative compared to,for example,what would have occurred if the bution of Working Group III to the Fourth Assessment Report of the consumer had purchased a standard energy efficiency product. Intergovernmental Panel on Climate Change,2007.www.ipcc.ch/ 84 Skumatz,L.A.;Khawaja,M.S.; Krop,R.(May 2010).Non-Energy publications_and_data/ar4/wg3/en/contents.html Benefits:Status,Findings,Next Steps,and Implications for Low 93 Schiller,S.R.(August 2011)."Documenting Greenhouse Gas Income Program Analyses in California. Revised Report. Prepared for Emission Reductions from End-Use Efficiency Activities in the Kyoto Brenda Gettig,Sempra Utilities.www.liob.org/docs/LIEE Non-Energy Protocol's Clean Development Mechanism Offset Program."2011 Benefits Revised Report.pdf. International Energy Program Evaluation Conference Proceedings, ss As an example of a quantifiable definition of"comfort,"ANSI/ August 16-18,2011, Boston, Massachusetts.Madison,Wisconsin: ASHRAE Standard 55-2010 defines a range of indoor thermal International Energy Program Evaluation Conference(IEPEC). environmental conditions acceptable to a majority of occupants. www•iepec.org/20llPapersTOC/papers/115.pdf#page=l. 86 Or when the"perception"is the factor that affects the benefit 94 Schiller,S.R.; Prindle,B.;Cowart,R.; Rosenfeld,A.H.(2008). (such as comfort),or the associated perception of a return on "Energy Efficiency and Climate Change Mitigation Policy."2008 investment that may affect program"uptake." ACEEE Summer Study Proceedings.August 17-22,2008, Pacific Grove,California.Washington, DC:American Council for an Energy- 87 Kushler, M.; Nowak,S.;Witte, R(February 2012).A National Efficient Economy(ACEEE).www.aceee.org/files/proceedings/2008/ Survey of State Policies and Practices for the Evaluation of Ratepayer- data/papers/8_306.pdf. Funded Energy Efficiency Programs.American Council for an Energy- Efficient Economy(ACEEE). Report Number U122.www.aceee.org/ 95 Under acap-and-trade program,an overall emission tonnage cap is set for an affected sector or set of emission sources.Governments resea rch-report/u 122. create allowances representing the temporary right to emit one unit 88 State and Local Energy Efficiency Action Network.(May (e.g.,one ton)within the total cap amount.Initial allowance alloca- 2012).Evaluation, Measurement and Verification(EM&V)of tions can be sold(auctioned)by the government or distributed for Residential Behavior-Based Energy Efficiency Programs:Issues and free to affected sources or to others.In the above-referenced paper Recommendations. Prepared by Todd,A.;Stuart,E.;Schiller,S.; (Schiller et.al.),the authors call the activities paid for by the allowance Goldman,C.; Lawrence Berkeley National Laboratory.wwwl.eere. money complementary policies and programs—in other words,pro- energy.gov/seeaction/pdfs/emv_behaviorbased_eeprograms.pdf. grams that while running parallel to cap-and-trade programs are still integral to the success of cap-and-trade and GHG mitigation policies. 89 A panel data model is an analysis model in which many data points over time are observed for a certain population(also called a time- series of cross-sections). 90"Building Energy Codes Program."(2012).U.S. Department of Energy.www.energycodes.gov. 7-32 www.seeaction.energy.gov December 2012 Chapter 8 Impact Evaluation Planning Chapter 8 builds on preceding chapters and presents impact evaluation concepts and the steps involved in the planning process.These include the development of evaluation approaches, budgets, and schedules.The first section discusses how evaluation planning and reporting is integrated into the program implementation process, while the second section presents the concept of a hierarchy of planning documents, starting with an evaluation framework.The third section presents 14 issues and questions that help determine the scope and scale of an impact evaluation.The last section provides some guidance on preparing program impact evaluation plans and site-specific measurement and verification (M&V) plans, as well as a checklist for preparing an evaluation plan. 8.1 INTEGRATION OF EVALUATION INTO THE Figure 8.1 shows the energy efficiency program implementation PORTFOLIO IMPLEMENTATION CYCLE cycle,emphasizing evaluation activities and feedback to the cur- Before describing the evaluation planning process,it is important rent and future programs(portfolio).This figure and the following to understand how it is integral to what is typically a cyclic program discussion concerning evaluation timing are related to the activities or portfolio planning-implementation-evaluation process. In most associated with formal impact evaluations,usually conducted by cases,the overall cycle time frame is determined by a program s—or an independent third-party evaluator.There are other documenta- more likely a portfolio's—funding and contracting schedules(cycles). tion activities that are carried out by program implementers.These These portfolio implementation cycles can be one or two years or include defining baselines,conducting project inspections,and calcu- even longer.The point at which programs are being designed is lating savings for the paying of incentives to contractors and prepar- ideally when the impact,market,and process evaluation planning ing their own(implementer)estimates of claimed savings.There is process should begin.This is primarily so that the program budget, an interaction between these different activities by the implementer schedule,and resources can properly take into account evaluation and evaluator,which is discussed in Section 8.3; however,this text is requirements and opportunities. It is also a way to reinforce the primarily focused on the activities carried out by just the evaluator. concept that evaluation is an integral part of the portfolio process, The first three evaluation activity steps displayed in Figure 8.1 are supporting the portfolio's success through an assessment of the pro- further described below: gram's impacts as well as the program's theory for how savings are Program goal setting.When a program is first envisioned,often to be achieved.961n addition,early consideration of evaluation,on a as part of a portfolio of programs,is when both program goals very practical level,helps ensure that the data collection required to and evaluation goals should be considered. For example,if the support expected evaluation efforts is accommodated at the time of program goal is to save electricity during peak-use periods, implementation. It is also helpful to decide early in the process which the evaluation goals can include accurately documenting how entities will collect and analyze data(e.g.,program implementers,an much electricity is saved during a defined peak period(gross independent third-party evaluator,or both). impact)and how much of these savings can be attributed to the program(net impact). High-level evaluation schedules and Evaluations should be produced within a portfolio cycle or very budgets can also be established at this time. soon after the completion of a cycle.This is so evaluation results Program design.The evaluation design effort should begin can document the operations and effects of the programs in a timely during the program design phase.The objective should be a manner and provide feedback for ongoing program improvement, preliminary evaluation plan that includes a schedule and a provide information to support energy efficiency portfolio assess- budget as well as a description of each work task to be com- ments(including market assessments and potential studies),and pleted.The evaluation issues described in Section 8.3 should help support the planning of future portfolio cycles,load forecasts, be raised,although not necessarily fully addressed,at this time. and energy resource plans. For impact evaluations that examine Whereas a program design is usually completed at this stage,it energy savings of certain measures and program mechanisms,the is likely that the evaluation plan will not be fully defined.This is evaluation information can also be used to inform deemed savings typically because of the necessary priority given to the program values though updating of technical reference manuals(TRMs). design and the need to have the program plan pretty well December 2012 www.seeaction.energy.gov 8-1 completed before the evaluation plan can be finalized. the evaluation planning process should be started well before Another reason why the evaluation design may very well not a program is launched.Trying to collect baseline data on equip- be completed at this time is that it is not unusual,although ment that has been replaced(and usually thrown away)after not best practice,to select the evaluator after the program has a project is completed can usually be ranked between difficult been fully designed(and sometimes after it is implemented). and impossible. Nonetheless,it is ideal to assign evaluation staff or consultants to the program evaluation during program design activities. The overall evaluation plan should be reviewed with program Regardless of when the formal evaluator gets involved,evalua- implementers and,equally important,with the appropriate oversight tion goals and metrics should be considered,and an initial set of body(ies),to ensure that it meets the requirements of policymakers, evaluation priorities(with a rough schedule and budget)should portfolio managers,and/or regulators.These requirements are hope- be established during this portfolio cycle phase. fully set in the evaluation framework,which is a high-level document • Program launch.Program launch is when program materials prepared outside of the portfolio cycle as a guiding document for the and timing strategies are finalized and made ready,implemen- relevant jurisdictions.This framework document concept is discussed tation contracts(if needed)are negotiated,trade allies and key in Section 8.2. stakeholders are notified,and materials and internal processes Although it often is preferable to start an evaluation prior to the are developed to prepare for program introduction and program launch in order to collect baseline information,in most implementation. If the detailed program evaluation plan has cases,either(1)the evaluation and program start simultaneously not been prepared,it should be prepared before the program due to the typical interest in initiating a program as soon as possible, is launched—or if not,soon after it is launched.An outline of or(2)the evaluation starts well after the start of the program due to such a plan is presented in Section 8.4. serial planning of programs first and then evaluation in the portfolio • It is in the evaluation plan that the program-specific evaluation implementation cycle.In general,the sooner the evaluation process issues are fully addressed and resolved, including specifying can begin,the higher the quality of the evaluation and the sooner the data needed to perform the evaluation as well as which information—such as indications of where programs are meeting,or data should be collected during program implementation. not meeting,their goals—can be provided. This is also the time when some baseline data collection can take place, but most likely only after the program's projects In terms of reporting,evaluation information can be summarized and participants are identified. In situations where project site and provided on any time cycle,but most cycles are driven by(1) baseline data collection is required as part of the evaluation, the time required to complete the evaluation work and prepare FIGURE 8.1: Program implementation cycle with high-level evaluation activities schedule,ACTIVITY: ACTIVITY: ACTIVITY: ACTIVITY- Program Goal Setting Program Design Program Launch Implement Evaluation Evaluation Activity: : =,A Evaluation Activity Evaluation Activity: Evaluation Activw ,ft2ts. Prepare preliminary Prepare detailed Implement evaluation evaluation plan evaluation . . collectand reporting and baselineexpectations data as --.-. FOR • PROGRAMS • 8-2 www.seeaction.energy.gov December 2012 reports after the last projects in a program cycle are completed,(2) EVALUABILITY regulatory needs,and(3)a desire to get the information needed to implementers so they can adjust existing programs and design "Evaluability"is an assessment of the probability that evalua- new ones(that may start right after the completion of the current tion information will be available when evaluations are program cycle)using current and relevant evaluation,measurement, actually undertaken.Some data(e.g.,the age of a building) and verification(EM&V)information. For future program designs, can be gathered at any time;some data(e.g.participant spill- projected savings estimates can be adjusted based on program over,current hours of operation)are best gathered at the time evaluation results.Assumptions underlying the energy efficiency of evaluation;and some data must be gathered at the time of potential analysis used for planning at the beginning of the program implementation or they will be lost forever or rendered unreli- cycle can then be updated based on the full net impact analysis. able due to changes in personnel or fading recollection(these These data then feed back into the goal setting and savings potential include free ridership,removed equipment,or non-participant analysis activities,and the cycle repeats to allow for an integrated customer contact).The list below is an example of some of the planning process for future programs. items included in an evaluability assessment template: 8.2 HIERARCHY OF EVALUATION PLANNING • Is there a way to track participants? AND REPORTING DOCUMENTS • Is there a way to track non-participants? • Are specific locations of measures being tracked?Can There are several evaluation planning and reporting documents that they be found? are typically prepared as part of the evaluation effort.The following is • Are program assumptions being tracked on a site-specific an overview of these documents in the context of a portfolio of pro- level (e.g.,hours of operation)? grams that uses public or utility customer funds and has a program . Is the delivered energy-saving service and/or installed administrator with some government(regulatory)oversight.While retrofit being recorded? privately funded portfolios can use many of the concepts presented Does the device recording savings include the outcome here,the formality and requirements for independent evaluation may or result of the activities? be different.Publicly funded portfolios usually have more stringent . Are savings assumptions documented? reporting requirements than privately funded ones. • Is the source of savings assumptions specified? 8.2.1 Evaluation Planning Documents • Are the pre-retrofit or baseline parameters being recorded? Entities such as states,regulatory commissions,utilities,and others • Does the database record the"as-found"values for can establish and document their evaluation requirements in a parameters used to estimate ex ante(projected)savings? hierarchy of documents. Figure 8.2 outlines the hierarchy of these • Does baseline monitoring need to take place? documents and indicates their typical time frame(or time horizon) • Can one of the impact evaluation methods specified in and applicability level(e.g.,state or utility program administra- this guide be used? tor,program or portfolio of programs,or individual projects).The • Are there code compliance or program overlap issues increasing level of detail in this hierarchy of documents provides for savings estimation? indications of the appropriate place for each stakeholder in the overall energy efficiency effort to provide input.For example,all stakeholders may be interested in an EM&V framework but—on content and scope of other EM&V documents(e.g.,annual the other end of the spectrum—only program implementers and portfolio and statewide evaluation reports).The issues that evaluators will generally be concerned with the details of a program would be considered when preparing such a framework are research or M&V plan. described in Section 8.3.This is perhaps the principle document that all stakeholders can focus on and use to provide high-level Descriptions of the documents are as follows: input—the"forest versus the trees"of an EM&V infrastructure. • EM&V Framework.A framework is a primary document that Portfolio Cycle EM&V Plan.This is the planning document that lays out EM&V principles,metrics,allowable approaches, indicates which major evaluation activities will be conducted net versus gross savings issues, reporting requirements and during the evaluation cycle(typically one,two,or three years) schedules,and the roles and responsibilities of various entities. and describes them at a high level,including budget and alloca- An EM&V framework document tends to be"fixed" but can tion of resources and effort between programs,measures, be updated periodically. It often sets the expectations for the market sectors,etc. It lists all the planned, major evaluation December 2012 www.seeaction.energy.gov 8-3 FIGURE 8.2: Hierarchy of EMN planning documents Protocol (IPMVP)measurement and verification options are applied,orjust for inspections if deemed savings values TIMEFRAME COVERAGE are to be used. Multiple EM&V Region, State, Year bm FRAMEWORK or Program Jurisdictions,such as state or regional organizations oversee- Administrator ing energy efficiency portfolios,should also seriously consider developing TRMs.The contents of these TRMs vary,but they PORTFOLIO Region, State, are typically a database of standardized,state-or region-specific AnnualCYCLE EM&V or Pro Multiple Year gram algorithms(deemed calculations)and associated savings estimates PLAN Administrator (deemed savings values)for conventional electric and natural gas energy efficiency measures.The TRMs are used by energy efficiency As Required . . Program program administrators and implementers to reduce EM&V costs (e.g., annual) or Portfolio and uncertainty in how measured savings will be credited with value as they prepare their projected and claimed savings values.Similarly, FM evaluators will also use the information in TRMs to prepare their As Required M&V Project evaluated savings values.TRIM values for individual measures are (e.g., annual) PLAN or Site not always formally vetted in a regulatory process,although this is a good practice.See Section 4.2 for more information on TRMs. Source:Schiller,S.R.;Goldman,C.A.(August2011).'Developing State and National In some cases,different program administrators(e.g.,utilities)in the Evaluation Infrastructures—Guidance for the Challenges and Opportunities of EM&V." 2011International Energy Program Evaluation Conference Proceedings,August16-18, same state or region will use different savings values for the same 2011,Boston,Massachusetts.Madison,Wisconsin:International Energy Program measures.Although different values may be perfectly appropriate Evaluation Conference(IEPEQ.Www.iepec.org/2o11PopersTOC/papers/o73.pdfapage=l. for the same measures because of different baselines,delivery mechanisms,weather,or other factors,it is important for each activities with a schedule for when they are to be conducted. jurisdiction to at least have established and consistent procedures Examples of such major evaluation activities are the impact, for building and maintaining TRMs,with deemed calculations and process,and market evaluations,as well as updates to any savings values that can be accepted for projected,claimed,and technical reference manuals(TRMs).This EM&V plan might evaluated savings estimates.State regulators and program adminis- indicate,for example, (1)which programs would have rigorous trators may also want to consider pooling their resources to support evaluations in each year and which would only have verification development of regional TRMs as well as statewide or regionally reviews,and(2)which programs will undergo process evalua- coordinated TRM updating processes and schedules.Examples tions and which will not. of this are the Regional Technical Forum in the Northwest(www. • Evaluation Activity-Specific Detailed Research Plans. nwcouncil.org/energy/rtf)and the Regional EM&V Forum in the Research plans are created for the major EM&V activities or northeast and mid-Atlantic regions(http://neep.org/emv-forum). studies planned in a given cycle prior to the time each effort Other TRMs available online are listed in Appendix C. is launched. Examples of these plans are(1) program-specific 8.2.2 Evaluation Reports impact evaluation plans that go into substantial detail on what As described in Chapter 3,there are impacts that are reported by evaluation approaches will be used,schedule, budgets,the the program administrators or implementers,and there are impacts data that will be collected,and results that will be reported; reported by the evaluator.Savings estimates for programs tend to be and (2)process,market effects,and market baseline study reported in one or more of the following classifications: plans that would similarly provide sufficient detail to guide the actual implementation of the evaluation activity. • Projected savings:values reported by a program implementer or administrator prior to the time the energy efficiency • Project-Specific M&V Plans. Project-specific plans may be activities are completed required for projects that are part of a custom program and that are selected for analysis and inspection.These are the • Claimed savings:values reported by a program implementer plans that describe the specific activities that would be or administrator after the energy efficiency activities have conducted at a single site when one or more of the four been completed International Performance Measurement and Verification 8-4 www.seeaction.energy.gov December 2012 • Evaluated saving:values reported by an independent third- portfolio cycle evaluation reports to be completed well after the party evaluator after the energy efficiency activities and impact end of the completion of the evaluated cycle and into the time of evaluation have been completed. the next cycle. Evaluators report the evaluation results and,as appropriate,work The above list indicates the typical reports associated with impact with(1)regulators to assess whether goals have been met,(2) evaluations. However,other reports often prepared by evaluators program administrators to implement recommendations for current include process evaluations and market assessments.These are or future program improvements,and/or(3)resource planners to prepared as required to support the implementation of effective understand the historical role and future role of energy efficiency energy efficiency programs.These reports are also made publicly as an energy resource. Reporting also provides information to the available,preferably with summaries also provided in a manner that energy consumers and general public,who are often funding these is accessible to laypersons. efforts,on what has been achieved with their investments. 8.3 PRINCIPLES AND ISSUES THAT Correlated with the evaluation planning documents described above, DETERMINE THE SCOPE OF AN IMPACT the following are the typical types of impact evaluation reports that EVALUATION are prepared. This section presents and discusses issues and principles that should • Project-Specific M&V Reports.These reports document the be defined by each jurisdiction in order to develop a framework impacts determined for a specific project,site,or measure and document for their evaluation activities.As with the last section,it the methods used to determine the impacts.They tend be provides an overview of these issues and principles in the context reviewed with the project administrators and implementers of a portfolio of programs that uses public or utility customer funds before they are finalized,and their results are made available and has a program administrator with some government(regula- on a limited basis to protect the confidentiality of the con- tory)oversight.And,again,while privately funded portfolios can sumer information that is usually included. use many of the concepts presented here,the formality and typical • Impact Evaluation Reports.The results of conducting the requirements for independent evaluation may not be relevant. evaluation activities described in each impact evaluation plan are documented in an impact evaluation report.The report Because of differences in approaches to EM&V and no"cross-border documents the impacts,and perhaps cost-effectiveness,of trading"of energy efficiency savings,it has not been necessary for a program,as well as the methods used to determine the regulatory commissions and administrators of ratepayer-funded impacts. Program administrators and implementers usually energy efficiency programs in states to define a saved unit of energy have opportunities to provide input on these reports.The final in exactly the same way.97 Thus,each jurisdiction(e.g.,state)can, reports are also usually publicly available. and many do,develop EM&V requirements(and TRMs)that are appropriate for their own situations,with the EM&V requirements • Portfolio Cycle Evaluation Reports.The results of carrying out for each jurisdiction linked to the needs of that jurisdiction. the evaluation activities described in an EM&V portfolio plan are documented in a portfolio cycle(e.g.,annual or biennial) How the evaluation issues and principles are defined for each juris- evaluation report.It documents the impact metrics(e.g.,gross diction depends on the specific programmatic and regulatory context and net energy and demand savings,first year,and lifetime)and (including any mandates)found within each jurisdiction,the objec- cost effectiveness associated with the portfolio of programs as tives and scale of the energy efficiency activities being evaluated,and well as the methods used to determine the impacts.Program how EM&V results will be used.For example,one state may have administrators and implementers also usually have opportunities very limited goals for energy efficiency and may not have perfor- to provide input on these reports.The final reports are made mance incentives for their energy efficiency portfolio administrator. publicly available with summaries or synthesis reports provided It also may have a limited level of naturally occurring or mandated in a manner that is accessible to laypersons and with guidance (via codes and standards)energy efficiency activity.Another state on context and interpretation of the evaluation findings.Interim may have established aggressive,long-term energy-savings targets reports are also suggested so that progress indicators can be in legislation,developed a performance-based incentives scheme for provided and problems,if they exist,are identified before the program administrators,and have high energy costs as well as a need end of the program cycle. Interim results can also support plan- for very solid savings data for resource planning purposes.The high ning efforts for the next portfolio cycle,as it is common for the energy cost scenario may also have resulted in a high level of natural December 2012 www.seeaction.energy.gov 8-5 and mandated energy efficiency activity.Given these differences, 7. What is the reporting"boundary"?Are transmission and the first state may only need a limited level of EM&V and use fairly distribution (T&D)losses included,and how"granular"will permissive baselines,while the second state might require very rigor- the results be?See the discussion below in Section 8.3.3. ous(and expensive) EM&V with well-defined baselines.Although 8. What are the schedules for implementing the evaluation in both scenarios EM&V can be valuable to regulators and program and reporting?See the discussion below in Section 8.3.4. administrators and implementers,these kinds of considerations drive 9. What impact evaluation approaches will be used?This simply the type,style,budget,and timing of EM&V requirements. relates to selecting one or more of the impact evaluation These EM&V requirements can be developed by program administra- approaches defined in Chapters 4,5,and 6 and Appendix B tors,agencies with responsibility for overseeing EM&V activities(e.g., (top-down evaluation)that will be allowed,or preferred,for the state utility commissions,energy offices),and other stakeholders,and evaluation of different programs or the portfolio as a whole. then documented in the framework document described in Section 10. What are expectations for savings determination certainty 8.2.Preferably,they all draw from established EM&V principles,such (confidence and precision)?Section 7.4 includes a discussion as those indicated in Section 7.6.To this end,stakeholders in each of uncertainty and managing and defining the risk of savings jurisdiction should work together to address the following 14 issues uncertainty. regarding EM&V.These issues below have been vetted in a number of 11. Which cost-effectiveness tests will be used?Cost-effectiveness jurisdictions and form a good basis for defining the needs of regula- test selection determines what data collection will be required tors,administrators,implementers,and resource planners. However, as part of the impact evaluations.Cost-effectiveness tests are the list should be reviewed and customized for each jurisdiction,and briefly described in Appendix B. to meet exigent regulatory needs. 12. How are evaluated savings estimates applied—looking back/ 1. What are the evaluation objectives,metrics,and research going forward?See the discussion below in Section 8.3.5. issues that support program policies and/or regulations? 13. What are the data management strategies?See the discussion Driving the response to this question is the need to know the below in Section 8.3.6. policy and/or regulatory goals that are the basis for the energy efficiency programs and thus the evaluation of the programs. 14. How are disputes addressed?See the discussion below in See Section 8.3.1.and Section 2.1.1 for discussions of energy Section 8.3.7. efficiency program policy contexts. These issues are presented in what can be considered a linear 2. What are the evaluation principles that drive the effort? sequence,but many are interrelated and the overall planning pro- See Section 7.6 for a discussion of evaluation characteristics cess is certainly iterative.These issues can be addressed through a and principles. variety of mechanisms,such as collaborative efforts(as is the case in 3. What is the scale of the evaluation effort?How much time, Washington),with advisory groups(as is the case in New Mexico),or effort,and money will be spent on evaluation?What is the with regulatory proceedings(as is the case in many states).The end relative allocation of resources among impact,market,and result of addressing the above 14 issues is the evaluation framework process evaluations?Section 7.5 includes a discussion on that documents the evaluation expectations for all stakeholders. defining the scale of the effort and establishing related budgets. Experience has indicated that,if the funding and time requirements 4. Who will conduct the evaluations,how is an independent for reliable evaluations are understood and balanced with informa- li evaluation defined,and what are the relative EM&V roles on needs and accuracy expectations and approved by the primary stakeholders,evaluation efforts can be well-supported and succeed between implementers,evaluators,regulators,stakeholders,and others?See the discussion below in Section 8.3.2. in providing the results desired. 5. Is performance determined on the basis of net or gross 8.3.1 Defining Evaluation Objectives and Metrics savings?What factors are included in defining net savings? Impact evaluations should focus on a program's performance Chapter 5 includes a discussion of net savings definitions, at meeting its key implementation goals and,if desired,provide uses of net versus gross savings,and issues associated with information for future program and resource planning.To this end, determining and using net savings. program managers and regulators need to be assured that the evalu- 6. What are the baselines against which savings are determined? ations conducted will deliver the type and quality of information Section 7.1 includes a discussion of baselines and their needed.This requires consideration of the specific program benefits influence on impact evaluations. (metrics)to be evaluated and reported,the concepts behind the 8-6 www.seeaction.energy.gov December 2012 results the program is expecting to achieve(usually referred to as the • Inform decisions regarding program administrator compensa- program theory),and whether any other related evaluations will be tion and final payments(for regulated programs and concurrently conducted and coordinated. performance-based programs) Under-designed evaluations can waste valuable resources by not • Assess if there is a continuing need for the program reliably providing the information needed—or worse,providing incor- • Provide specific data for demand forecasts and resource rect information.Related to under-designed evaluations are delays that planning in an integrated resource planning(IRP)effort(in the can make it impossible to collect valuable baseline data and postpone format and with baselines and metric definitions consistent the results so that they cannot be used for current program improve- with the resource planners' requirements). ment or future program design.Evaluations can also be over-designed, In practice,the selection of evaluation objectives will be shaped addressing issues that are not priority issues or employing methods by many situational factors. Probably the most important factors that could be replaced by less-costly approaches.Evaluation activities are the program goals—hopefully goals that are well defined and should be prioritized so that evaluation resources—typically limited— quantifiable into evaluation metrics.Some program goals,beyond can be focused on the issues of importance.Like most activities,an numerical metrics of energy and demand savings,cost effectiveness, evaluation that is well defined in terms of objectives and scale,and also and maximizing energy or peak savings within portfolio budgets,are affordable,is more likely to be completed successfully than one with as follows: undefined or unrealistic objectives and budget requirements. • Maximize leverage of portfolio dollars in creating private Setting impact evaluation objectives involves defining the specific investment in energy-efficient products to achieve savings information(metrics)that will be reported from the evaluation.The . Defer specific resource planning needs(e.g., peaking or scale of the evaluation is a more general concept,indicating how baseload power plants,transmission or distribution investments) much effort(e.g.,time,funding,human resources)will be expended on the evaluation.It involves balancing the various needs for Reduce greenhouse gas(GHG)emissions information,accuracy of the information,and timing of the informa- • Maximize the fraction of the population participating in tion against the available resources,including funding. portfolios by sector • Maximize consumer satisfaction. 8.3.1.1 Evaluation Objectives As discussed in the beginning of this guide,evaluations can have In Section 2.1.1,three common examples of program goals are three primary objectives: briefly described. For those three examples,the following are some comments on evaluation objectives: 1. To document the benefits/impacts of a program and determine • Cost effectiveness.All cost-effective energy efficiency evalu- whether the program(or portfolio of programs)met its goals ations for programs with this policy context have to focus 2. To help understand why program-induced effects occurred and resources on determining the long-term cost effectiveness identify ways to improve current and future programs of the programs. Evaluation methodologies can be complex, 3. To support energy demand forecasting and resource planning by as assumptions about the baselines,net-to-gross ratios, understanding the historical and future resource contributions and non-energy benefit factors used in program evaluations of energy efficiency compared to other energy resources. can be critical to the calculations selected for determining Therefore,the first step in planning an evaluation is simply picking cost-effectiveness. which of these objectives(if not all)are applicable,prioritizing them, Energy efficiency resource standard(EERS).Evaluations for and making them more specific to the evaluated program.The follow- programs with an EERS policy objective may be able to focus ing are some typical and more specific impact evaluation objectives: relatively less effort on long-term cost effectiveness, because • Measure and document energy and peak savings the EERS mandates are often promulgated on the assumption (which can only be changed by future legislation)that energy • Document program milestones,such as homes weatherized efficiency will always be less expensive than other resources or people trained up to the level of the EERS target percentage. Evaluations of • Measure and document avoided emissions EERS programs thus focus on the year-by-year tally of progress • Provide data needed to assess cost-effectiveness toward the EERS targets. • Provide ongoing feedback and guidance to the program Target spending budget.Evaluations for programs with target administrator budgets have to focus on the most cost-effective use of the December 2012 www.seeaction.energy.gov 8-7 budgets,so that they can demonstrate that the administrators • Lifetime savings(savings that occur during the effective are getting the maximum benefit from each program dollar. useful life of the efficiency measure): MWh saved during the measure's lifetime, in years Another policy issue that can affect program evaluation budgets is • Monetary savings:dollars saved in electricity or natural gas whether the utility or other program administrators are eligible to costs and maintenance per year. receive a financial incentive(bonus)for managing a successful pro- gram portfolio—or suffer a penalty for not meeting the established The metrics listed above are related to energy savings and avoided goals.The scale of the penalty or bonus can determine the level of emissions.However,other non-energy metrics may also be a subject rigor required in the evaluation as well as the level of attention paid of the impact evaluation;examples include direct or indirect job to the evaluation activities. Evaluators should be cognizant of the gains created over the program implementation period for a training level of intensity that some stakeholders apply to seemingly obscure program or annual water savings associated with a low-flow faucet issues of evaluation methodology when they can affect program program that is focused on water heating savings.In addition,a focus goals and consequences,such as administrator incentives. of many programs is transforming markets(see Section 7.10.3 and However,beyond just what the program goals are and uses of Appendix B),and thus other metrics can include market penetration the evaluation results,other factors also affect the selection of of energy efficiency products or pricing of such products. evaluation objectives: As a reminder,as discussed in Section 7.4,evaluation results,like any • Whether the program is new,expanding,or contracting estimate,should be reported as expected values with an associated • The policy and/or regulatory framework in which the level of variability. evaluation results will be reported 8.3.1.3 Other Evaluation Efforts and Other Programs • The relative priority placed upon the evaluation's comprehensiveness and accuracy by the responsible While this guide is focused on impact evaluations,there are authorities(i.e.,the budget and resources available). other types of evaluations(as described in Appendix B). If other evaluations,such as process or market effects evaluations,are to One approach to address these considerations is to use theory-based be conducted,their plans should be integrated with the impact evaluation(TBE). In short,TBE involves defining the events a program evaluation objectives and plan. If cost-effectiveness analyses are to is intended to cause,along with the precise causal mechanisms lead- be conducted,it is critical to define which cost-effectiveness test(s) ing to these events,with the evaluation,then focusing on testing the will be used,and thus what impact evaluation data are needed. consistency of observed events with the overall program theory(and Furthermore,if more than one program is being evaluated,and thus the goal).Theory-based evaluation is briefly discussed in Section the programs may have some interaction,then the programs,their 7.10.3,in the market transformation program section. evaluations,and the assignment of benefits to one program versus another need to be coordinated to avoid—or at least minimize- 8.3.1.2 Evaluation Metrics double-counting of savings. In terms of reporting impact evaluation results,the key parameters are the units and time frame of the units.Some examples include 8•3.2 Evaluator Roles and Selection the following: While evaluation is relevant to many kinds of business and govern- • Electricity savings:kWh saved per year and per month ment efficiency portfolios,the importance of defining the role and independence of the evaluator is primarily an issue for programs that • Demand savings(example 1): kW saved per month of each are funded with public or energy consumer funds.In such programs, year of program,averaged over peak weekday hours a government regulatory entity,typically a utilities commission, • Demand savings(example 2): kW savings coincident with oversees the actions of the program administrator.This subsection annual utility peak demand, reported for each year of addresses the situation in which a government regulator oversees a the program program administrator's conduct of efficiency activities and wants • Avoided emissions(example 1): metric tons of COz and some level of independent assessment of program/portfolio impacts. SOx avoided during each year of the program It first describes the possible roles of evaluators in program imple- • Avoided emissions(example 2): metric tons of NOx avoided mentation and evaluation,and then presents criteria for defining during ozone season months of each year of the program independent third-party evaluators and criteria for their selection. 8-8 www.seeaction.energy.gov December 2012 Starting with impact evaluation—particularly determining energy However,in general practice,"independent third party"is thought and demand savings—there are two possible sets of results reported to mean that the evaluator has no financial stake in the evaluation after a project or program is implemented:claimed savings and results(e.g.,magnitude of savings)and that its organization,its evaluated savings.Claimed savings,those reported by the program contracts,and its business relationships do not create bias in favor administrator or implementer,are needed not only for possible of or opposed to the interests of the administrator,implementers, reporting to a government/regulatory agency that oversees program participants,or other stakeholders such as utility customers efficiency programs(such as utility program administrator reporting (consumers). However,different states'regulatory bodies have taken to a utilities commission)but also for the basic implementation of different approaches to(1)defining the requirements for evaluators the programs. For example,on a project-by-project basis,claimed who are asked to review the claimed savings and prepare evaluated savings need to be determined in order to pay contractors or savings reports,and(2)who hires that evaluator. consumers whose payments depend on actual implementation and/or achieved savings.Claimed savings also need to be reported A 2012 American Council for an Energy-Efficient Economy(ACEEE) internally to justify expenditures,as would be the case with any study indicated the following:98 organization.Staff or consultant evaluators(e.g.,engineers,analysts, .....there is a great diversity among the states in how they econometricians),working directly for the program implementers handle the evaluation of ratepayer-funded energy efficiency and/or administrators,almost always prepare these claimed savings. programs.This begins with the administration of the evaluation Their role is to directly support the program implementation and function itself,where just over a third of states(37%)feature prepare required internal and external(e.g.,regulatory)reporting. utility[evaluation]administration,36%feature[evaluation] For the above-listed administrator/implementer functions,those administration by the utility regulatory commission or a doing the work may or may not have the formal title of evaluators. combination of the commission and utilities,and over a Evaluator is a very broad term that can describe people conducting quarter(27%)feature[evaluation]administration by some evaluation activities who are part of a consulting,program admin- other government agency or third-party entity. Most states istration,implementation,government,utility,or other type of (79%)rely on independent consultants/contractors to conduct organization.If not called evaluators,they may be known as analysts, the actual evaluations,although a substantial minority(21%) engineers,or M&V engineers,or may simply be part of the adminis- use utility and/or government agency staff. trator's or implementer's program implementation team. Irrespective of how the relationships are determined or who hires whom,the objective is for all parties to the evaluation to believe that Evaluated savings,on the other hand,are only required if some the reported results are based on valid,unbiased information that is entity,such as a government regulatory agency,wants an indepen- sufficiently reliable to serve as the basis for informed decisions. dent third party to either determine themselves or double-check the claimed savings that are determined by the implementer/administra- CLOSING THE LOOP-INTEGRATION OF tor.This leads to several resulting questions: IMPLEMENTER AND EVALUATOR • Who is responsible for which evaluation activities? • Will there be overlap such that an independent third-party There has been a noticeable paradigm shift in evaluation in evaluator,in preparing evaluated savings,only simply confirms recent years.The old model brought in the evaluator at the the work done by the implementer/administrator in preparing tail end of the project to assess delivery,cost-effectiveness, the claimed savings(verification)?Or,will the independent and achievement of stated goals.A different model brings third-party evaluator conduct some of its own data collection the evaluator in at the onset of the program,as an integral and impact evaluation analyses? part of the team. Program goals are linked to specific metrics, • What is meant by an independent third-party evaluator,and which are linked to specific data collection methods.The who retains them to prepare the evaluated savings reports? evaluator can provide feedback in real time to provide timely assessment and recommendations,if needed,for corrective Starting with the last question first,there is no formal definition of actions.This model needs to be balanced with the possible independent or third-party evaluator,as well as there are no well conflicting nature of evaluation goals—the implementer's well-established precedents as to who hires the entity(ies)that goal of understanding and improving the program perfor- provides the evaluated savings reports.The hiring entity could be the mance and a regulating authority's goal of ensuring that the regulator or the administrator,or perhaps some other entity. savings reported are"real." December 2012 www.seeaction.energy.gov 8-9 In terms of the first and second questions listed above—who does bring to the project,availability of the key staff,labor rates and what and whether there are overlaps—there are many options,as overall budget expectation,and approach to the impact evaluation alluded to in the ACEEE study referenced above. Having complete work,including schedule and deliverables. analyses conducted independently by a third party provides perhaps the greatest level of due diligence and integrity for evaluated Budget expectation and approach to the work(instead of a set work savings values.However,such analyses do add costs,and it is not scope,deliverables,and budget)are listed in these criteria because uncommon for there to be overlaps in the determination of savings. it is usually not practical to ask bidders to prepare a complete scope/ Also,because a common objective of evaluation is to improve the budget during the proposal process.This is because either the pro- performance of a program and help with program improvement,a grams to be evaluated are not fully defined and/or the typical first totally independent approach does not directly favor a tight working task is to prepare the portfolio cycle(e.g.,annual or biennial)EM&V relationship between the evaluator and the implementer/admin- plan and the evaluation activity-specific detailed research plans,if istrator.Thus,the selection of an evaluator can require balancing not the evaluation framework itself. evaluation independence(so that the evaluation is considered Traditionally,evaluation consulting firms tend to use econometricians objective)with the desire to have the evaluator close enough to the (professionals who apply statistical and mathematical techniques to process that the evaluation provides ongoing and early feedback problem solving)and engineers.In the last 10 years or so,there has without the implementer feeling"defensive." been acknowledgement of the need for individuals trained in fields One way to look at the relative roles of the different entities involved other than just efficiency and/or with other skills,such as market in preparing claimed and evaluated savings(as well as project analysis and public policy. People with skills such as interviewing and savings)is to consider the roles as associated with oversight or survey data analysis are also needed. administrator activities.These can be generally defined as follows: Many evaluators are members of industry professional organiza- • Oversight activities:activities that are under the purview of tions or are Certified Measurement and Verification Professionals the entity responsible for all energy efficiency programs and (CMVPs).99 The following are two of the professional organizations the associated EM&V implemented by program administrators that energy evaluators participate in(and which will post evaluation in the subject jurisdiction (e.g.,state or utility service terri- RFPs/RFQs): tory).Oversight activities will usually include coordination with . Association of Energy Service Professionals,www.aesp.org the government/regulatory authority.They may also include • International Energy Program Evaluation Conference, feedback or guidance to and from stakeholders(including www.iepec.org. administrators and implementers)about the evaluation plans and implementation as well as the process of approving In addition,the California Measurement Advisory Council(CALMAC) reported results. now offers a directory of evaluators,at www.calmac.org/ • Administrator activities:activities undertaken by the program contractorcontact.asp. administrators during the process of developing,implementing, and conducting M&V activities pertinent to their implementa- 8 Setting the Boundary and Granularity of tion of energy efficiency programs.These M&V activities may Reported rted Results also be known as the primary evaluation activities. When evaluating energy,demand,and emissions savings,it is important to properly define the project boundaries(i.e.,the equip- As suggested in Section 8.2,it is in the evaluation framework that ment,systems,or facilities that will be included in the analyses). these roles and responsibilities are sorted out.One example is how Ideally,all primary effects(the intended savings)and secondary Maryland defined these roles in a simple matrix included in Table effects(unintended positive or negative effects—sometime called 8.1 at the end of this chapter.This is only an example,however,not interactive factors),100 and all direct(at the project site)and indirect necessarily a recommended allocation of duties. (at other sites)effects will be taken into account. From a practical point of view,with respect to energy and demand savings,this In terms of selecting third-party independent evaluators,the usual translates into deciding whether savings will be evaluated for approach is via a request for proposal(RFP)or request for qualifica- specific pieces of equipment(where the"boundary"may include, tions(RFQ)process.Typical selection criteria include qualifications for example,just motor savings or light bulb savings),the end-use and experience of particular staff assigned to the proposed work, system(such as the HVAC or the lighting system),whole facilities,or resources that the bidding organization(or often,bidding team) an entire energy supply and distribution"system." 8-10 www.seeaction.energy.gov December 2012 From an electricity system point of view,the energy and demand then move the evaluation cycle and the program/portfolio cycle into savings at the power plant producing the electricity will be greater better alignment within one or two years.Other evaluation activity than the savings at the end-use(in the facility)due to transmission schedule considerations are the length of the portfolio cycle(one, and distribution(T&D)losses.These average or marginal(which can two,or three years),whether to estimate persistence of savings(see be much higher than the average)losses are on the order of 5%to Section 7.3),whether information is needed for a next portfolio cycle, 20%.101 Thus,the savings at the power plant busbar can be 120% and what information will be included in the evaluation reports. of the end-use savings.Whether T&D losses will be included in the boundary and reported as part of program impacts is something that In terms of reporting requirements,certainly there will be a need can be defined in the evaluation framework. for final program and portfolio reports for each program year and/ or cycle.However,there will likely also be needs for interim reports For avoided emissions calculations,the boundary assessment issues to track the performance of the programs being evaluated and the are discussed in Chapter 6. Boundaries are also important for defin- evaluation itself. ing other non-energy benefits(NEBs)as well.For example,if there is job impact to be reviewed,it needs to be decided whether the jobs There are several considerations for setting up a reporting schedule: are local,statewide, national,or international. • The timing for policy decisions and evaluation planning The time granularity of evaluation analyses relates to whether 15-minute, • The desire to have early feedback for program implementers hourly,monthly,seasonal,annual,and/or lifetime data collection and • Program lifecycle stage(evaluating a first-time program or a savings reporting are required.The"granularity decision"is based on long-established program) how the information from the evaluation is to be used.Annual savings • Evaluation data collection time lags data are generally only useful for an overview of the program benefits. • Regulatory and/or management oversight requirements More detailed data are usually required for cost-effectiveness analyses, Contract requirements for reporting savings for"pay for demand forecasting,and resource planning.For avoided emissions, performance"programs annual values are typical;however,for certain programs,such as smog programs,there are specific seasons or time periods of interest. Timing requirements to use the evaluation results to update energy and demand savings as well as measure life estimates If demand savings are to be calculated,the choice of definition for specific energy efficiency measures in a TRM. (e.g.,annual average,peak summer,coincident peak)is related to time granularity of the evaluation results.Section 7.2 includes a A standard reporting plan would have final reports for each of discussion of the different definitions and describes how this deci— sion greatly influences the data collection requirements,and thus portfolio cycle evaluation reports,with interim reporting on the the effort required to complete the evaluation. impact and portfolio cycle reports.Quarterly or biannual evaluation status reports can also be helpful for keeping all stakeholders 8.3.4 Schedule for Evaluation and Reporting informed. In addition, providing online summary data to the public The evaluation time frame has two components:the time period can be a valuable tool for keeping publicly funded efforts informed over which the evaluation activities will take place and the reporting and perhaps for maintaining public support. schedule.An ideal evaluation schedule begins before the start of the program implementation(to collect any baseline data and set 8.3.5 How Are Evaluated Savings Estimates Applied—Looking Back Or Going Forward? up the overall evaluation infrastructure)and continues for some time after the program is completed to analyze persistence of sav- Estimates of costs and savings from energy efficiency measures are ings. However,the actual timing of the evaluation is influenced by typically made both prior to program implementation(i.e.,projected several—often competing—considerations. savings)and post-program implementation(i.e.,evaluated and/or claimed savings).And,as one would expect,evaluated estimates The first consideration is when to start the evaluation efforts. of savings are considered a more accurate representation of actual Programs and portfolios tend to get into a regular cycle;for example, savings than projected savings.An issue arises when deemed savings the start of each program year may be on January 1,but the evalu- values are used to project and claim energy savings for an energy ation process may not get started until the spring.One approach efficiency measure in a given program year(e.g.,based on per-unit in this situation is to accept the late start of an evaluation for the savings values in a TRM approved for that program year),but an current year and conduct a more limited effort than desired,but evaluation during the program year indicates the TRM per-unit December 2012 www.seeaction.energy.gov 8-11 savings values are too high or too low for the subject measure. administrator goal achievement. However,once TRMs become The question thus becomes"Should the deemed savings value be "stable"(i.e.,fewer and fewer changes are made)the differ- adjusted retroactively to the current program year or only applied on ence should be minor,if there is any at all. a going-forward basis?" Savings from custom projects or programs(where savings are determined after project implementation using agreed-upon Consider the following example: protocols)should use the evaluated values as the savings for all • TRM per-unit savings values developed in 2011 indicate that purposes. the savings from measures verified to have been installed in a program in 2012 are 10,000 MWh. 8.3.6 Data Management Strategies • However,an ex post(after the fact)evaluation indicates that Evaluations are based on data—often in very large amounts. Decision the per-unit savings values in the TRM were overly optimistic, makers often do not need or want to see the detailed calculations or and the actual program savings are 9,500 MWh based on the raw data inputs that drive evaluation results.But all parties need to same number of units verified to have been installed. know that reported results are built on a foundation of sound data • Assuming the verification is correct and all the measures inputs.Proper data management strategies allow administrators, were installed,does the oversight body(e.g.,a regulatory implementers,and evaluators—as well as oversight bodies—to commission)credit the program with 10,000 MWh of savings delve into the underlying data,both to be able to review underlying or only 9,500 MWh for the subject program year? assumptions and to combine the data in new ways as they see fit for current program reviews or future program developments. There are equity issues associated with the above options for how and when to apply updated TRM values.On one hand,the pro- Thus,while data management could easily be defined as"down in the gram administrator/implementer relied on an approved value for weeds;'it can be a major attribute of a portfolio's implementation budgeting and savings estimates;however,for resource planners and evaluation and a major cost.Thus,it deserves attention in the and other stakeholders,what matters most is the most accurate evaluation framework.For stakeholders,the questions addressed in indication of what occurred.While no perfect solution exists,this the framework tend to be associated with:"Will the data resources is an issue that is best to decide upon for inclusion in an evaluation be fully used for current and future project development,such as framework before it occurs during the portfolio cycle.One approach updating deemed savings values?""Will consumer confidentiality be is to use this guidance: properly maintained?"and"What will be publicly accessible?" • Per-unit cost and savings stipulated(deemed)values in a TRM The fundamentals of good data management are the same across should be based on the best available information at the time industries.Within an energy efficiency portfolio,two areas of these estimates and/or calculations are made,and they should primary importance are the ability to compare results across time be determined in a rigorous and transparent manner. It is (longitudinal analysis)and ability to compare results by factors such recognized that TRM values are not static and will be updated as program type or delivery mechanism.Some items are typically from time to time. specified as part of an evaluation plan: • If new information indicates that per-unit cost and savings • Data required stipulated (deemed)values in a TRM should be updated,these new values will be adopted for use in future program savings • Format of data to be provided by tracking systems; projections,claims,and evaluations. compatibility and standardization • Savings claimed in the current program year(based on deemed • Access to data and summaries savings)for measures or programs already implemented before • Data confidentiality protection protocols the TRM is updated are not adjusted retroactively up or down • Data quality assurance and control. for purposes of defining program administrator or implementer goal achievement. In terms of ensuring consistency of data reporting across programs,the • In terms of using the evaluated results for load forecasting and following data components are recommended for standardization: resource planning,estimates with the most up-to-date TRM Measure naming convention(i.e.,the same energy measures, values are used,although this could result in two sets of results end-uses,and applications have the same name from program being reported:one for forecasting/planning and one for to program) 8-12 www.seeaction.energy.gov December 2012 • Measure categorization (i.e., place measures into logical can and do arise,and it is best to define how they will be addressed measure categories,subcategories,and technologies: before they occur.Most jurisdictions will have their own approaches, Category=Lighting,Subcategory=Indoor Lighting, with mechanisms for regular discussions,regulatory hearings,media- Technology=CFL) tions,arbitration,or other solutions.Even a few lines in an evaluation • Sector and building type classification across programs framework document defining the steps for dispute resolution can (e.g.,Sector=Commercial, Building Type=Education— eliminate a great deal of difficulty should a dispute arise. Primary School). 8.4 PORTFOLIO CYCLE IMPACT EVALUATION • Normalization units across programs(i.e., utilize normalization AND M&V PLAN OUTLINES units as the basis of quantity or measure size,such Flow-GPM for water flow measures or Ctrl-kW for occupancy sensors). The evaluation planning documents should clearly present the evalu- • Program delivery methods across programs(e.g., program ation efforts and details of the actions to be undertaken during the types are prescriptive rebates,direct install,and point of sale) evaluation activity,as well as consideration of regulatory(reporting) requirements.A plan is a stand-alone decision document,meaning • Indication of project dates(e.g.,date of installation). it must contain the information the evaluator and others need to understand what is to be undertaken,why,when,and how.Plans While there is not currently a single reporting standard for energy are also important historical documents since it is not unusual for efficiency programs,national,regional,and international reporting programs with long lifecycles to undergo staff changes over the specifications are under development.In the meantime,three data- course of the program. base projects that demonstrate good data management practices are as follows: The following subsections outline the contents of a portfolio cycle • Standardized Project Tracking(SPT)database in California. impact evaluation plan and M&V plan.The M&V plan is included California has long maintained various databases related to because it is a common approach for calculating gross energy savings. energy use and energy efficiency. In recent years,the California Following the M&V plan outline is an evaluation planning checklist. Public Utilities Commission has worked closely with the program administrators to create standard claims-reporting 8.4.1 Portfolio Cycle Impact Evaluation Plan and templates. Recent versions of the SPT allow all stakeholders Report Outlines in California to review program savings and costs at the The following is an outline that can be used to produce an impact measure, program,and portfolio levels.See the websites at evaluation plan. http://open-emv.com/and www.cpuc.ca.gov/PUC/energy/ Part A.Program Background Energy+Efficiency. 1. Short description of the program(s)being evaluated (e.g.,the • The Regional Energy Efficiency Database(REED). Developed market,program delivery approach,technologies,budget, by the Regional EM&V Forum, REED has commitments from objectives) ten jurisdictions for a common reporting system of energy 2. Presentation of how the program will achieve its objectives,the efficiency savings and associated impacts to support a range of program theory energy and environmental policies.See the website at 3. List of the technologies offered by the program http://neep.org/emv-forum. 4. Program schedule • The Northwest Regional Technical Forum(RTF).The RTF is an 5. Numerical energy and non-energy savings projections. advisory committee established in 1999 to develop standards to verify and evaluate conservation savings.See the website at Part B.Evaluation Overview www.nwcouncil.org/energy/rtf/Default.htm. 1. List of evaluation objectives and how they support program goals 8.3.7 Addressing Disputes 2. List of which metrics will be reported (e.g.,annual MWh, Disputes can arise from the process used to develop impact monthly peak kW,annual therms,annual COz) evaluation results and/or the results themselves. Disputes are best 3. Description of verification activities addressed before they arise through understood agreements on 4. Version of the TRM to be used and/or any TRM development/ how the evaluations will be conducted(i.e.,defined in an evaluation review activities framework and through good communication).However,disputes December 2012 www.seeaction.energy.gov 8-13 5. Gross and net impact evaluation approaches selected for 5. An activities timeline with project deliverable dates, determining energy(and demand)savings including reporting activity and key milestones,including 6. Methodology for calculating non-energy benefits such as communications with administrator/implementer avoided emissions,as appropriate 6. Detailed budget and schedule 7. List of primary factors that will be considered in analysis of 7. Evaluation team—information concerning the independence of gross and net savings(e.g.,weather,occupancy,free riders, the evaluator. spillover)as well as list of major assumptions 8. Description of how program impact results will be combined to The final product or output of an evaluation is an evaluated savings report portfolio impacts,addressing the need for adjustments report.The following is a sample report outline: such as accounting for program overlap or other factors • List of Figures and Tables 9. Expectations for overall certainty of savings estimates • Acronyms • Abstract 10. Assumptions concerning availability of data and other • Acknowledgments information provided by the administrator/implementer; relative roles of evaluator and administrator/implementer 1. Executive Summary(Include highlights of key recommended 11. Budget and schedule summary improvements to the program,if relevant) 12. Listing of evaluators(if known)or evaluator selection method. 2. Introduction • Program Overview(e.g.,program description,objectives) Part C.Detailed Evaluation Approach,Scope,Budget, • Evaluation Objectives and Methods Schedule,and Staffing • Report Structure (This is the detailed presentation of evaluation activities to be 3. Study Methodology undertaken,including the evaluation approach to be used.) • Data Collection Approach(es) 1. Gross impact savings analysis—a description of the data • Analysis Methods collection and analysis activities and approaches(if an M&V • Limitations,Caveats evaluation approach is selected, identify the IPMVP option to 4. Key Evaluation Results with metrics and realization rates be used) (answers for all of the questions specified for the evaluation. 2. Net impact savings analysis—a description of how spillover, This could include several sections on findings. Findings could free ridership,and other effects will be addressed in the be presented for each method used, by program components evaluation activities and in the data analysis(as appropriate) covered,by market segments covered,and so forth,followed by a section on integrated findings or organized and presented 3. Data collection,handling,and sampling: by the different observed effects or type of results.) a. Measurement collection techniques 5. Synthesis of analysis and findings as well as implications of b. Sampling approach and sample selection methods for findings each evaluation activity that includes sampling efforts 6. Recommendations(should include clear,actionable,and (as appropriate) prioritized recommendations that are supported by the analysis) c. How the comparison group or non-participant information 7. Summary and Conclusions will be used in the evaluation(s)and in the analysis 8. Appendices(examples listed below): (as appropriate) d. Data handling and data analysis approach to be used to • Recommended improvements to the evaluation process, address the researchable issues including any lessons learned for future evaluation studies 4. Uncertainty of results—presentation and discussion of the • Detailed documentation of the research design and threats to validity, potential biases, methods used to minimize assumptions,data collection methods,evaluation analysis bias,and level of precision and confidence associated with methodology,references,and results tables the sample selection methods and the evaluation approaches; • Survey or interview instrument,coding scheme,and quality control information should also be included here compiled results tables and data 8-14 www.seeaction.energy.gov December 2012 • Sources and quality(caveats on data)of primary and The M&V plan should include an overall project description,facility secondary information equipment inventories,descriptions of the proposed measures,energy- • Details on quantitative data analysis:analytical framework, savings estimates,a budget,and proposed project implementation and modeling approach,and statistical results M&V schedules.A project-specific M&V plan should also demonstrate • Qualifications to results that any metering and analysis will be done consistently,logically,and • Possible sources of overestimation and underestimation with a level of accuracy acceptable to all parties. • Analysis of reliability of energy savings estimates,treatment The following is a sample M&V plan outline: of issues that threaten reliability of results(e.g.,double counting, use of savings factors,and synergistic effects) 1. Description of project, measures to be installed,and How attribution was addressed(for net savings impact project objectives • evaluation) 2. elected IPMVP option and measurement boundary • Other assumptions and justifications. 3. Description of base year conditions,data collection, and analyses 8.4.2 Project-Specific M&V Plan and Report Outlines 4. Identification of any changes to base year conditions and If the M&V gross impact evaluation approach is selected,an M&V how they will be accounted for in the analyses plan needs to be prepared for each project selected for analysis.This 5. Description of reporting period conditions,data collection, section discusses the M&V planning process for individual projects and analyses and presents an M&V plan outline and report outline. 6. Basis for adjustments that may be made to any measurements M&V activities involve the following steps: and how this will be done 1. Selecting one of the four IPMVP options for the project.The 7• Specification of exact analysis procedures options define general approaches to documenting savings. 8. Metering schedule and equipment specifications 2. Preparing a project-specific M&V plan that outlines the details 9. Description of expected accuracy and how it will be determined of what will be done to document savings. 10. Description of quality assurance procedures 3. Defining the pre-installation baseline,including equipment 11. Description of budget and schedule and systems, baseline energy use,and factors that influence 12. Description of who will conduct M&V. baseline energy use. 4. Defining the reporting period situation,including equipment The tables at the end of this chapter summarize what could be and systems, post-installation energy use,and factors that contained in the M&V plans.Table 8.2 lists general requirements influence post-installation energy use.Site surveys;spot,short- for an overall M&V plan.Table 8.3 lists requirements that could be term,or long-term metering;and/or analysis of billing data can addressed in the M&V plan for each measure(e.g.,building lighting also be used for the reporting period assessment. retrofit,building air-conditioning retrofit,control system upgrade) 5. Conducting periodic(typically annual) M&V activities to verify that is included in the project being evaluated. More information on the continued operation of the installed equipment or system, the contents of an M&V Plan can be found in the IPMVP determining current year savings,identifying factors that may (www•evo-world.org). adversely affect savings in the future,and estimating savings 8.4.3 Checklist of Planning Decisions for an for subsequent years. Impact Evaluation A project-specific M&V plan describes in reasonable detail what Table 8.4 presents a checklist for preparing an impact evaluation will be done to document project savings. It can be a plan for each plan.The list is organized around the decisions associated with the energy efficiency measure included in the project(e.g.,when a gross savings calculation,net savings calculation,calculation retrofit isolation approach is used).Or,it can cover the entire project of avoided emissions,and generic issues. (e.g.,when the whole-facility analysis approach is used). In either case,the M&V plan will consider the type of energy efficiency measures involved and the desired level of savings certainty. December 2012 www.seeaction.energy.gov 8-15 TABLE 8.1: Maryland Public Service Commission EM&V, Data Tracking, and Reporting Act Roles and Responsibilities for Utilities and Statewide Evaluation Consultants AND/ORUTILITIES PSC STATEWIDE TASK DELIVERABLE (OR THEIR EVALUATION CONSULTANT) • STATEWIDE STUDIES Prepare statewide baseline study X102 Prepare additional statewide market assessments(e.g.,market X impact studies)and updates(bi-or tri-annual) Prepare statewide technical reference manual(TRM) X103 (and annual or bi-annual updates) Review TRM and updates of TRM X Approve TRM(and review statewide market assessments) X PLANNING AND MANAGEMENT Prepare Master Evaluation Schedule for utility and statewide evaluation contractor activities;include overview of reporting X X X schedule with annual and tri-annual portfolio reporting,as well as semi-annual interim reports and/or presentations Develop utility(statewide and individual utility)impact and process evaluation plans;including gross and net energy and demand (including peak demand from distributed resources)savings, X cost-effectiveness analyses,database and reporting protocols, survey templates,and schedules Review utility evaluation plans X Develop plan for due diligence(quality assurance/quality control, QA/QC)of utility impact results(energy and demand savings,cost X effectiveness),including verification approach(with sampling plan) and schedules for review of utility submittals and reporting to PSC 8-16 www.seeaction.energy.gov December 2012 UTILITIES PSC • ' EVALUATION• •TASK • ' • PSC • • PLANNING AND MANAGEMENT(CONTINUED) Develop plan for determining and reporting additional outcomes (e.g.,system reliability,;T&D and generation needs impacts;emissions X X avoidance;price mitigation;jobs impacts;effects on ratepayers, especially low-income;coordination with federal stimulus funding) Review and approve utility and PSC evaluation consultant plans X Coordinate all utility evaluation efforts X Coordinate statewide due-diligence(QA/QC)impact evaluation/ X verification efforts PROCESS EVALUATION Prepare program process evaluations X Conduct(independent)customer and trade-ally satisfaction surveys X and reports Review customer and trade-ally satisfaction survey results X x DATABASES Prepare a data reporting,interface,and database plan that includes coordination between the utilities'and PSC's statewide evaluation X X contractor database(s)and utility databases Design,implement,and maintain utility primary program management and tracking database(s)with project and program X data(includes individual utility databases and a statewide database operated by utilities) Design,implement,and maintain statewide data management and quality control database of information"uploaded"from utility database(s)and used for(1)obtaining and managing data for X due-diligence activities,and(2)establishing a public web-accessible database and reporting system with aggregated,higher-level informa- tion on program impacts(e.g.,statewide energy and demand savings) Review and approve statewide database and reporting plan I X December 2012 www.seeaction.energy.gov 8-17 AND/ORUTILITIES PSC STATEWIDE TASK DELIVERABLE (OR THEIR EVALUATION CONSULTANT) • PRIMARY DATA COLLECTION AND IMPACT ANALYSES: ENERGY EFFICIENCY(EE)AND DEMAND RESPONSE(DR) rPROGRAMS Prepare ex ante(preliminary)savings estimates X Conduct primary data collection and site baseline and ex post(after the X fact)verifications for energy efficiency and DR projects Prepare persistence of savings analysis:Conduct primary X data collection Prepare analyses and documentation of project,program,and X portfolio gross and net energy and demand savings,cost effectiveness INDEPENDENT DATA COLLECTION AND IMPACT ANALYSES Conduct quality control and due diligence of utility analyses and documentation of project,program,and portfolio gross and net energy and demand savings and cost effectiveness;inspect sample of project sites X and review primary data and analyses,prepare verified achieved versus claimed savings and cost-effectiveness report per reporting schedule I OTHER OUTCOME ANALYSES Prepare additional efficiency and DR program/portfolio outcome reports:System reliability,T&D and generation needs impacts;emissions avoidance; price mitigation;jobs creation;impacts on low-income X ratepayers;and leveraging of federal stimulus funding such as American Recovery and Reinvestment Act(ARRA)funds CREPORTING Prepare utility interim semi-annual and final annual(and tri-annual) reports of energy efficiency program and portfolio net and gross impacts X and cost-effectiveness evaluation results Prepare semi-annual(interim)and annual(final)reports of verified achieved utility program and portfolio results for energy and demand X savings and cost effectiveness. Prepare annual report on additional Empower MD outcome results(e.g., X avoided emissions,reliability and job impacts) Review utility and PSC consultant semi-annual reports; review and approve X utility and PSC consultant annual(and tri-annual)reports 8-18 www.seeaction.energy.gov December 2012 UTILITIES PSC STATEWIDE TASK • - • - THEIR EVALUATION CONSULTANT) • BEST PRACTICES Participate in quarterly(or semi-annual)impact evaluation process review X X X and improvement meetings Prepare best practices recommendations for improvements to X X evaluation processes Review and approve best practices recommendations for program X modifications and improvements OTHER Prepare materials and reports in support of PSC analysis of efficiency programs X X Organize and conduct periodic statewide public workshops on evaluation X X results of energy efficiency programs Source:Consensus Report on the EmPower Maryland EM&V Process.(June 24,2009). . December 2012 www.seeaction.energy.gov 8-19 TABLE 8.2: Energy Efficiency M&V Plan Contents: General Components . • M&V PLAN COMPONENTS Project goals and objectives Project Description Site characteristics and data available(e.g.,whole building and end-use metering) Measure descriptions that include how savings will be achieved Estimated savings by measure Project Savings and Costs Estimated M&V cost by measure Project installation schedule Schedules M&V activities Raw and compiled data formats Reporting M&V report contents Reporting interval Confidence and precision of results M&V Approach M&V option(s)used Person(s)responsible for M&V activities 8-20 www.seeaction.energy.gov December 2012 TABLE 8.3: Energy Efficiency Project-Specific and Measure-Specific M&V Plan Contents M&V PLAN COMPONENTS OMMMMMMI Data requirements kW,operating hours,temperature Basis of stipulated values Lighting operating hours equal 4,000/year based on metered XYZ building Analysis Method kWh savings(year)=[(kW/Flxture(baseline)x Quantitylbaseline)] Savings calculated equations —(kW/Fixture(post)x Quantity(post)]x Operating Hours per yea r(post) Regression expressions Three parameter change-point cooling model Computer simulation models DOE-2 simulation model Metering protocols ASHRAE Guideline 14-2002 pump multiple point test throughout short-term monitoring Equipment ABC Watt Hour Meter Equipment calibration protocols National Institute of Science and Technology protocols Metering and Monitoring Metered data Flow rate,RMS power Sample size 25 lighting circuits out of 350 Sampling accuracy 90%confidence/10%precision Metering duration and interval 2 weeks/15-minute data Performance factors Boiler efficiency Operating factors Load,operating hours Baseline Determination Existing service quality Indoor temperature set points Minimum performance standards State energy code Party responsible for developing Smith Engineers,hired by sponsor adjustments Savings Adjustments Savings adjustment approach Baseline adjusted for reported period weather and building occupancy levels December 2012 www.seeaction.energy.gov 8-21 TABLE 8.4: Evaluation Planning Checklist CHECKLIST FOR GROSS SAVINGS DETERMINATION SAVINGS TO BE REPORTED Energy savings(annual,seasonal,monthly,hourly,other) Demand savings(peak,coincident,average,other) SELECTED GROSS ENERGY SAVINGS CALCULATION APPROACH Measurement and verification approach Deemed savings approach Large-scale billing analysis approach Quality assurance approach L MEASUREMENT AND VERIFICATION APPROACH IPMVP Option A,B,C,or D DEEMED SAVINGS APPROACH Source of deemed savings identified and verified LARGE-SCALE BILLING ANALYSIS APPROACH Randomized controlled trials Quasi-experimental method SAMPLE SIZE CRITERIA SELECTED LIFETIME ENERGY AND DEMAND SAVINGS FACTORS TO BE EVALUATED Degradation LRebound Other 8-22 www.seeaction.energy.gov December 2012 . CHECKLIST FOR NET SAVINGS DETERMINATION NET SAVINGS FACTORS TO BE EVALUATED Free riders Spillover effects Other market effects Other(s) NET SAVINGS CALCULATION APPROACH SELECTED Self-reporting surveys Enhanced self-reporting surveys Large-scale consumption data analysis approaches(randomized controlled trial methods and quasi-experimental methods) Stipulated net-to-gross ratio Cross-sectional studies Top-down evaluations(or macro-economic models) CHECKLIST FOR AVOIDED EMISSIONSCALCULATIONS ELECTRICITY EFFICIENCY SAVINGS—GRID-CONNECTED LLoad emission rates Regional marginal baseload emission rates(using capacity factors or equivalent) Regional historical hourly emission rates Energy scenario modeling NATURAL GAS, FUEL OIL,AND NON-GRID-CONNECTED ELECTRIC GENERATING UNITS Default emission factor Source testing December 2012 www.seeaction.energy.gov 8-23 EVALUATIONGENERIC • • OVERALL GOALS L Does the evaluation address the key policy,regulatory,and oversight needs for evaluation information? Will the program's success in meeting energy,demand,and emissions goals be quantifiably evaluated in the same manner as they are defined for the program? Does the evaluation plan represent a reasonable approach to addressing the information needs? Are there missing opportunities associated with the evaluation approach that should be added or considered? Are any additional non-energy benefits being evaluated? Does the impact evaluation provide the data needed to inform other evaluations that may be performed, particularly cost-effectiveness analyses? Has a balance been reached between evaluation costs,uncertainty of results,and value of evaluation results? UNCERTAINTY OF EVALUATION RESULTS Can the confidence and precision of the evaluation results be quantified?If so,how? Are there key threats to the validity of the conclusions?Are they being minimized given budget constraints and study trade-offs?Will they be documented and analyzed? Is the evaluation capable of providing reliable conclusions on energy and other impacts? BUDGET,TIMING,AND RESOURCES Does the evaluation take advantage of previous evaluations and/or concurrent ones for other programs? Does the cost of the study match the methods and approaches planned? LDo the scheduled start and end times of the evaluation match the need for adequate Are adequate human resources identified? Does the evaluation rely on data and project access that are reasonably available? REPORTING Are the time frames and scopes of evaluation reported defined? Do the data collection,analysis,and quality control match the reporting needs? Are the persistence of savings and avoided emissions being evaluated? LHave measurements and impacts(emissions)boundaries been properly set? 8-24 www.seeaction.energy.gov December 2012 EVALUATIONGENERIC • . • SAMPLING AND ACCURACY L Is the sampling plan representative of the population served? Is the sampling plan able to support the evaluation policy objectives? LAre there threats to the validity of the evaluation results addressed in the sample design? December 2012 www.seeaction.energy.gov 8-25 Chapter 8: Notes 96 See discussion of theory-based evaluation in Section 8.3.1. 97 Cross-border trading means that energy savings obtained in one jurisdiction cannot be used to meet energy efficiency requirements in another jurisdiction.Should a national energy efficiency resource portfolio be established at some point,then consistent national EM&V standards and deemed savings values/calculations may become required. For more information on EM&V requirements that may be associated with a national energy efficiency or clean-resource standard,see State and Local Energy Efficiency Action Network.(April 2011).National Energy Efficiency Evaluation, Measurement, and Verification(EM&V)Standard:Scoping Study of Issues and Implementation Requirements. Prepared by Schiller,S.R.;Goldman, C.A.;Galawish, E.; LBNL Environmental Energy Technologies Division. wwwl.eere.energy.gov/seeaction/pdfs/emvstandard_scopi ngstudy.pdf. 98 Kushler, M.; Nowak,S.;Witte, P.(February 2012).A National Survey of State Policies and Practices for the Evaluation of Ratepayer- Funded Energy Efficiency Programs.American Council for an Energy- Efficient Economy(ACEEE). Report Number U122.www.aceee.org/ resea rch-report/u 122. 99 See the Efficiency Valuation Organization(EVO):www.evo-world.org. 100 A very common interactive factor is the relationship between(1) reductions in heat generated by retrofitted lighting systems in a building as they consume less electricity and(2)the possible requirements for increased space heating and/or reduced space cooling. 101 Lazar,J.; Baldwin,X.(August 2011).Valuing the Contribution of Energy Efficiency to Avoided Marginal Line Losses and Reserve Requirements. Regulatory Assistance Project(RAP).www.raponline.org/ document/down load/id/4537. 102 As of now,the baseline study contractor is separate from statewide contractors hired by the PSC for other activities listed in this table. 103 May be done in conjunction with Northeast Energy Efficiency Partnerships(NEEP)effort. 8-26 www.seeaction.energy.gov December 2012 Appendix A: Glossary This glossary defines and explains terms used in this guide and in the evaluation,measurement,and verification(EM&V)of energy efficiency programs.Included are terms commonly used in the design Savings,or more accurately stated"savings estimates,"from and implementation of EM&V,terms often found in EM&V reports, energy-efficiency measures,projects,programs,and portfolios and terms associated with efficiency measures,programs,and are reported at various times in the lifecycle of the efficiency program strategies.Some of the definitions in this glossary reflect activity and with varying degrees of certainty.The two major usage in the specific context of energy efficiency program administra- points when they are reported are prior to and after the tors and implementers with an oversight(regulatory)body;thus,they implementation of the activity.Different jurisdictions have may not be applicable to non-regulatory contexts such as in commer- different names for savings reports,what they contain,and cial agreements between energy services companies and their clients. whether and what adjustments or evaluation activities take place between pre-implementation and post-implementation. References used for preparation of this glossary are given at the For example,prior to implementation,savings can be called end of this appendix.Glossaries tend to be living documents. For Ex Ante Savings or Projected Savings.After implementation, example,the well-regarded "EM&V Forum Glossary of Terms,"a there are even more variations in names and content,with project of the Regional Evaluation, Measurement and Verification titles including ex post savings,adjusted savings,evaluated Forum is regularly updated as a result of reviews by forum members savings,and tracking estimates.Examples of differences in and outside experts.104 these"after implementation"reports can include whether the results have been corrected for data errors,whether they have been verified(e.g.,confirmed actual number of installations Accuracy:A concept that refers to the relationship between the true versus planning projections),and/or whether a third-party value of a variable and an estimate of the value.The term can also evaluator has generated or reviewed the reported values. be used in reference to a model or a set of measured data,or to Savings can also be indicated as first-year,annual demand describe a measuring instrument's capability. or energy savings,and/or lifetime energy or demand savings values.They also can be indicated as gross savings and/or net Additionality:A criterion applied to,for example,greenhouse gas savings values.In this glossary,several savings reporting (GHG)projects,stipulating that project-based GHG reductions classifications are defined.Ideally,these terms and meth- should only be quantified if the project activity"would not have ods should be applied;however,it is critical that whenever happened anyway;"in other words,that the project activity(or the savings are reported,the basis for the values indicated same technologies or practices it employs)would not have been be made clear,rather than just relying on these common implemented in its baseline scenario and/or that project activity classifications,since their definition(and titles)—as well emissions are lower than baseline emissions. as how values are determined for each classification—can Adjusted Savings:See claimed savings and evaluated savings. vary from jurisdiction to jurisdiction. Administrator:An entity selected by a regulatory or other government organization to contract for and administer an energy efficiency portfolio within a specific geographic region and/or sufficient allowances to cover their pollutant emissions during that period.Allowances are often confused with credits earned in market.Typical administrators are utilities selected by a public the context of project-based or offset programs,in which sources service commission or a nonprofit or state government agency,as trade with other facilities to attain compliance with a conventional determined by legislation. regulatory requirement.Cap-and-trade program basics are discussed The basic tradable commodity within a pollutant at www.epa.gov/capandtrade/index.html. emissions trading systems.Allowances grant their holder the right to Annual Demand Savings:The maximum reduction in electric or gas emit a specific quantity of pollution once(e.g.,one ton of COz(eq)). The total quantity of allowances issued by regulators dictates the demand associated with energy efficiency activities in a given year within a defined boundary.More specific definitions of demand and total quantity of emissions possible under the system.At the end of each compliance period,each regulated entity must surrender thus demand savings occur within specific regulatory jurisdictions or electric control areas. December 2012 www.seeaction.energy.gov A-1 Annual Energy Savings:The reduction in electricity use(kilowatt- Baseline Data:The baseline conditions of the facilities,market hours)or in fossil fuel use(in thermal units)associated with energy segment,generating equipment,or other area of focus of the subject efficiency activities in a given year within a defined boundary. efficiency activity. ASHRAE Guideline 14:American Society of Heating, Refrigerating Baseline Efficiency:The energy use of the baseline equipment, and Air-Conditioning Engineers(ASHRAE)Guideline 14,2002 process,or standard that is being replaced by a more efficient Measurement of Energy and Demand Savings(www.ashrae.org). approach to providing the same energy service(a subset of baseline). It is used to determine the energy savings obtained by the The boundary within which all the primary more efficient approach. effects and significant secondary effects associated with a project are evaluated. Baseline Period:The period of time selected as representative of facility operations before the energy efficiency activity takes place. Attribution:Ascribing or establishing a causal relationship between action(s)taken by an entity and an outcome. For efficiency program Baseline Scenario:A hypothetical description of what would have evaluation,this is associated with the difference between net and most likely occurred without implementation of the subject project gross savings. For example,an impact evaluation indicates that 30% or program. of the gross energy savings associated with a ceiling fan incentive program could be attributed to the ENERGY STAR'labeling program Behavior Programs:Energy efficiency programs that use strategies and not the incentive program. intended to affect consumer energy use behaviors in order to achieve energy and/or energy demand savings. Programs typically Avoided Costs:The forecasted economic benefits of energy savings. include outreach,education,competition,rewards, benchmarking, In the context of energy efficiency,these are the costs that are and/or feedback elements.Such programs may rely on changes avoided by the implementation of an energy efficiency activity.Such to consumers' habitual behaviors(e.g.,turning off lights)or"one- costs are used in benefit-cost analyses of energy efficiency activities. time"behaviors(e.g.,changing thermostat settings). In addition, Because efficiency activity reduces the need for electric generation, these programs may target purchasing behaviors(e.g.,purchases these costs include those associated with the cost of electric of energy-efficient products or services),often in combination with generation,transmission,distribution,and reliability.Typically,costs other programs(e.g., rebate programs or direct install programs). associated with avoided energy and capacity are calculated.Other These programs are also distinguished by normally being evaluated costs avoided by the efficiency activity can also be included,among using large-scale data analysis approaches involving experimental them the value of avoided emissions not already embedded in the or quasi-experimental methods,versus deemed savings or generation cost,impact of the demand reduction on the overall measurement and verification approaches. market price for electricity,and avoided fuel or water. For natural gas efficiency programs,avoided costs can include components of the Benchmarking:A process that compares the energy,emissions, production,transportation,storage,and service that are variable to and other resource-related conditions of a facility against industry the amount of natural gas delivered to customers. best practices or other benchmarks such as average per square foot energy consumption of similar buildings in the same city. Baseline:Conditions,including energy consumption and demand, which would have occurred without implementation of the subject Benefit-Cost Ratio:The mathematical relationship between the energy efficiency activity.Baseline conditions are sometimes referred benefits and costs associated with the implementation of energy to as"business-as-usual"conditions and are used to calculate efficiency measures,programs,practices,or emission reductions. project-and program-related savings. Baselines can also include The benefits and costs are typically expressed in dollars. definition of non-energy metrics that are being evaluated,such as air emissions and jobs. Benefit-Cost Test:The methodology used to compare the benefits of an investment with the costs;also called cost-effectiveness test. Baseline Adjustments:For measurement and verification analyses, Five key benefit-cost tests have,with minor updates,been used factors that modify baseline energy or demand values to account for for more than 20 years as the principal approaches for energy independent variable values(conditions)in the reporting period. efficiency program evaluation.These five cost-effectiveness tests are A-2 www.seeaction.energy.gov December 2012 the participant cost test,the program administrator cost test the Calibration: In economic,planning,or engineering modeling,the ratepayer impact measure test the total resource cost test, and the process of adjusting the components of the model to reflect reality societal cost test. as closely as possible,in order to prepare for the model's use in future applications.The term also applies to the process whereby Bias:The extent to which a measurement or a sampling or analytic metering and measurement equipment is periodically adjusted to method systematically underestimates or overestimates a value. maintain industry measurement standards. Some examples of types of bias include engineering model bias; meter bias;sensor placement bias;inadequate or inappropriate California Measurement Advisory Council(CALMAQ An informal estimates of what would have happened absent a program or committee comprised of representatives of the California utilities, measure installation;a sample that is unrepresentative of a state agencies,and other interested parties.CALMAC provides population;and selection of other variables in an analysis that are a forum for the development,implementation,presentation, too correlated with the savings variable(or each other)in explaining discussion,and review of regional and statewide market assessment the dependent variable(such as consumption). and evaluation studies for California energy efficiency programs conducted by member organizations(www.calmac.org). Billing Analysis:A term used to define either(1)a specific measurement and verification(M&V)approach used to estimate Claimed Savings:Values reported by an implementer or project savings or(2)any analytic methodology used to determine administrator,using their own staff and/or an evaluation consulting project or program energy savings based on the use of the energy firm,after the subject energy efficiency activities have been consumption data contained in consumer billing data.It compares completed;also called tracking estimates,reported savings,or in billing data from program participant(s)over a period of time before some cases,ex post savings(although ex post usually applies to the energy-efficient measures are installed at customer site(s)to evaluated savings[see evaluated savings]). billing data for a comparable period of time afterward.If used to describe an M&V approach,it is equivalent to IPMVP Option C,Whole As with projected savings estimates,these values may use results Building Analysis.If used to describe an evaluation approach,it is of prior evaluations and/or values in a technical reference manual. comparable to the large-scale data analysis approach. However,they may be adjusted from projected savings estimates by correcting for any known data errors and actual installation rates and Billing Data:Data obtained from the electric or gas meter that may also be adjusted with revised values for factors such as per-unit is used to bill the customer for energy used in a particular billing savings values,operating hours,and savings persistence rates. period.In an evaluation context,billing data also refers to the Claimed savings can be indicated as first-year,annual,and/or lifetime customer billing records over time.Those records are used to energy or demand savings values,and can indicated as gross savings conduct analyses of energy use before and after implementation and/or net savings values. of energy efficiency measures. Co-Benefits:The impacts of an energy efficiency program other than Building Commissioning:See commissioning. the direct purpose(i.e.,energy and demand savings)for which it was designed.See non-energy benefits. Building Energy Simulation Model-Computer models based on physical engineering principals and/or standards used to estimate Coefficient of Variation(CV The mean(average)of a sample, energy use and/or savings.These models usually incorporate divided by its standard error. site-specific data on customers and physical systems,such as square footage,weather,surface orientations,elevations,space Coincident Demand:The demand of a device,circuit,or building volumes,construction materials,equipment use,lighting,and that occurs at the same time as the peak demand of a utility's system building occupancy.Building simulation models can usually account load or at the same time as some other peak of interest,such as for interactive effects between end uses(e.g.,lighting and HVAC), building or facility peak demand.The peak of interest should be part-load efficiencies,and changes in external and internal heat specified(e.g.,"demand coincident with the utility system peak"). gains/losses.Examples of building simulation models include DOE-2, The following are examples of peak demand: EnergyPlus,and Carrier HAP. • Demand coincident with utility system peak load December 2012 www.seeaction.energy.gov A-3 • Demand coincident with independent system operator/regional variable are associated with high values of another variable for the transmission organization summer or winter peak or according same participant. For example,facility size and energy consumption to performance hours defined by wholesale capacity markets usually have a high positive correlation. • Demand coincident with high electricity demand days. Cost-Benefit and Cost-Effectiveness Analysis:Analysis that compares Commissioning:Often abbreviated as"Cx,"a systematic quality the benefits associated with a program or measure's outputs or assurance process associated with new construction that spans outcomes with the costs(resources expended)to produce them. the entire design and construction process,helping ensure that a Cost-benefit analysis is typically conducted to determine the new building's performance meets owner expectations;sometimes relationship of the program's benefits and costs,as a ratio.Cost- referred to as building commissioning.Commissioning ensures effectiveness analysis is generally undertaken to compare one that the new building operates as the owner intended and that program or program approach to other approaches,or options for building staff are prepared to operate and maintain its systems and the use of funds,to determine the relationship among the options. equipment. The terms are often interchanged in evaluation discussions.See benefit-cost test. Common Practice:The predominant technology(ies)implemented or practice(s)undertaken in a particular region or sector.Common Cost Effectiveness:An indicator of the relative performance or practices can be used to define a baseline. economic attractiveness of any energy efficiency investment or practice. In the energy efficiency field,the present value of the Comparison Group:See control group. estimated benefits produced by an energy efficiency program is compared to the estimated total costs to determine if the proposed Conditional Savings Analysis(CSA):A type of analysis in which investment or measure is desirable from a variety of perspectives change in consumption is modeled using regression analysis against (e.g.,whether the estimated benefits exceed the estimated costs presence or absence of energy efficiency measures. from a societal perspective).See benefit-cost ratio. Confidence:An indication of how close,expressed as a probability, Cumulative Energy Savings:The summation of energy savings(e.g., the true value of the quantity in question is within a specified megawatt-hours,therms)from multiple projects or programs over a distance to the estimate of the value.Confidence is the likelihood specified number of years,taking into account the time of measure that the evaluation has determined the true value of a variable installation in the first year,annual energy savings for subsequent within a certain estimated range. For example,a program that years,and the average life of the installed measures. produces an estimate of 2%energy savings,with a 95%confidence interval of(1%,3%)means that there is a 95%probability that the Custom Program:An energy efficiency program intended to provide true program energy savings is between 1%and 3%.A smaller efficiency solutions to unique situations not amenable to common confidence interval implies that the estimate is more precise(e.g., or prescriptive solutions. Each custom project is examined for its a 2%energy savings estimate with a confidence interval of[1.5%, individual characteristics,savings opportunities,efficiency solutions, 2.5%]is more precise than a 2%energy savings estimate with a and often,customer incentives. confidence interval of[1%,3%]).See precision. Database for Energy-Efficient Resources(DEER):A California Control Group:A group of consumers who did not participate in the database designed to provide publicly available estimates of energy evaluated program during the program year and who share as many and peak demand savings values,measure costs,and effective useful characteristics as possible with the treatment(participant)group; life(www.deeresources.com). also called comparison group.The comparison group is used to isolate program effects from other factors that affect energy use. Deemed Savings Calculation:An agreed-to(stipulated)engineering algorithm(s)used to calculate the energy and/or demand savings Cooling Degree Days:The cumulative number of degrees in a month associated with an installed energy efficiency measure.These or year by which the mean temperature is above a set temperature, calculations are developed from common practice that is widely usually 18.3°C/65°F.See degree days. considered acceptable for the subject measure and its specific application.It may include stipulated assumptions for one or Correlation For a set of observations,such as for participants in an energy efficiency program,the extent to which high values for one A-4 www.seeaction.energy.gov December 2012 more parameters in the algorithm,but typically it requires users Demand-Side Management(DSM):Strategies used to manage to input data associated with the actual installed measure into the v energy demand,including energy efficiency,load management, algorithm(s). fuel substitution,and load building. Deemed Savings Value:An estimate of energy or demand savings Dependent Variable:Term used in regression analysis or other for a single unit of an installed energy efficiency measure that(1)has analyses seeking to explain the relationship among variables to been developed from data sources and analytical methods that are quantify the variable that is being explained by the other widely considered acceptable for the measure and purpose,and(2) (independent)variables. is applicable to the situation being evaluated. Individual parameters or calculation methods can also be deemed;also called stipulated Direct Emissions:Emissions from sources within an entity's organiza- savings value. tional boundaries that are owned or controlled by the entity,includ- ing stationary combustion emissions, mobile combustion emissions, Degree Days:For any individual day,an indication of how far that process emissions,and fugitive emissions. Direct emissions are the day's average temperature departed from a fixed temperature, source of avoided emissions for thermal energy efficiency measures usually 18.3°C/65°F. Heating degree days,which measure heating (e.g.,avoided emissions from burning natural gas in a water heater). energy demand,quantify how far the average temperature fell below 65°F.Similarly,cooling degree days,which measure cooling energy Direct Install Program:An energy efficiency program design strategy demand,quantify how far the temperature averaged above 65°F. In involving the direct installation of measures in customer premises both cases,smaller values represent less energy demand;however, by a contractor sponsored by the program.Such programs generally values below 0 are set equal to 0,because energy demand cannot involve one-for-one replacement of existing equipment with more be negative. Furthermore,because energy demand is cumulative, efficient equipment,and may include a customer rebate. degree day totals for periods exceeding one day are simply the sum of each individual day's degree days total.Degree days are used in Discount Rate:An interest rate applied to a stream of future costs calculations of heating and cooling loads and in evaluation regression and/or monetized benefits to convert those values to a common analyses to adjust for differences in heating and cooling requirements period,typically the current or near-term year,to measure and between baseline and project scenarios. reflect the time value of money. It is used in benefit-cost analysis to determine the economic merits of proceeding with the proposed Demand:The time rate of energy flow. It is the requirement for project,and in cost-effectiveness analysis to compare the value energy consumption of energy source(s)by an energy using system of projects.The discount rate for any analysis is either a nominal at a given instant or averaged over any designated interval of time. discount rate or a real discount rate.A nominal discount rate is used Demand usually refers to the amount of electric energy used by in analytic situations when the values are in then-current or nominal a customer or piece of equipment at a specific time,expressed in dollars(reflecting anticipated inflation rates).A real discount rate kilowatts(kW equals kWh/h)but can also refer to natural gas use at a is used when the future values are in constant dollars,and can be point in time,usually as Btu/hr,kBtu/hr,therms/day,or cubic feet per approximated by subtracting expected inflation from a nominal day(ccf/day). discount rate. Demand Response(DR) The reduction of consumer energy use at Diversity:That characteristic of a variety of electric loads whereby times of peak use in order to help system reliability, reflect market individual maximum demands of each load usually occur at conditions and pricing,or support infrastructure optimization or different times. deferral of additional infrastructure.Demand response programs may include contractually obligated or voluntary curtailment,direct load Diversity Factor:The ratio of the sum of the demands of a group of control,and pricing strategies. users to their coincident maximum demand during a specified period of time(e.g.,summer or winter). Demand Savings:The reduction in electric or gas demand from the baseline to the demand associated with the higher-efficiency Education Programs:Programs primarily intended to educate equipment or installation.This term is usually applied to billing customers about energy efficient technologies or behaviors or demand to calculate cost savings or peak demand for equipment provide information about programs that offer energy efficiency sizing purposes. or load-reduction information or services.These programs may provide indirect energy or demand savings. December 2012 www.seeaction.energy.gov A-5 Effective Useful Life(EUL):An estimate of the duration of savings systems,or operations that reduces the amount of energy that from a measure. EUL is estimated through various means,including would otherwise have been used to deliver an equivalent or median number of years that the energy efficiency measures improved level of end-use service. installed under a program are still in place and operable.Also, EUL is sometimes defined as the date at which 50%of Installed units are Energy Savings:Reduction in electricity use in kilowatt-hours or in still in place and operational.See measure life. fossil fuel use in thermal unit(s). Efficiency:The ratio of the useful energy delivered by a dynamic ENERGY STAR':A joint program of the U.S.Environmental Protection system(e.g.,a machine,engine,or motor)to the energy supplied Agency and the U.S.Department of Energy designed to reduce energy to it over the same period or cycle of operation.The ratio is usually use and the impact on the environment.The ENERGY STAR label is determined under specific test conditions. awarded to products that meet applicable energy efficiency guide- lines and to homes and commercial buildings that meet specified Emission Factor:A representative value that relates the quantity energy efficiency standards.The program provides a range of energy of a pollutant released to the atmosphere with an activity associ- management tools,primarily computer-based,for businesses. ated with the release of that pollutant.These factors are usually expressed as the weight of pollutant divided by a unit weight, Engineering Methods:The use of standard formulas or models volume,distance,or duration of the activity emitting the pollutant based on those formulas,typically accepted by ASHRAE,as the basis (e.g.,pounds of a pollutant per million Btu of heat input or pounds for calculating energy use. of emissions per MWh of electricity produced).Such factors facilitate estimation of emissions from various sources of air pollution.They Engineering Model: Engineering equations used to calculate energy are based on available data of acceptable quality,and are generally use and savings.These models are usually based on a quantitative assumed to be representative of averages for that pollutant. description of physical processes that transform delivered energy into useful work,such as heat, lighting,or motor drive.In practice, End Use:General categories of energy efficiency measures these models may be reduced to simple equations in spreadsheets reflecting the type of services provided(e.g.,lighting,HVAC, that calculate energy use or savings as a function of measurable motors,and refrigeration). attributes of customers,facilities,or equipment(e.g.,lighting use= watts x hours of use). End-Use Metering:The direct measuring of energy consumption or demand by specific end-use equipment,typically as part of load Equipment Life:The number of years that a measure is installed and research studies or to measure the impacts of demand-side operates until failure. management programs. Error:The deviation of measurements from the true value of the Energy Conservation:Term used to reflect doing with less of a variable being observed;also called measurement error. service in order to save energy.The term is often unintentionally used instead of energy efficiency. Evaluated Savings:Savings estimates reported by an independent, third-party evaluator after the subject energy efficiency activities Energy Efficiency:The use of less energy to provide the same or an have been implemented and an impact evaluation has been improved level of service to the energy consumer;or,the use of less completed;also called ex post,or more appropriately,ex post energy to perform the same function. evaluated savings. Energy Efficiency Activity:Any of a wide range of actions that are The designation of"independent"and"third party"is determined by anticipated to result in the more efficient use of energy.Energy those entities involved in the use of the evaluations and may include efficiency measures are a subset of energy efficiency activities,as evaluators retained,for example,by the administrator or energy efficiency activities include non-technology specific actions a regulator.These values may rely on claimed savings for factors such as education. such as installation rates and a technical reference manual for values such as per-unit savings values and operating hours.These Energy Efficiency Measure:At an end-use energy consumer facility, saving estimates may also include adjustments to claimed savings an installed piece of equipment or system;a strategy intended to or projected savings for data errors,per-unit savings values, affect consumer energy use behaviors;or modification of equipment, A-6 www.seeaction.energy.gov December 2012 operating hours,installation rates,savings persistence rates,or which the participant's activity would have partially replicated the other considerations. Evaluated savings can be indicated as first-year, program measure;or(3)deferred,in which the participant's activity annual,and/or lifetime energy or demand savings values.They also would have partially or completely replicated the program measure, can be indicated as gross savings and/or net savings values. but at a future time beyond the program's time frame. Evaluation:The conduct of any of a wide range of assessment Fuel Switching: Using an alternative fuel(usually of lower carbon studies and other activities aimed at determining the effects of a intensity)to produce required energy. program and understanding or documenting program performance, program or program-related markets and market operations, The change in energy consumption and/or program-induced changes in energy efficiency markets,levels of demand that results from energy efficiency programs,codes and demand or energy savings,or program cost-effectiveness. Market standards,and naturally occurring adoption,which have a long- assessment,monitoring and evaluation,and measurement and lasting savings effect.Gross market savings generally do not include verification are aspects of evaluation. temporary reductions in energy use from changes in weather, income,energy prices,and other structural economic changes such Evaluator:A person or entity that conducts evaluations. as in industry composition. Ex Ante Savings:See projected savings. Gross Savings:The change in energy consumption and/or demand that results directly from program-related actions taken by Experimental Design:A method of estimating the impact of a participants in an energy efficiency program,regardless of why they program or other event(e.g.,a medication,a procedure),in participated. which outcomes between at least two randomly assigned groups are compared. Heating Degree Days:The cumulative number of degrees in a month or year by which the mean temperature falls below a fixed Ex Post Savings:See claimed savings and evaluated savings. temperature,usually 18.3°C/65°F.See degree days. Ex Post Evaluated Savings:See evaluated savings. Home Energy Rating System(HERS):An indexing system,associated with ENERGY STAR®,used in residential new construction to rate the External Validity:The condition in which an impact estimate that is pre-and post-construction of new homes to highlight and indicate internally valid for a given program population and time frame can the degree of energy efficiency embedded in the construction.The be generalized and applied to new situations(e.g.,new populations, HERS Index is a scoring system established by the Residential Energy future years). Services Network(RESNET)in which a home built to the specifica- FEMP M&V Guidelines:U.S. Department of Energy Federal Energy tions of the HERS Reference Home(based on the 2006 International Management Program's 2008 M&V Guidelines:Measurement and Energy Conservation Code)scores a HERS Index of 100,while a net Verification for Federal Energy Projects. zero energy home scores a HERS Index of 0.The lower a home's HERS Index,the more energy efficient it is in comparison to the Free Driver,Non-Participant:A program non-participant who has HERS Reference Home. Each 1-point decrease in the HERS Index adopted particular energy efficiency measure(s)or practice(s)as a corresponds to a 1%reduction in energy consumption compared result of the evaluated program.See spillover. to the HERS Reference Home. Free Driver,Participant:A program participant who has adopted HVAC:Heating,ventilation,and air conditioning. additional or incremental energy efficiency measure(s)or practice(s) as a result of the evaluated program, but which were not directly Impact Evaluation:An evaluation of the program-specific,directly or induced by the program.See spillover. indirectly induced changes(e.g.,changes in energy and/or demand use)associated with an energy efficiency program. Free Rider:A program participant who would have implemented the program's measure(s)or practice(s)in the absence of the program. An entity selected and contracted with or qualified Free riders can be(1)total,in which the participant's activity would by a program administrator to provide products and services to have completely replicated the program measure;(2)partial,in consumers either directly or indirectly. December 2012 www.seeaction.energy.gov A-7 Incentive:A financial strategy intended to encourage a change in number of measures installed by the number of measures incented behavior related to energy use. Incentives can take various forms. by an energy efficiency program in a defined period of time. Customer incentives are commonly used in energy efficiency pro- grams as rebates for individual measures or as buy-downs in more Interactive Effects:The influence of one technology's application on custom-oriented projects. Performance or shareholder incentives the energy required to operate another application.An example is are monies that are established in a planning period to encourage the reduced heat in a facility as a result of replacing incandescent program administrators to attain specified levels of savings during lights with CFLs,and the resulting need to increase space heating the program year. from another source,usually oil-or gas-fired.With respect to IPMVP Options A and B,interactive effects in energy use or demand occur- Incremental Annual Savings:The difference between the amount ring beyond the measurement boundary of the M&V analysis. of energy savings acquired or planned to be acquired as a result of energy efficiency activities in one year,and the amount of energy Internal Validity:Refers to how well an evaluation was conducted savings acquired or planned to be acquired as a result of the energy (e.g.,design,how variables were measured,what was/wasn't efficiency activities in the prior year. measured)and how confidently one can conclude that the observed effect(s)were produced solely by the independent variable and not Incremental Cost:The difference between the cost of existing or extraneous ones.For impact evaluations,this is related to whether baseline equipment or service and the cost of alternative energy- the savings impacts are valid for the specific program being evalu- efficient equipment or service. ated,the given program participant population,and the given time frame of the evaluation.This is often compared to external validity. Independent Variables:The explanatory factors(e.g.,weather or occupancy)in a regression model that are assumed to affect the International Performance Measurement and Verification Protocol variable under study(e.g.,energy use). (IPMVP):A guidance document with a framework and definitions describing the four M&V approaches;a product of the Efficiency Indirect Emissions:Emissions that are a consequence of activities Valuation Organization(www.evo-world.org). that take place within the organizational boundaries of an entity,but occur at sources owned or controlled by another entity.For example, In its broadest terms,the concept that the effect of an emissions of electricity used by a manufacturing entity that occur activity or outcome expected to occur and remain within a defined at a power plant represent the manufacturer's indirect emissions. boundary flows outside the boundary,leading to unintended results. Indirect emissions are typically the source of avoided emissions for In efficiency programs,an example of leakage is when a measure is electric energy efficiency measures. incented by a program(with the associated costs and assumed savings)but is installed outside of the program's jurisdiction.In Indirect Energy(Demand)Savings(Indirect Program Energy the context of air regulation,such as a cap-and-trade program,an Savings):The use of the words"indirect savings"or"indirect example of leakage is a shift of electricity generation from sources program savings"refers to programs that are typically information, subject to the cap-and-trade program to higher-emitting sources education,marketing,or outreach programs in which the program's not subject to the program.Sometimes used interchangeably actions are expected to result in energy savings achieved through the with secondary effects,although leakage is a more"global"issue, actions of the customers exposed to the program's efforts,without whereas secondary,interactive effects tend to be considered direct enrollment in an program that has energy-savings goals. within the facility where a project takes place. Inspections.Site visits to facilities treated under an energy efficiency Levelized Cost:The result of a computational approach used to program that document the existence,characteristics,and operation compare the cost of different projects or technologies.The stream of baseline or project equipment and systems,as well as factors of each project's net costs is discounted to a single year using a that affect energy use.Inspections may or may not include review of discount rate(creating a net present value)and divided by the commissioning or retro-commissioning documentation. project's expected lifetime output(megawatt-hours or therms). Installation Rate:The percentage of measures that are incented by Lifetime Cost Per Kilowatt-hour or Therm:The cost associated an energy efficiency program that are actually installed in a defined with a piece of equipment(supply-side or demand-side),energy period of time.The installation rate is calculated by dividing the efficiency program,or total portfolio during its expected life in A-8 www.seeaction.energy.gov December 2012 relation to(divided by)the electricity or gas that it produces or saves flow is shown by arrows from one to the others in the program logic. over its lifetime.The annual costs are usually discounted back to a It can also be displayed as a table,with the linear relationship single year using an appropriate discount rate. presented by the rows in the table. Lifetime Demand Savings:The expected demand savings over the Main Meter:The meter that measures the energy used for the lifetime of an installed measure(s),project(s),or program(s).It may whole facility.There is at least one meter for each energy source and be calculated by multiplying the annual peak demand reduction possibly more than one per source for large facilities.Typically,utility associated with a subject measure(s)by the expected useful lifetime meters are used,but dataloggers may also be used as long as they of the measure(s). It may include consideration of technical degrada- isolate the load for the facility being studied.When more than one tion and possibly the rebound effect.Savings can be gross or net. meter per energy source exists for a facility,the main meter may be For electricity,it can be expressed in units of kilowatt-years. considered the accumulation of all the meters involved. Lifetime Energy Savings The expected energy savings over the Marginal Cost:The sum that has to be paid for the next increment of lifetime of an installed measure(s),project(s),or program(s). It may be product or service.The marginal cost of electricity is the price to be calculated by multiplying the annual energy use reduction associated paid for kilowatt-hours above and beyond those supplied by with a subject measure(s)by the expected useful lifetime of the sub- presently available generating capacity. ject measure(s).It may include consideration of technical degradation and possibly the rebound effect.Savings can be gross or net. The commercial activity(e.g.,manufacturing,distributing, buying,and selling)associated with products and services that affect Load Factor:A percentage indicating the difference between the energy use. amount of electricity or natural gas a consumer used during a given time span and the amount that would have been used if the use had Market Assessmeni An analysis that provides an assessment of how stayed at the consumer's highest demand level during the whole and how well a specific market or market segment is functioning time.The term also means the percentage of capacity of an energy with respect to the definition of well-functioning markets or with facility,such as a power plant or gas pipeline,that is used in a given respect to other specific policy objectives.A market assessment period of time. It is also the ratio of the average load to the peak generally includes a characterization or description of the specific load during a specified time interval. market or market segments,including a description of the types and number of buyers and sellers in the market,the key actors that Load Management:Steps taken to reduce power demand at peak influence the market,the type and number of transactions that load times or to shift some of it to off-peak times. Load management occur on an annual basis,and the extent to which market participants may coincide with peak hours,peak days,or peak seasons. Load consider energy efficiency an important part of these transactions. management may be pursued by persuading consumers to modify This analysis may also include an assessment of whether a market behavior or by using equipment that regulates some electric has been sufficiently transformed to justify a reduction or elimina- consumption.This may lead to complete elimination of electric use tion of specific program interventions.Market assessment can be during the period of interest(load shedding)and/or to an increase in blended with strategic planning analysis to produce recommended electric demand in the off-peak hours as a result of shifting electric program designs or budgets.One particular kind of market assess- use to that period(load shifting). ment effort is a baseline study,or the characterization of a market before the commencement of a specific intervention in the market Load Shapes:Representations such as graphs,tables,and databases for the purpose of guiding the intervention and/or assessing its that show the time-of-use pattern of customer or equipment energy effectiveness later. use.These are typically shown over a 24-hour or whole-year(8,760 hours)period. Any characteristic of the market for an energy- related product,service,or practice that helps to explain the gap Logic Model:The graphical representation of a program theory between the actual level of investment in,or practice of,energy showing the connection among activities,their outputs,and subse- efficiency and an increased level that would appear to be cost quent short-term,intermediate,and long-term outcomes.Often,the beneficial to the consumer. logic model is displayed with these elements in boxes,and the causal December 2012 www.seeaction.energy.gov A-9 Market Effect:A change in the structure of a market or the behavior Market Sectors:General types of markets that a program may target of participants in a market that is reflective of an increase(or or in which a service offering may be placed. Market sectors include decrease)in the adoption of energy efficient products,services, categories such as agricultural,commercial,industrial,government, or practices and is causally related to market interventions(e.g., and institutional. programs). Examples of market effects include increased levels of awareness of energy efficient technologies among customers and Market Segments:A part of a market sector that can be grouped suppliers,increased availability of energy efficient technologies together as a result of a characteristic similar to the group.For through retail channels,reduced prices for energy efficient models, example,within the residential sector are market segments such as build out of energy efficient model lines,and—the end goal— renters,owners,multifamily,and single-family. increased market share for energy efficient goods,services,and design practices. Market Theory:A theoretical description of how a market operates relative to a specific program or set of programs designed to influ- Market Effect Evaluation:An evaluation of the change in the ence that market. Market theories typically include the identification structure or functioning of a market,or the behavior of participants of key market actors,information flows,and product flows through in a market,that results from one or more program efforts.Typically, the market, relative to a program designed to change the way the the resultant market or behavior change leads to an increase in the market operates.Market theories are typically grounded upon the adoption of energy efficient products,services,or practices. information provided from a market assessment but can also be based on other information. Market theories often describe how Market Event:The broader circumstances under which a customer a program intervention can take advantage of the structure and considers adopting an energy efficiency product,service,or practice. function of a market to transform the market.Market theories can Types of market events include,but are not necessarily limited to(1) also describe the key barriers and benefits associated with a market new construction,or the construction of a new building or facility; and describe how a program can exploit the benefits and overcome (2)renovation,or the updating of an existing building or facility;(3) the barriers. remodeling,or a change in an existing building;(4)replacement,or the replacement of equipment,either as a result of an emergency Market Transformation:A reduction in market barriers resulting such as equipment failure or as part of a broader planned event; from a market intervention,as evidenced by a set of market effects and(5)retrofit,or the early replacement of equipment or refitting that is likely to last after the intervention has been withdrawn, of a building or facility while equipment is still functioning,often as reduced,or changed. a result of an intervention into energy efficiency markets. Measure: [verb] Use of an instrument to assess a physical quantity Market Participants The individuals and organizations participating or use of a computer simulation to estimate a physical quantity. in transactions with one another within an energy efficiency market or markets,including customers and market actors. Measure:[noun]See energy efficiency measure. Market Penetration Rate:A measure of the diffusion of a technol- Measure Life:The length of time that a measure is expected to ogy,product,or practice in a defined market,as represented by the be functional;sometimes referred to as expected useful life. percentage of annual sales for a product or practice,the percentage Measure life is a function of equipment life and measure persistence. of the existing installed stock for a product or category of products, Equipment life is the number of years that a measure is installed and or the percentage of existing installed stock that uses a practice. will operate until failure.Measure persistence takes into account business turnover,early retirement of installed equipment,and Market Saturation:A percentage indicating the proportion of a other reasons measures might be removed or discontinued. specified end-user market that contains a particular product.An example would be the percentage of all households in a given Measurement and Verification(M&V):A subset of program impact geographical area that have a certain appliance.Studies conducted evaluation that is associated with the documentation of energy to obtain this information within the residential sector are referred savings at individual sites or projects using one or more methods to as residential appliance saturation studies(RA55). that can involve measurements,engineering calculations,statistical analyses,and/or computer simulation modeling. M&V approaches are defined in the IPMVR A-10 www.seeaction.energy.gov December 2012 Measurement Boundary:The boundary of the analysis for determin- Naturally Occurring Efficiency:The effects of energy-related deci- ing direct energy and/or demand savings. sions,by both program participants and non-participants,that would have been made in the absence of the program;alternatively,the Measure Penetration:The fraction of annual market sales captured expected average efficiency of one or more measures or systems in by a more efficient measure or system at a given point in time;also the absence of all publicly funded energy efficiency programs. It can called market penetration. For example,the market penetration of be part of a baseline determination. CFLs was 20%in the Florida market in 2005. Net Present Value(NPV):The value of a stream of cash flows Measure Persistence:The duration of an energy-consuming converted to a single sum in a specific year,usually the first year of measure,taking into account business turnover,early retirement the analysis. It can also be thought of as the equivalent worth of all of installed equipment,technical degradation factors,and other cash flows relative to a base point called the present. reasons measures might be removed or discontinued. The change in energy consumption and/or demand that Measure Retention Study:An assessment of(1)the length of time is attributable to a particular energy efficiency program.This change the measure(s)installed during the program year are maintained in in energy use and/or demand may include,implicitly or explicitly, operating condition,and(2)the extent to which there has been a consideration of factors such as free ridership,participant and non- significant reduction in the effectiveness of the measure(s). participant spillover,and induced market effects.These factors may be considered in how a baseline is defined(e.g.,common practice) Measure Saturation:The fraction of a total market(e.g.,consumers, and/or in adjustments to gross savings values. buildings)captured by a specific efficiency activity at a given point in time;also called market saturation. For example,in 2005,20%of the Net-to-Gross(NTG)Ratio:A factor representing net program commercial buildings that are more than 100,000 square feet in savings divided by gross program savings that is applied to gross Portland had high-efficiency chillers that exceeded minimum program impacts to convert them into net program load impacts. efficiency standards. The factor itself may be made up of a variety of factors that create differences between gross and net savings,commonly including free Metered Data:Data collected over time through a meter for a riders and spillover.Can be applied separately to either energy or specific energy using,end-use system(e.g.,lighting and HVAC), demand savings. or location(e.g.,floors of a building or a whole premise). Metered data may be collected over a variety of time intervals. Metered New Construction: Residential and nonresidential buildings that data usually refers to electricity or gas data. have been newly built or have added major additions. Metering:The collection of energy-consumption data over time Nominal:For dollars,"nominal"means the figure representing the through the use of meters.These meters may collect information actual number of dollars exchanged in each year,without accounting with respect to an end use,a circuit,a piece of equipment,or a for the effect of inflation on the value or purchasing power.For inter- whole building(or facility).Short-term metering generally refers to est or discount rates,"nominal"means that the rate includes the data collection for no more than a few weeks.End-use metering rate of inflation(the real rate plus the inflation rate approximately refers specifically to separate data collection for one or more end equals the nominal rate). uses in a facility,such as lighting,air conditioning,or refrigeration. Spot metering is an instantaneous measurement(rather than over Non-Energy Effects or Non-Energy Benefits(NEB):The identifi- time)to determine an energy consumption rate. able non-energy impacts associated with program implementation or participation;also referred to as non-energy impacts(NEI)or The collection of relevant measurement data over time co-benefits.Examples of NEBs include avoided emissions and other at a facility,including but not limited to energy consumption or environmental benefits,productivity improvements,jobs created, emissions data(e.g.,energy and water consumption,temperature, reduced program administrator debt and disconnects,and higher humidity,volume of emissions,hours of operation)for the purpose comfort and convenience level of the participant.The value is most of savings analysis or to evaluate equipment or system performance. often positive,but may also be negative(e.g.,the cost of additional maintenance associated with a sophisticated,energy-efficient control system). December 2012 www.seeaction.energy.gov A-11 Non-Participant:Any consumer who was eligible but did not programs),or(2)the set of all programs conducted by one participate in the subject efficiency program,in a given program year. organization,such as a utility(and which could include programs that cover multiple markets,technologies,etc.). Normalized Annual Consumption(NAC)Analysis:A regression- based method that analyzes monthly energy consumption data and Potential,Achievable:The amount of energy or demand savings adjusts the consumption data to eliminate annual or other periodic within a defined geographical area or population that can be fluctuations in an influencing factor(e.g.,weather on heating and achieved in response to specific energy efficiency program designs, cooling needs)based on a historical normal or average pattern of the delivery approaches,program funding,and measure incentive levels; influencing factor. sometimes referred to as market potential,program potential,or realistic potential. Offset: Program mechanism that allows an entity to neutralize a portion or all of its regulated,capped emissions contribution by Potential,Economic:Refers to the subset of the technical orchestrating or funding projects that are not subject to regulation potential that is economically cost-effective as compared to (i.e.,an emissions cap). conventional options. Panel Data Model:An estimation analysis model that contains many Potential Studies:Studies conducted to assess market baselines and data points over time rather than averaged,summed,or otherwise future savings that may be expected for different technologies and aggregated data. customer markets over a specified time horizon. Participant:A consumer that received a service offered through Potential,Technical:An estimate of energy savings based on the the subject efficiency program,in a given program year;also called assumption that all existing equipment or measures will be replaced with the most efficient equipment or measure that is both available program participant.The term"service"is used in this definition to suggest that the service can be a wide variety of inducements, and technically feasible over a defined time horizon,without regard including financial rebates,technical assistance,product installations, to cost or market acceptance. training,energy efficiency information or other services,items,or Precision:The indication of the closeness of agreement among conditions. Each evaluation plan should define"participant"as it repeated measurements of the same physical quantity. Precision applies to the specific evaluation. is a measure of how statistically confident evaluators can be that the estimated impact of a program is close to the true impact of a Participant Cost Test(PCT):A cost-effectiveness test that measures program.An estimate with a smaller confidence interval is said to the economic impact to the participant of adopting an energy be more precise. It is also used to represent the degree to which efficiency measure. an estimated result in social science(e.g.,energy savings)would be Peak Demand:The maximum level of hourly demand during a replicated with repeated studies. specified period.The peak periods most commonly identified are Prescriptive Program:An energy efficiency program focused annual and seasonal(summer and winter). on measures that are one-for-one replacements of the existing equipment and for which fixed customer incentives can be Peak Load:The highest electrical demand within a particular period developed based on the anticipated similar savings that will accrue of time.Daily electric peaks on weekdays typically occur in late from their installation. afternoon and early evening.Annual peaks typically occur on hot summer days. Primary Effects:Effects that the project or program are intended to achieve.For efficiency programs,this is primarily a reduction in Persistence:See savings persistence rate and measure persistence. energy use(and/or demand)per-unit of output. Persistence Study:A study to assess changes in program impacts Process Evaluation:A systematic assessment of an energy efficiency over time.See savings persistence rate and measure persistence. program for the purposes of documenting program operations at Portfolio:Either(1)a collection of similar programs addressing the the time of the examination,and identifying and recommending same market(e.g.,a portfolio of residential programs),technology improvements to increase the program's efficiency or effectiveness (e.g.,motor efficiency programs),or mechanisms(e.g.,loan for acquiring energy resources while maintaining high levels of participant satisfaction. A-12 www.seeaction.energy.gov December 2012 Program:An activity,strategy,or course of action undertaken by an willingness to sign up for the program). Randomization creates a implementer or administrator.Each program is defined by a unique control group that is statistically identical to the treatment group, combination of the program strategy, market segment,marketing in both observable and unobservable characteristics,such that any approach,and energy efficiency measure(s)included. Programs difference in outcomes between the two groups can be attributed to consist of a group of projects with similar characteristics and the treatment with a high degree of confidence. installed in similar applications. Randomized Controlled Trial(RCT):A type of experimental program Program Administrator Cost Test(PACT, See utility/program evaluation design in which energy consumers in a given population administrator cost test. are randomly assigned into two groups:a treatment group and a control group.The outcomes for these two groups are compared, Program Incentive:An incentive,generally monetary,that is offered resulting in program energy savings estimates. to a customer through an energy efficiency program to encourage the customer to participate in the program.The incentive is intended Ratepayer Impact Measure(RIM)Test:A cost-effectiveness test that to overcome one or more barriers that keep the customer from measures the impact on utility operating margin and whether rates taking the energy efficiency activity on his own. would have to increase to maintain the current levels of margin if a customer installed energy efficient measures.The test measures Program Theory:A presentation of the goals of a program, what happens to customer bills or rates due to changes in utility incorporated with a detailed presentation of the activities that the revenues and operating costs caused by the program. program will use to accomplish those goals and the identification of the causal relationships between the activities and the program's For dollars,"real"means that the dollars are expressed in effects. Program theory is often the basis for the logic model. a specific base year in order to provide a consistent means of comparison after accounting for inflation.For interest and discount Program Year(PY):The calendar year approved for program rates,"real"means the inflation rate is not included(the nominal implementation. Note that program years can be shorter than rate minus the inflation rate approximately equals the real rate). 12 months if programs are initiated mid-year. Realization Rate:Used in several contexts for comparing one savings Project:An activity or course of action involving one or multiple estimate with another.The primary and most meaningful application energy efficiency measures at a single facility or site. is the ratio of evaluated gross savings to claimed gross savings (versus comparing net and gross savings estimates,which are best Projected Savings:Values reported by an implementer or defined with a net-to-gross ratio). Basis for the ratio not being 1.0 administrator before the subject energy efficiency activities are can include several considerations,such as(1)adjustments for data implemented;also called planning estimates or ex ante estimates. errors,(2)differences in implemented measure counts as a result of Projected savings are typically estimates of savings prepared for verification activities,and/or(3)other differences revealed through program and/or portfolio design or planning purposes.These values the evaluation process,such as changes in baseline assumptions. are typically based on pre-program or portfolio estimates of factors Rebate:See incentive. such as per-unit savings values,operating hours,installation rates, and savings persistence rates.These values may use results of prior Rebate Progran An energy efficiency program in which the program evaluations and/or values in a technical reference manual.They can administrator offers a financial incentive for the installation of be indicated as first-year,annual,and/or lifetime energy or demand energy-efficient equipment. savings values.They can be indicated as gross savings and/or net savings values. Rebound Effect:A change in energy-using behavior that yields an increased level of service accompanied by an increase in energy Random Assignment:A condition where each household or entity use that occurs as a result of taking an energy efficiency action; in the study population is randomly assigned to either the control also called take back.The result of this effect is that the absolute group or the treatment group based on a random probability,as savings associated with the direct energy efficiency action is opposed to being assigned to one group or the other based on impacted(usually reduced)by the resulting behavioral change. some characteristic of the household(e.g.,location,energy use,or The rebound effect can be considered as a factor in determining savings persistence rate. December 2012 www.seeaction.energy.gov A-13 Regression Analysis:Analysis of the relationship between a to isolate the energy flows for the system(s)under consideration. dependent variable(response variable)to specified independent IPMVP Option A involves"Key Parameter Measurement"and IPMVP variables(explanatory variables).The mathematical model of their Option B involves'All Parameter Measurement." relationship is the regression equation. Retrofit Program:An energy efficiency program that provides incentives,information,and technical support to encourage Regression Model:A mathematical model based on statistical customers to replace existing and operating equipment or systems analysis where the dependent variable is quantified based on its with more energy-efficient equipment or systems that provides the relationship to the independent variables that are said to determine same function,or to add energy efficient equipment or systems to an its value. In so doing,the relationship between the variables is existing facility(e.g.,addition of thermal insulation). estimated statistically from the data used. Rigor:The level of expected confidence and precision.The higher Reliability:The quality of a measurement process that would the level of rigor,the more confident one is that the results of the produce similar results on(1)repeated observations of the same evaluation are reliable. condition or event,or(2)multiple observations of the same condition or event by different observers.Reliability refers to the Sample:In program evaluation,a portion of the population selected likelihood that the observations can be replicated. to represent the whole. Differing evaluation approaches rely on simple or stratified(based on some characteristic of the population) Reporting Period:The time following implementation of an energy samples. efficiency activity during which savings are to be determined. Sample Design:The approach used to select the sample units. Representative Sample:A sample that has approximately the same distribution of characteristics as the population from which it was Sampling Error:An error that arises because the data are collected drawn.Such samples may be randomly selected or not. Random from a part,rather than the whole of the population. It is usually samples selected from the same population as participants are more measurable from the sample data in the case of probability sampling. representative of the program participants. Savings Persistence Rat Percentage of first-year energy or demand Resource Acquisition Program:A program designed to achieve savings expected to persist over the life of the installed energy directly energy and/or demand savings;also called retrofit program. efficiency equipment.It is developed by conducting surveys of Such a program generally involves encouraging customers to replace installed equipment several years after installation to determine existing equipment with more-efficient equipment. presence and operational capability of the equipment. Retro-commissioning:Often abbreviated as"RCx,"a systematic Seasonal Energy Efficiency Ratio(SEER)-The total cooling output of method for investigating how and why an existing building's systems a central air-conditioning unit(in Btu)during its normal use period are operated and maintained,and for identifying ways to improve for cooling divided by the total electrical energy input(in watt-hours) overall building performance. Retro-commissioning improves during the same period,as determined using specified federal a building's operations and maintenance(0&M)procedures to test procedures. enhance overall building performance. Secondary Effects: Unintended impacts of the project or program Energy efficiency activities undertaken in existing residential such as rebound effect(e.g.,increasing energy use as it becomes or nonresidential buildings,where existing inefficient equipment or more efficient and less costly to use),activity shifting(e.g.,when systems are replaced by energy-efficient equipment or systems or generation resources move to another location),and market leakage where efficient equipment or systems are added to an existing (e.g.,emission changes due to changes in supply or demand of facility(e.g.,addition of thermal insulation). commercial markets).Secondary effects can be positive or negative. Retrofit Isolation:The savings measurement approach defined Snap Back:See rebound effect. in IPMVP Options A and B,as well as ASHRAE Guideline 14,that determines energy or demand savings through the use of meters Societal Cost Test(SCT):A cost-effectiveness test that measures the net economic benefit to the utility service territory,state,or region, A-14 www.seeaction.energy.gov December 2012 as measured by the total resource cost test, plus indirect benefits Structured Interview:An interview in which the questions to be such as environmental benefits. asked,their sequence,and the detailed information to be gathered are all predetermined.These are used where maximum consistency Spillover(Participant and Non-Participant): Reductions in energy across interviews and interviewees is needed. consumption and/or demand caused by the presence of an energy efficiency program,beyond the program-related gross savings of Take Back Effect:See rebound effect. the participants and without direct financial or technical assistance from the program.There can be participant and/or non-participant Technical Degradation Factor:A multiplier used to account for time- spillover.Participant spillover is the additional energy savings and use-related change in the energy savings of a high-efficiency that occur as a result of the program's influence when a program measure or practice relative to a standard-efficiency measure or participant independently installs incremental energy efficiency practice due to technical operational characteristics of the measures, measures or applies energy-saving practices after having participated including operating conditions and product design. in the energy efficiency program.Non-participant spillover refers to energy savings that occur when a program non-participant installs Technical Reference Manual(TRM):A resource document that energy efficiency measures or applies energy savings practices as a includes information used in program planning and reporting result of a program's influence. of energy efficiency programs. It can include savings values for measures,engineering algorithms to calculate savings,impact factors Spillover Rate: Estimate of energy savings attributable to spillover to be applied to calculated savings(e.g.,net-to-gross ratio values), effects expressed as a percent of savings installed by participants source documentation,specified assumptions,and other relevant through an energy efficiency program. material to support the calculation of measure and program savings—and the application of such values and algorithms in Statistically Adjusted Engineering(SAE)Models:A category of appropriate applications. statistical analysis models that incorporate the engineering estimate of savings as a dependent variable.The regression coefficient in these Total Resource Cost(TRC)Test:A cost-effectiveness test that models is the percentage of the engineering estimate of savings measures the net direct economic impact to the utility service observed in changes in energy use. For example,if the coefficient on territory,state,or region. the SAE term is 0.8,this means that the customers are on average realizing 80%of the savings from their engineering estimates. Tracking Estimate:See claimed savings. Stipulated Values:See deemed savings. Treatment Group:The group of consumers that receive the subject program's intervention or"treatment"(i.e.,the program participants). Stratified Random Sampling:A sampling method where the population is divided into X units of subpopulations,called Uncertainty:The range or interval of doubt surrounding a measured strata,that are non-overlapping and together make up the entire or calculated value within which the true value is expected to fall population.A simple random sample is taken of each strata to create within some degree of confidence. a sample based upon stratified random sampling. Upstream Program:A program that provides information and/or Stratified Ratio Estimation:A sampling method that combines financial assistance to entities in the delivery chain of high-efficiency a stratified sample design with a ratio estimator to reduce the products at the retail,wholesale,or manufacturing level.Such a coefficient of variation by using the correlation of a known measure program is intended to yield lower retail prices for the products. for the unit(e.g.,expected energy savings)to stratify the population Utility/Program Administrator Cost Test:Acost-effectiveness test and allocate samples from strata for optimal sampling.Stratified that measures the change in the amount the utility must collect from ratio estimation can reduce the number of sites,observations,and the customers every year to meet an earnings target(e.g.,a change thus evaluation costs required to achieve a given level of precision by in revenue requirement);also called program administrator cost test using the correlation of a known measure for the unit(e.g.,expected (PACT)and also known as the utility cost test. In several states,this energy savings)to stratify the population and allocate sample test is referred to as the program administrator cost test. In those requirements for each strata. cases,the definition of the"utility"is expanded to program administrators(utility or third-party). December 2012 www.seeaction.energy.gov A-15 Verification:An independent assessment that the program has been implemented per the program design.For example,the objectives of measure installation verification are to confirm(1)the installation rate,(2)that the installation meets reasonable quality standards,and (3)that the measures are operating correctly and have the potential to generate the predicted savings.Verification activities are generally conducted during on-site surveys of a sample of projects.Project site inspections,participant phone and mail surveys,and/or implementer and consumer documentation review are typical activities associated with verification.Verification may include one-time or multiple activities over the estimated life of the measures. It may include review of commissioning or retro-commissioning documentation. Verification is a subset of evaluation and,as such,can also include review and confirmation of evaluation methods used,samples drawn,and calculations used to estimate program savings. Whole-Building Calibrated Simulation Approach:A savings measurement approach(defined in IPMVP Option D and ASHRAE Guideline 14)that involves the use of an approved computer simulation program to develop a physical model of the building in order to determine energy and demand savings.The simulation program is used to model the energy used by the facility before and after the retrofit.The pre-or post-retrofit models are calibrated with measured energy use and demand data as well as weather data. Whole-Building Metered Approach:A savings measurement approach(defined in the IPMVP Option C and ASHRAE Guideline 14)that determines energy and demand savings through the use of whole-facility energy(end-use)data,which may be measured by utility meters or data loggers.This approach may involve the use of monthly utility billing data or data gathered more frequently from a main meter. Workforce Education and Training Program Programs primarily intended for building maintenance engineers, HVAC contractors, engineers,architects,maintenance personnel,and others.These programs provide information about energy efficiency concepts, recommended energy-efficient technologies or behaviors,and/or programs that offer energy efficiency or load-reduction information, products or services.These programs may provide indirect energy or demand savings. A-16 www.seeaction.energy.gov December 2012 Glossary References California Commissioning Collaborative. (2006). California Northeast Energy Efficiency Partnerships. (July 2011).EM&V Forum Commissioning Guide:New Buildings and California GLOSSARY OF TERMS Version 2.1.PAH Associates.A project of the Commissioning Guide:Existing Buildings. www.cacx.org/ Regional Evaluation,Measurement and Verification Forum. resources/commissioning-guides.html. http://neep.orgluploadslEMV°o20ForumlEMVYa20ProductsI EMV Glossary-Version_2.1.pdf. California Public Utilities Commission.(April 2006).California Energy Efficiency Evaluation Protocols:Technical, Methodological,and State and Local Energy Efficiency Action Network. (May 2012). Reporting Requirements for Evaluation Professionals. Prepared Evaluation,Measurement,and Verification(EM&V)of Residential by The TecMarket Works Team. www.calmac.org/events/ Behavior-Based Energy Efficiency Programs:Issues and EvaluatorsProtocols_Final AdoptedviaRuling_06-19-2006.pdf. Recommendations.Prepared by Todd,A.;Stuart,E.;Schiller,S.; Goldman, C.;Lawrence Berkeley National Laboratory. Efficiency Valuation Organization. (September2010).International wwwl.eere.energy.gov/seeaction/pdfs/emv_behaviorbased Performance Measurement and Verification Protocol.EVO eeprograms.pdf. 10000-1:2010. Concepts and Options for Determining Energy and Water Savings Volume 1. www.evo-world.org. World Resources Institute and World Business Council for Sustainable Development. (November 2005). The GHG Protocol Eto,J.;Prahl,R.;Schegel,J. (1996).A Scoping Study on Energy- for Project Accounting. www.ghgprotocol.org/files/ghgp/ Efficiency Market Transformation by California Utility DSM ghg_project protocol.pdf. Programs. Lawrence Berkeley National Laboratory. http://eetd.lbl.gov/ea/ems/reports/39058.pdf. National Action Plan for Energy Efficiency. (2007). Guide for Conduct- ing Energy Efficiency Potential Studies. Optimal Energy,Inc. www.epa.gov/cleanenergy/documents/suca/potential guide.pdf. National Action Plan for Energy Efficiency. (2007).Model Energy Efficiency Program Impact Evaluation Guide.Steven R.Schiller, Schiller Consulting, Inc. www.epa.gov/eeactionplan. National Action Plan for Energy Efficiency. (2008). Understanding Cost-Effectiveness of Energy Efficiency Programs:Best Practices, Technical Methods,and Emerging Issues for Policy-Makers. Prepared by Energy and Environmental Economics,Inc. (E3)and Regulatory Assistance Project(RAP). www.epa.gov/cleonenergy/ documents/suca/cost-effectiveness.pdf National Action Plan for Energy Efficiency. (2009).Energy Efficiency as a Low-Cost Resource for Achieving Carbon Emissions Reductions.ICF International,Inc. www.epa.gov/cleanenergy/ documents/suca/ee_and carbon.pdf. December 2012 www.seeaction.energy.gov A-17 Appendix B: Other Evaluation Categories and Approaches This appendix provides a brief introduction to process and market 13,1.1 Process Evaluations effects evaluations,cost-effectiveness analysis,and impact evalua- The goal of process evaluations is to produce better and more cost- tions using"top-down"approaches.The material in this appendix is effective programs. Process evaluations meet this goal by assessing intended to supplement the other sections of the guide,which are the processes a program undergoes during implementation,docu- focused on"bottom-up"impact evaluations. menting program goals and objectives from a variety of perspectives, B.1 PROCESS, MARKET EFFECTS, AND and describing program strengths and weaknesses so that success COST EFFECTIVENESS EVALUATIONS is highlighted and improvements can be made in a timely manner. Thus, process evaluations examine the efficiency and effectiveness The following subsections introduce three non-impact types of evalu- of program implementation procedures and systems.Typical process ations:process,market,and cost-effectiveness. However,because evaluation results involve recommendations for changing a pro- cost-effectiveness analyses rely on the documentation of program gram's structure,implementation approaches,and goals. impacts,these analyses are often considered a component of impact evaluations,and program cost-effectiveness indicators are thus often These evaluations usually consist of asking questions of those included in impact evaluation reports.Table 13.1 compares these involved in the program,analyzing their answers,and comparing three evaluation types,plus impact evaluations. results to established best practices.Whereas it is typically required that an independent third-party evaluator is involved in conducting TABLE BA: Program Evaluation Types EVALUATION TYPE DESCRIPTION EXAMPLES USES Impact Evaluation Quantifies direct and indirect changes Determines the amount of energy and associated with the subject program(s) demand saved Process Evaluation Indicates how the procedures associated Identifies how program designs and with program design and implementation processes can be improved are performing from both the administra- tor's and the participants' perspectives Market Effects Evaluation Analyzes how the overall supply chain Characterizes changes that have occurred and market for energy efficiency products in efficiency markets and whether they have been affected by the program are attributable to and sustainable with or without the program Cost-Effectiveness Evaluation Quantifies the costs of program Determines whether an energy efficiency implementation and compares them program is a cost-effective investment with program benefits compared with other programs and energy supply resources B-1 www.seeaction.energy.gov December 2012 impact evaluations,for process evaluations,jurisdictions might variety of outputs that lead to key outcomes(see sidebar in Chapter recommend(but not require)them to be conducted by independent 7 on Theory-Based Evaluation:A Guiding Principle for MT Evaluation). third-party evaluators; however,the use of third-party process evalu- Logic models can be linked to performance indicators that provide ators is a best practice. Use of a trusted party for process evaluation ongoing feedback to program managers.The models usually flow top is important for successful process evaluation so that the evaluator to bottom and are often organized according to five basic categories: can gather the necessary data and provide feedback in a manner • Program inputs:financial,staffing,and infrastructure resources that is productive(e.g.,not considered threatening by the recipient that support the activity of the feedback). • Program activities:overarching activities that describe what Process evaluations are particularly valuable in the following situations: the program is doing(e.g.,marketing and rebate processing) • Benefits are higher/lower than expected and/or are being • Outputs:metrics resulting from the activities,and that tend to achieved more quickly/slowly than expected. be measurable"bean counting" results(e.g., provide outreach • There is limited program participation or stakeholders are slow events at five community fairs) to begin participating. • Short-to intermediate-term outcomes:expected outcomes • The program is a greater success than anticipated. resulting from program activities,with goals attached to those outcomes when possible.(e.g.,target energy savings and • The program has a slow start-up. recruitment into the program) • Participants are reporting problems. • Long-term outcomes and goals:ideal,sustainable outcomes • The program appears not to be cost effective. resulting from program activities(e.g.,"all eligible customers • The program is built around a new concept that could be participate in the program"and"increase customer awareness replicable for other populations,technologies,etc. of program offerings"). As part of a process evaluation,a logic model may be developed for the These logic model categories indicate the intended and expected program(or possibly a set of logic models for a complete portfolio of results of activities.Expected short-, medium-,and long-term out- programs).A program's theory and logic model serve as a roadmap to comes tend to define program goals at a high level and also specify guide the systematic approach of a process evaluation.A program logic market effects(i.e.,expected program outcomes). In this manner, model is a visual representation of the program's theory that illustrates process evaluation is part of a continuum linking impact and market a set of interrelated program activities that combine to produce a effects evaluations. TABLE B.2: Elements of Typical Process Evaluations Program Design Program Implementation • The program mission • Quality control • Assessment of program logic • Operational practice—how the program is implemented • Use of new practices or best practices • Program targeting,marketing,and outreach efforts • Program timing Program Administration Participant Response • Program oversight • Participant interaction and satisfaction • Program staffing • Market and government allies interaction and satisfaction • Management and staff training • Program information and reporting December 2012 www.seeaction.energy.gov B-2 The primary mechanism of process evaluations is data collection programs.In this guide,the definition of market transformation is: (e.g.,surveys,questionnaires,and interviews)from administrators, a reduction in market barriers resulting from a market intervention, designers,participants(e.g.,facility operators,business owners, as evidenced by a set of market effects,that is likely to last after the renters,or homeowners),implementation staff(including contrac- intervention has been withdrawn, reduced,or changed. tors,subcontractors,and field staff),trade allies(e.g.,mechanical contractors,architects,and engineers)and key policymakers.Other Market effects evaluations often involve a significant undertaking, elements of a process evaluation can include workflow and produc- because they require collection and analysis of data from a wide tivity measurements;reviews,assessments,and testing of records, range of market actors,as well as analysis of those data against a databases,program-related materials,and tools;and collection and background developed out of secondary sources.Market effects are analysis of relevant data from third-party sources(e.g.,equipment sometimes called the ultimate test of a program's success,answering vendors or retailers).Process evaluations can be operated continu- the question:"Will energy efficiency(best)practices continue in the ously,perhaps as part of a continuous improvement effort,or at marketplace,even after the current program ends?"The difference intervals(e.g.,as a new program is being implemented,whenever between a market change and a market effect is attribution:the there are major changes in a program,in response to issues noted in ability to trace back a change in the market to a specific program or first set of bullets above,and/or just every two to three years). group of programs.The following is a definition of market effects from a well-referenced 1996 study:los Table B.2 lists examples of program elements typically assessed during a process evaluation. Market effect:a change in the structure of a market or the behavior of market participants that is reflective of an increase in B.1.2 Market Effects Evaluations the adoption of energy-efficient products,services,or practices The goal of market effects evaluations is to characterize and quantify and is causally related to market intervention(s)(e.g.,programs). the effects of a program on supplier promotion and customer Examples of market effects include increased levels of awareness adoption of the targeted energy efficiency measures,regardless of of energy-efficient technologies among customers and suppli- whether those suppliers and customers participated in the program. ers,increased availability of efficient technologies through retail Effects that cannot be captured by program records are particularly channels,reduced prices for efficient models,build-out of efficient important for certain kinds of initiatives,including"upstream" model lines,and—the end goal—increased market share for promotions of mass-market goods,such as light bulbs and consumer efficient goods,services,and design practices. electronics as well as training programs aimed at inducing engineers and contractors to adopt energy efficiency design and specification POTENTIAL STUDIES practices.Studies have shown that even straightforward equipment rebate programs may have effects"outside the program"by Another form of market study(although not formally an exposing contractors and large customers to the benefits of efficient "evaluation")is called a potential study.Potential studies technologies.This in turn leads to increased specification of efficient are conducted before a program is implemented in order technologies on projects that do not receive program support. In to assess market baselines and future savings potentials for some cases, market effects evaluation results can be combined with different efficiency technologies,strategies,or approaches impact evaluation findings to estimate program-induced energy in different customer markets.These studies can also assess savings that were not tracked by the program itself. customer needs and barriers to adoption of energy efficiency, Other market studies include potential studies(see sidebar)and as well as how best to address these barriers through pro- market baseline studies. Potential studies investigate how much gram design. Potential studies indicate what can be expected saving may be available through various measures and baseline in terms of savings from a program. Potential is often defined studies look at indicators of market development before the in terms of technical potential(what is technically feasible program intervention. given commercially available products and services),and economic potential(which is the level of savings that can Market effects studies are usually associated with programs that be achieved assuming a certain level of participant and/or have a specific market transformation focus.There are many defini- societal cost effectiveness is required).Findings also help tions of market transformation,although it is often considered the managers identify the program's key markets and clients and ultimate goal of publicly and consumer-funded energy efficiency how to best serve the intended customers. B-3 www.seeaction.energy.gov December 2012 Examples of the questions that a market effects evaluation might without several years of information;even then,they will usually answer are as follows: not have the same level of accuracy as impact evaluations of direct • Are the entities that undertook energy efficiency projects resource acquisition savings.To credit measure adoption and undertaking additional projects or incorporating additional associated savings to a program,it must be shown that the increased technologies in their facilities that were not directly induced energy efficiency adoption,the longer-term market effects,and the by the program?This might indicate that facility operators have participant effects have all occurred essentially in the manner and become convinced of the value of,for example, high-efficiency in the order specified by the program theory.And this,for most motors,and are installing them on their own. programs,takes a number of years to reach this point. • Are entities that did not undertake projects now adopting In 2009,a comprehensive white paper study on market transformation concepts and technologies that were encouraged by the and market effects was prepared.106 Table 13.3,from a presentation program?This might indicate that the program convinced by the principle author of this white paper,indicates approaches for other facility operators of the advantages of the energy assessing market effects,including attribution. efficiency concepts. • Are manufacturers,distributors,retailers,vendors,and others As can be deduced from the above discussion and Table 13.3,the involved in the supply chain of energy efficiency products market effects evaluation can easily overlap with the spillover analyses (and services)changing their product offerings—for example, conducted as part of an impact evaluation. In fact,many of the tech- how are they marketing them,pricing them,stocking them? niques used to quantify market effects can be applied in estimating The answers can indicate how the supply chain is adapting to spillover savings. changes in supply of and demand for efficiency products. 107,108 B.1.3 Cost-Effectiveness Analyses Structuring a market effects evaluation entails consideration of sev- Cost-effectiveness(sometime called benefit-cost)evaluations compare eral levels or stages,with the ultimate goal generally understood to program benefits and costs,showing the relationship between the be the increased adoption of energy efficiency goods and services in value of a program's benefits and the costs incurred to achieve those the general market leading to energy savings. Energy savings are the benefits.The findings help judge whether to retain,revise,or eliminate ultimate goal of programs seeking to cause market effects(i.e.,the program elements and provide feedback on whether efficiency is an intended long-term outcome).The following list suggests a hierarchy effective investment,compared with energy supply options.Cost- of precursors to that goal: effectiveness evaluation is also often a key component of the evalua- • Early Acceptance:proliferation of models and manufacturers, tion process for programs using public or utility customer funds. purchase by frontrunners and enthusiasts In 1983,California's Standard Practice for Cost-Benefit Analysis of • Take-off Phase:customer awareness,visibility on shelves and Conservation and Load Management Programs manual(SPM)devel- in inventories, perceptible levels of market share in the supply oped five cost-effectiveness tests for evaluating energy efficiency channels programs.These approaches,with minor updates,continue to be • Maturity:all major competitors offer energy efficient models; used today and are the principal approaches used for evaluating codes and standards include energy efficient models energy efficiency programs across the United States.109 • Energy Savings:energy savings attributable to the program The five tests vary in terms of(1)perspectives(project participants, are associated with acceleration of these developments. ratepayers,utilities,or society),(2)their applicability to different In general,the achievement of goals at each of the higher levels program types,(3)the cost and benefit elements included in the of the hierarchy requires accomplishments at the lower levels.As calculation,(4)the methods by which the cost and benefit elements a result,tracking goals at each stage not only provides feedback are computed,and(5)the uses of the results. Most regulated utility on performance with respect to that goal itself,but also provides energy efficiency programs use one or more versions of these tests, evidence that effects at the next-higher levels can be attributed to sometimes with variations unique to the requirements of a particular the program. regulatory commission.Definitions of these tests(paraphrased from the SPM)are as follows on page 144: Goals will typically be set and tracked for different time frames and • Total Resource Cost(TRC)Test.The TRC test measures the net for different purposes.While energy savings are the ultimate market costs of a demand-side management program as a resource effects goal,in most cases,savings cannot be measured meaningfully option based on the total costs of the program,including December 2012 www.seeaction.energy.gov B-4 TABLE B.3:Approaches for Assessing Market Effects, Including Attribution BASIC SOURCE/RELATIVE ADVANTAGES LIMITATIONS SURVEYS OF CUSTOMER PURCHASES • Limited accuracy on key details:exact number,timing, • Can be deployed quickly,relatively inexpensively,and efficiency rating of purchases repeatedly over extended time frames • Non-response bias a problem,particularly in early stages • Can be deployed in program and non-program areas of market development • Generally produces reliable data on number of • Difficult to validate results in absence of some comparison purchases/adoptions to sales or program volumes SURVEYS OF SUPPLY-SIDE ACTORS • Difficult to build measures of sales volume—may need to • Taps into close knowledge of local markets be content with estimates of market share • Respondents sufficiently knowledgeable to provide • In many jurisdictions,population available to be sampled accurate information on product features is small • Difficult to validate results in absence of some comparison to sales or program volumes SHIPMENT AND SALES DATA • Requires negotiated cooperation of manufacturers and • Conceptually,the most accurate and detailed measure retailers;risk of dropouts of adoption:quantity,efficiency,timing • Difficult to obtain coverage of all sectors,time periods,and regions(and may be costly) • Quality control is difficult CUSTOMER-REPORTED FREE RIDERSHIP AND SPILLOVER • For non-participants, requires that customers be aware • Can be deployed quickly,relatively inexpensively,and of the program and able to judge its impact on adoption repeatedly over extended time frames decisions • Can probe adoption process and decisions • Consistent with current Performance Earnings Basis(PEB) methods now in force in California CROSS-SECTIONAL METHODS • Increasingly difficult to find non-program areas • Closest to conventional social science research methods; • Difficult to verify comparability of non-program areas intuitively satisfying • Appears to be effective only in time-limited periods • Data provide insight into exogenous factors,working of . Logistically demanding and time consuming market beyond program boundary EXPERT JUDGING • Not a statistical estimation process • Focuses insights from experienced market participants • Difficult to identify and account for factors affecting and observers individual judgments • Results can be expressed in terms of net adoptions • In some cases,can be deployed fairly rapidly Source:Rosenberg,M.(June 2010)"Market Effects and Market Transformation:Their Role in Program Design and Evaluation."EPA EM&V Webinar.www.emvwebinar.org. B-5 www.seeaction.energy.gov December 2012 both the participants'and the utility's costs. It combines the calculation.Results are reported either in net present value(NPV) perspectives of participants and non-participants,which is why dollars(method by difference)or as a ratio(i.e.,benefits/costs).Table it is also often called an"all ratepayers" perspective.The TRC B.4 outlines the basic approach underlying cost-effectiveness tests. ratio equals the benefits of the program,in terms of value of energy and demand saved,divided by the net costs.The ratio Each of the tests provides a different kind of information about is usually calculated on a lifecycle basis,considering savings the impacts of energy efficiency programs from different vantage and costs that accrue over the lifetime of installed energy points in the energy system.On its own,each test provides a single efficiency equipment or systems.This is a commonly applied stakeholder perspective.Together,multiple tests provide a com- cost-effectiveness test. prehensive approach to answering key questions:"Is the program • Program Administrator Cost Test(PACT).The PACT mea- effective overall?""Is it balanced?"`Are some costs or incentives too sures the net costs of a demand-side management program high or too low?""What is the effect on rates?""What adjustments as a resource option based on the costs incurred by the are needed to improve the alignment?" program administrator(often a utility,though it can be any Overall,the results of all five cost-effectiveness tests provide a more organization),excluding any net costs incurred by the partici- complete picture than the use of any one test alone.The TRC and SCT pant.The benefits are the same as the TRC benefits(energy cost tests help to answer whether energy efficiency is cost-effective and demand savings value).The PACT is also a commonly overall.The PCT,PACT,and RIM help to answer whether the selection applied test. of measures and design of the program is balanced from participant, • Participant Cost Test(PCT).The PCT assesses cost effectiveness utility,and non-participant perspectives,respectively.Looking at the from the participating consumer's perspective by calculating cost-effectiveness tests together helps to characterize the attributes the quantifiable benefits and costs to the consumer of par- of a program or measure to enable decision making,to determine ticipating in a program. Because many consumers do not base whether some measures or programs are too costly,whether some their decision to participate entirely on quantifiable criteria, costs or incentives are too high or too low,and what adjustments this test is not necessarily a complete measure of all the need to be made to improve distribution of costs and benefits among benefits and costs a participant perceives. stakeholders.The scope of the benefit and cost components included • Societal Cost Test(SCT).The SCT,a modified version of the TRC, in each test is summarized in Table B.5 and Table B.6. adopts a societal rather than a utility service area perspective.The primary difference between the societal and TRC tests is that,to The broad categories of costs and benefits included in each cost- calculate lifecycle costs and benefits,the societal test accounts effectiveness test are consistent across all regions of the country for externalities(e.g.,environmental benefits),excludes tax credit and applications. However,the specific components included in benefits,and uses a(often lower)societal discount rate. each test may vary across different regions, market structures,and • Ratepayer Impact Measure(RIM)Test.The RIM test only utility types. For example,transmission and distribution investment may be considered deferrable through energy efficiency in some applies to utility programs. It examines the potential impact that the energy efficiency program has on rates overall.The net areas and not in others. Likewise,the TRC and SCT may consider just benefits are the avoided cost of energy(same as PACT).The natural gas or electricity resource savings in some cases, but also net costs include the overhead and incentive costs(same as include co-benefits of other savings streams(such as water and fuel PACT)but also include utility lost revenues from customer bill oil)in others. savings. Historically, reliance on the RIM test has limited energy Also,for the SCT,how the"non-monetized benefits"in Tables B.5 and efficiency investment,as it is the most restrictive of the five B.6 are determined is an evolving area.In particular,benefits that cost-effectiveness tests. in the past could not be monetized(e.g.,air quality impacts)now The basic structure of each cost-effectiveness test involves a calcula- can be assigned monetary values and,in fact,need to be assigned lion of the total benefits and the total costs in dollar terms from a such values in order to be used in cost-effectiveness equations.Also, certain vantage point to determine whether or not the overall ben- non-energy benefits,which in the past might have been ignored, efits exceed the costs.A test is positive if the benefit-to-cost ratio is are being shown to have significant value.These include economic greater than one,and negative if it is less than one—with,of course, development and employment benefits as more money is spent on local services and products because of the efficiency investments. proper consideration of uncertainties in the inputs used in the December 2012 www.seeaction.energy.gov B-6 TABLE B.4: The Five Principal Cost-Effectiveness Tests Used in Energy Efficiency ACRONYMTEST • OF - • Participant cost test PCT Will the participants benefit over the Comparison of costs and benefits of the measure life? customer installing the measure Program administrator PACT Will utility bills increase? Comparison of program administrator cost test costs to supply-side resource costs Ratepayer impact measure RIM Will utility rates increase? Comparison of administrator costs and utility bill reductions to supply-side resource costs Total resource cost test TRC Will the total costs of energy in the Comparison of program administrator and utility service territory decrease? customer costs to utility resource savings Societal cost test SCT Is the utility,state,or nation better off Comparison of society's costs of energy as a whole? efficiency to resource savings and non- cash costs and benefits California Public Utilities Commission(CPUC).(2001).California Standard Practice Manual:Economic Analysis of Demand-Side Programs and Projects.www.energy.ca.gov/greenbuilding/ documents/background/07-J CPUC STANDARD PRACTICE MANUAL.PDF. National Action Plan for Energy Efficiency.(2008).Understanding Cost-Effectiveness of Energy Efficiency Programs:Best Practices,Technical Methods,and Emerging Issues for Policy-Makers. Prepared by Energy and Environmental Economics,Inc.(E3)and Regulatory Assistance Project(RAP).www.epa.gov/cleonenergy/documents/suca/cost-effectiveness.pdf. B-7 www.seeaction.energy.gov December 2012 TABLE B.5: Description of Benefits and Costs Included in Each Cost-Effectiveness Test PCT Benefits and costs from the perspective of the customer installing the measure • Incentive payments • Incremental equipment costs • Bill savings • Incremental installation costs • Applicable tax credits or incentives i PACT Perspective of utility,government agency,or third party implementing the program • Energy-related costs avoided by the utility • Program overhead cost • Capacity-related costs avoided by the utility, including generation, • Utility/program administrator incentive costs transmission,and distribution • Utility/program administrator installation costs RIM Impact of efficiency measure on non-participating ratepayers overall • Energy-related costs avoided by the utility • Program overhead cost • Capacity-related costs avoided by the utility, • Utility/program administrator incentive costs including generation,transmission,and distribution • Utility/program administrator installation costs • Lost revenue due to reduced energy bills TRC Benefits and costs from the perspective of all utility customers(participants and non-participants)in the utility service territory • Energy-related costs avoided by the utility • Program overhead costs • Capacity-related costs avoided by the utility, including generation, • Program installation costs transmission,and distribution • Incremental measure costs(weather paid by • Additional resource savings(i.e.,gas and water if utility is electric) the customer or utility) • Monetized environmental and non-energy benefits (see Section 4.9) • Applicable tax credits(see Section 6.4) SCT Benefits and costs to all the utility service territory,state,or nation as a whole • Energy-related costs avoided by the utility • Program overhead costs • Capacity-related costs avoided by the utility,including generation, • Program installation costs transmission,and distribution • Incremental measure costs(weather paid by • Additional resource savings(i.e.,gas and water if utility is electric) the customer or utility) • Non-monetized benefits(and costs)such as cleaner air or health impacts California Public Utilities Commission(CPUC).(2001).California Standard Practice Manual:Economic Analysis of Demand-Side Programs and Projects. www.energy.ca.gov/greenbuilding/documents/background/07-J CPUC STANDARD_PRACTICE MANUAL.PDF. National Action Plan for Energy Efficiency.(2008).Understanding Cost-Effectiveness of Energy Efficiency Programs:Best Practices,Technical Methods,and Emerging Issues for Policy- Makers.Prepared by Energy and Environmental Economics,Inc.(E3)and Regulatory Assistance Project(RAP).www.epa.gov/cleanenergy/documents/suca/cost-effectiveness.pdf. December 2012 www.seeaction.energy.gov B-8 TABLE B.6: Summary of Benefits and Costs Included in Each Cost-Effectiveness Test PCT PACT RIM Energy and capacity-related avoided costs Benefit Benefit I Benefit Benefit Additional resource savings Benefit Benefit Non-monetized benefits Benefit i Incremental equipment and installation costs Cost Cost i Program overhead costs Cost Cost Cost Cost Incentive payments Benefit Cost Cost LBill savings Benefit Cost California Public Utilities Commission(CPUC).(2001).California Standard Practice Manual:Economic Analysis of Demand-Side Programs and Projects. www.energy.co.gov/greenbuilding/documents/background/07-J CPUC STANDARD_PRACTICE MANUAL.PDF. National Action Plan for Energy Efficiency.(2008).Understanding Cost-Effectiveness of Energy Efficiency Programs:Best Practices,Technical Methods,and Emerging Issues for Policy- Makers.Prepared by Energy and Environmental Economics,Inc.(0)and Regulatory Assistance Project(RAP).www.epa.gov/cleanenergy/documents/suca/cost-effectiveness.pdf. B.2 TOP-DOWN IMPACT EVALUATION estimating net program savings at the sector,state,regional,and Top-down impact evaluation refers to methods that rely on aggregate national levels.The primary potential drawbacks of top-down evalu- energy consumption data or per-unit energy consumption indices ation are the difficulty in attributing energy consumption changes (e.g.,energy consumption per-unit of output or per capita)defined by to specific energy efficiency policies and/or particular programs and sector,utility service territory,state,region,or country as the starting actions. point for determining energy savings.Top-down evaluation focuses A metric that can be considered the output of top-down evaluation on the bottom line—reductions in energy use(and/or demand)for a is gross market savings.These are the energy savings resulting from state,region,or utility service territory.This gives top-down evaluation energy efficiency programs,codes and standards,and naturally a direct link to(1)demand forecasting and resource planning,and occurring adoption,and which have a long-lasting savings effect. (2)emissions accounting and forecasting,as used for greenhouse gas Such gross market savings sometimes do not include temporary mitigation goals. reductions in energy use from changes in weather,income,energy Figure B.1 compares the top-down with the bottom-up impact prices,and other structural economic changes,such as in industry evaluation approaches that are discussed in the body of this guide. composition.Figure B.2 shows a graphical illustration of the concept At present,virtually all energy efficiency program evaluations behind estimating gross market savings.110 conducted in the United States rely on bottom-up approaches. During the last two decades,many energy efficiency practitioners Top-down approaches start from aggregate data,such as state-level and policymakers have expressed growing interest in the use of data for energy consumption,and then attempt to correlate any top-down methods for documenting the system-wide impacts and changes in energy consumption with measures of energy efficiency gross market savings of energy efficiency initiatives. Interest in actions,such as expenditures or savings,using macro-economic top-down methods has grown from policymakers'and evaluation models.The main advantages of top-down evaluation methods over researchers'concerns that bottom-up evaluations have not properly bottom-up methods are their potentially lower evaluation costs due accounted for effects of free ridership,spillover,and energy effi- to relatively modest data requirements and the potential for directly ciency measure interactions—particularly in large program portfolios and in situations where energy consumer funds are used,such as in B-9 www.seeaction.energy.gov December 2012 utility-sponsored efficiency programs.Top-down evaluations should FIGURE B.1: Comparison of bottom-up versus also be less expensive to implement than bottom-up evaluations. top-down evaluation Thus,research on top-down evaluation has been directed largely toward estimation of energy consumer-funded energy efficiency program savings,both the gross market energy savings and the Cl) Cl) N portion attributable to the programs being evaluated. J Q .I. 1 J Top-down energy efficiency evaluation methods are generally Z Z less developed than bottom-up methods in the energy efficiency Z Q field. However,at time of the publication of this guide,two macro- d consumption(top-down)pilot studies are under way.Sponsored by 0 the California Public Utilities Commission(CPUC),these studies are d O attempts to test the effectiveness and reliability of different top- O ~ H down evaluation approaches and to determine whether they can ~ , , Om be applied consistently to project market gross savings and to attribute savings to utility-sponsored energy efficiency investments. The results of these pilots can be found at the CALMAC website (www.calmac.org)and the CPUC website(www.cpuc.ca.gov/PUC/ energy/Energy+Efficiency). FIGURE B.2: Graphical illustration of estimation of market gross savings 10,000 9,000 Observed consumption 8,000 Counterfactual 7,000 consumption GWh 6,000 5,000 4,000 3,000 1987 1989 1991 1993 1995 1997 1999 2001 2003 2005 2007 2009 Source:Stewart J.;Haeri,M.H.(July 2011).Critical Review and Recommendations on Top-Down Evaluation.White paper.The Cadmus Group,Inc.Prepared for the California Public Utilities Commission(CPUQ. December 2012 www.seeaction.energy.gov B-10 Appendix B: Notes 105 Eto,J.; Prahl, R.;Schegel,J. (1996).A Scoping Study on Energy- Efficiency Market Transformation by California Utility DSM Programs. Lawrence Berkeley National Laboratory.http://eetd.lbl.gov/ea/ems/ reports/39058.pdf. 106 Rosenberg,M.; Hoefgen,L.(March 2009).Market Effects and Market Transformation:Their Role in Energy Efficiency Program Design and Evaluation. Prepared for California Institute for Energy and Environment(CIEE)by KEMA,Inc.; Nexus Market Research. http://uc-ciee.org/downloads/mrkt_effts_wp.pdf. 107 Much of this subsection is taken(in some cases word for word)from the National Action Plan for Energy Efficiency. (2008). Understanding Cost-Effectiveness of Energy Efficiency Programs:Best Practices, Technical Methods,and Emerging Issues for Policy-Makers. Prepared by Energy and Environmental Economics, Inc. (E3)and Regulatory Assistance Project(RAP).www.epa.gov/cleanenergy/ documents/suca/cost-effectiveness.pdf. 108 Another recent summary report of cost-effectiveness tests is: Woolf,T.;Malone, E.;Takahashi, K.;Steinhurst,W.(July 23,2012). Best Practices in Energy Efficiency Program Screening:How to Ensure that the Value of Energy Efficiency is Properly Accounted For. Synapse Energy Economics,Inc. Prepared for National Home Performance Council.www.synapse-energy.com/Downloads/ SynapseReport.2012-07.N HPC.EE-Program-Screening.12-040.pdf. 109 California Public Utilities Commission(CPUC).(2001).California Standard Practice Manual:Economic Analysis of Demand-Side Programs and Projects.www.energy.ca.gov/greenbuilding/docu- ments/background/07-J_CPUC_STAN DARD_PRACTICE_MAN UAL. PDF.See also www.cpuc.Ca.gov/PUC/energy/Energy+Efficiency/ EM+and+V for the 2007 SPM Clarification Memo. 110 Stewart,J.; Haeri,M.H.(July 2011).Critical Review and Recommendations on Top-Down Evaluation.White paper.The Cadmus Group,Inc. Prepared for the California Public Utilities Commission(CPUC). B-11 www.seeaction.energy.gov December 2012 Appendix C: Resources This appendix provides a listing of references that provide a body Consortium for Energy Efficiency EM&V Resources:www.ceel.org/ of knowledge developed over the last several decades of energy eval/eval-res.php3. efficiency program implementation and evaluation.They can be considered the current primary resources for energy efficiency Efficiency Valuation Organization(EVO)International Performance program impact evaluation and project measurement and verifica- Measurement and Verification Protocol(IPMVP):www.evo-world.org. tion(M&V),and thus the basis for the definitions,approaches,and European Union Energy Efficiency Directive,measurement, issues adopted and explained in this guide. In addition,throughout monitoring,and evaluation:www.evaluate-energy-savings.eu/ the guide there are numerous documents referenced in endnotes— emeees/en/home/index.php. see individual guide sections for these references,as well as the references section that follows this appendix. Federal Energy Management Program M&V:http://ateam.lbl.gov/mv. C.1 BACKGROUND International Energy Program Evaluation Conference:www.iepec.org/ The information in this guide is documented in numerous guides, IEPECHome.htm?programsabstracts.htm. protocols,papers,and reports. From a historical perspective,many Northeast Energy Efficiency Partnerships EM&V Forum: of the basic references on energy and energy efficiency impact http://neep.org/emv-forum. evaluations were written in the 1980s and 1990s. Unfortunately, most of the early reference documents are not easily available to the Northwest Energy Efficiency Alliance evaluation:www.nwalliance.org/ general public(i.e.,they are not posted on the Web).However,here resea rch/eva I uation reports.aspx. are three reference documents in the public domain that provide a State and Local Energy Efficiency Action Network Resources: historical perspective and solid fundamentals: www.seeaction.energy.gov/resources.html. • Hirst, E.; Reed,J.eds. (1991).Handbook of Evaluation of Utility DSM Programs.ORNL/CON-336. Prepared for Oak Ridge U.S. Environmental Protection Agency EM&V Webinar Series: National Laboratory. www.emvwebinar.org. • Khawaja, M.S.; Mulholland,C.;Thayer,J.;Smith, K.(2008). C.2.2 Select Impact Evaluation Resources Guidebook for Energy Efficiency Program Evaluation California Institute for Energy and Environment.(2009).Energy Measurement& Verification. 1016083. Palo Alto,CA: Electric Efficiency O Evaluation Training Power Research Institute(EPRI). Opportunities.http://uc-ciee.org/ planning-evol uation/7/342/105/nested. • Violette, D. M.(1995).Evaluation, Verification, and Performance Measurement of Energy Efficiency Programmes. California Public Utilities Commission.(2001).California Standard Prepared for International Energy Agency. Practice Manual:Economic Analysis of Demand-Side Programs and Projects.www.energy.ca.gov/greenbuilding/documents/ RE PRIMARY IMPACT EVALUATION background/07-J_CPUC_STANDARD_PRACTICE_MANUAL.PDF. RESOURCES See also www.cpuc.ca.gov/PUC/energy/Energy+Efficiency/ C.2.1 General Efficiency Evaluation Resource EM+and+V for the 2007SPM Clarification Memo. Websites American Council for an Energy-Efficient Economy California Public Utilities Commission.(2006).California Energy (ACEEE)publications:www.aceee.org/publications. Efficiency Evaluation Protocols:Technical,Methodological,and Reporting Requirements for Evaluation Professionals. Prepared California Institute of Energy and Environment, Planning and by The TecMarket Works Team.www.calmac.org/events/ Evaluation Library:http://uc-ciee.org/planning-evaluation/7/ Eva IuatorsProtocols_Final_AdoptedviaRuIing_06-19-2006.pdf. Ibrsearch. California Measurement Advisory Council(CALMAC): California Public Utilities Commission.(2004).California Evaluation www.calmac.org. Framework(2004).Prepared by The TecMarket Works Team.www.calmac.org/publications/California_Evaluation_ Framework_J une_2004.pdf. December 2012 www.seeaction.energy.gov C-1 Messenger, M.; Bharvirkar, R.;Golemboski,B.;Goldman,C.A.; C,2,3 U.S. Environmental Protection Agency Schiller,S.R.(April 2010).Review of Evaluation,Measurement and Non-Energy Benefits and Avoided Emissions Verification Approaches Used to Estimate the Load Impacts and Calculation References Effectiveness of Energy Efficiency Programs. Lawrence Berkeley Assessing the Multiple Benefits of Clean Energy National Laboratory. Report LBNL-3277E.http://emp.lbl.gov/ This document helps state energy,environmental,and economic sites/all/files/Ibnl-3277e.pdf. policymakers identify and quantify the benefits of clean energy, including the energy,environmental—specifically greenhouse gas, State and Local Energy Efficiency Action Network.(April 2011). air,and health—and economic benefits of clean energy.It provides National Energy Efficiency Evaluation,Measurement,and an analytical framework that states can use to estimate those Verification(EM&V)Standard:Scoping Study of Issues and benefits during the development,implementation,and evaluation of Implementation Requirements. Prepared by Schiller,S.R.; clean energy policies and programs. http://epa.gov/statelocalclimate/ Goldman,C.A.;Galawish,E.; LBNL Environmental Energy resources/benefits.html. Technologies Division.wwwl.eere.energy.gov/seeaction/pdfs/ emvstandard_scopingstudy.pdf. Co-Benefits Risk Assessment(COBRA)Screening Model EPA's Co-Benefits Risk Assessment(COBRA)screening model is a free State and Local Energy Efficiency Action Network.(June 2011). tool that helps state and local governments estimate and map the Scoping Study to Evaluate Feasibility of National Databases for air quality, human health,and related economic benefits(excluding EM&V Documents and Measure Savings. Prepared by Jayaweera, energy cost savings)of clean energy policies or programs. T.; Haeri, H.; Lee,A.; Bergen,S.; Kan,C.;Velonis,A.;Gurin,C.; Visser,M.;Grant,A.; Buckman,A.;The Cadmus Group Inc. Roadmap for Incorporating Energy Efficiency/Renewable Energy wwwl.eere.energy.gov/seeaction/pdfs/emvscoping_ Policies and Programs into State and Tribal Implementation Plans data basefeasibility.pdf. The EPA Roadmap reduces the barriers for state,tribal,and local agencies to incorporate energy efficiency and renewable energy(EE/ State and Local Energy Efficiency Action Network.(May 2012). RE)policies and programs in state and tribal implementation plans Evaluation,Measurement,and Verification(EM&V)of Residential by clarifying existing EPA guidance and providing new and detailed Behavior-Based Energy Efficiency Programs:Issues and information.The goal of this document is to facilitate the use of Recommendations. Prepared by Todd,A.;Stuart,E.;Schiller,S.; energy efficiency and renewable energy emissions reduction strate- Goldman,C.;Lawrence Berkeley National Laboratory. gies in air quality plans.www.epa.gov/airquality/eere/manual.html. wwwl.eere.energy.gov/seeaction/pdfs/emv_behaviorbased_ eeprograms.pdf. Projected Impacts of Existing State Energy Efficiency and Renewable Energy Policies U.S.Department of Energy.(2006).EERE Guide for Managing General These EPA methods and projected energy impacts may be useful to Program Evaluation Studies. Prepared by H. Barnes,Lockheed states preparing State Implementation Plan(SIP)submittals to meet Martin Aspen;Gretchen Jordan,Sandia National Laboratories. the National Ambient Air Quality Standards(NAAQS)for ozone and wwwl.eere.energy.gov/analysis/pdfs/evl_mg_app.pdf. particulate matter.www.epa.gov/statelocalclimate/state/ statepolicies.html. U.S. Department of Energy.(2007).Impact Evaluation Framework for Technology Deployment Programs. Prepared by J.Reed, Emissions&Generation Resource Integrated Database(eGRID) Innovologie,LLC;G.Jordan,Sandia National Laboratories;E.Vine, eGRID is a comprehensive source of data on the environmental Lawrence Berkeley National Laboratory.wwwl.eere.energy.gov/ characteristics of almost all electric power generated in the United analysis/pdfs/impact_framework_tech_deploy_2007_overview.pdf. States.These include air emissions of nitrogen oxides,sulfur dioxide,carbon dioxide,and many other attributes.www.epa.gov/ U.S. Environmental Protection Agency.(2012). The Roadmap for cleanenergy/energy-resources/egrid/index.html. Incorporating Energy Efficiency/Renewable Energy Policies and Programs into State and Tribal Implementation Plans. (See Appendix I of the roadmap).www.epa.gov/airquality/eere. C-2 www.seeaction.energy.gov December 2012 Power Plant Emissions Calculator(P-PEC) Hourly Marginal Emissions Tool P-PEC is a simplified tool that uses eGRID"non-baseload"emissions This tool is a statistical dispatch simulator that predicts the hourly factors to help states quickly estimate the magnitude of emission changes in generation and air emissions at electric generating units reductions from EE/RE for each power plant within a region,and (EGUs)resulting from EE/RE policies and programs.Contact Robyn to understand potential emission reductions within a county or DeYoung at EPA:deyoung.robyn@epa.gov air quality nonattainment area.Contact Robyn DeYoung at EPA: deyoung.robyn@epa.gov C.2.4 Technical Reference Manual (TRMs) Resources TABLE CA: United States Technical Reference Manuals RESOURCE NAME INFORMATION® ADMINISTRATORINCLUDED www.energystar.gov/index. Online Ex ante savings National ENERGY STAR° cfm?c=products. Calculators based on Agency pr_find_es_products algorithms Regional Technical Ex ante savings Regional— www.nwcouncil.org/energy/rtf/ Online Advisory Forum(RTF)Deemed based on Northwest Measures measures/Default.asp Database algorithms Committee http://neep.org/uploads/ Regional— EMV%20Forum/EMV%2OProducts/ Algorithms and Nonprofit Mid-Atlantic Mid Atlantic TRIM PDF Mid%20Atlantic%20TRM_ ex ante savings Organization V1%202_FINAL.pdf Arkansas Deemed www.aepefficiency.com/ Arkansas Savings Quick Start oklahoma/ci/downloads/ PDF Algorithms Public Utility Programs Deemed_Savings_Report.pdf DEER Database for Software California Energy-Efficient www.deeresources.com Ex ante savings State Commission resources Program Connecticut Light& Power and United www.ctenergyinfo.com/2012 CT Connecticut Illuminating Company Program Savings Documentation PDF Algorithms and ex public Utility Program Savings FINAL.pdf ante savings Documentation Hawaii Energy Efficiency www.hawaiienergy.com/media/ Algorithms and ex Hawaii assets/AnnualReportAttachments- PDF State Commission Program TRIMALL.pdf ante savings www.efficiencymaine.com/docs/ Efficiency Maine TRIM board_meeting_documents/ PDF Algorithms and Trust —Commercial Maine-Commercial-TRM-8-31- ex ante savings 2010-Final.pdf Maine www.efficiencymaine.com/docs/ Efficiency Maine TRIM board_meeting_documents/ PDF —Residential Maine-Residential-TRM-02-04-09. pdf Massachusetts www.ma-eeac.org/docs/ Statewide TRIM for 2012 MTM Files/Oct 28,2011 DPU Algorithms and Massachusetts Estimating Filings/WMECo/13-WMECO PDF ex ante savings Agency Savings from Energy Exhibit H MA TRM_2012 PLAN_ Efficiency Measures FINAL.pdf December 2012 www.seeaction.energy.gov C-3 INFORMATION 7Mric 0 . . - •- INCLUDEDwww.michigan.gov/mpsc/Michigan Energy Excel Measures Database Database an 0,1607,7-159-52495_ Ex ante savings State Commission 55129---,00.html New Jersey Clean Energy www.njcleanenergy.com/files/ Algorithms and New Jersey Program Protocols to file/Library/NJ Protocols Revisions PDF Agency Measure Resource Savings 7-21-11_Clean.pdf ex ante savings New York Standard Approach for Estimating New York Energy Savings from http://efile.mpsc.state.mi.us/efile/ PDF Algorithms and Agency docs/16671/0026.pdf ex ante savings Energy Efficiency Programs Ohio Ohio TRM Online Algorithms and www.ohiotrm.org State Commission Database ex ante savings www.puc.state.pa.us/electric/ Algorithms and Pennsylvania Pennsylvania TRM DOC State Commission Actl29/TRM.aspx ex ante savings Deemed Savings, www.entergy-texas.com/content/ Energy_Efficiency/documents/ Algorithms and Texas Installation,and PDF State Commission Efficiency Standards Deemed— Savings_Measures_ ex ante savings List.pdf Efficiency Vermont www.veic.org/Libraries/Resumes/ Algorithms and Nonprofit Vermont Technical Reference User PDF Manual Tech Manua lEVT.sflb.ashx ex ante savings Organization Focus on Energy www.focusonenergy.com/files/ Evaluation Business Document—Management_System/ Algorithms and Wisconsin PDF State Commission Programs: Deemed Evaluation/bpdeemedsavings- ex ante savings Savings Manual V1.0 manuavl0_evaluationreport.pdf Source:State and Local Energy Efficiency Action Network.(June 2011).Scoping Study to Evaluate Feasibility of National Databases for EM&V Documents and Measure Savings. Prepared byJayaweera,T.;Hoed,H.;Lee,A.;Bergen,S.;Kan,C.;Velonis,A.;Gunn,C.;Visser,M.;Grant A.;Buckman,A.;The Cadmus Group Inc. wwwl.eere.energy.govlseeaction/pdfs/emvscoping databasefeasibility.pdf.Tables are current as of Summer 2012. C-4 www.seeaction.energy.gov December 2012 C.3 MEASUREMENT AND VERIFICATION ASHRAE Performance Measurement Protocols(PMP)for RESOURCES Commercial Buildings.These provide a standardized,con- Several M&V resource documents used in the development of this sistent set of protocols for facilitating the comparison of guide are available via the internet and are presented in this section; measured energy,water,and indoor quality performance of they can be considered the current primary resources for energy commercial buildings.www.ashrae.org. efficiency project M&V. • International Standards Organization(ISO).The International • International Performance Measurement and Verification Standards Organization entered the world of energy manage- Protocol(IPMVP).The IPMVP provides an overview of current ment with the release of ISO 50001-Energy Management best practices for verifying results of energy efficiency,water, Systems.This protocol includes the following: and renewable energy projects in commercial and industrial — ISO 50001:2011 specifies requirements for establishing, facilities. Internationally, it is the most recognized M&V implementing, maintaining,and improving an energy protocol for demand-side energy activities.The IPMVP was management system,the purpose of which is to enable an developed with DOE sponsorship and is currently managed organization to follow a systematic approach in achieving by the nonprofit Efficiency Valuation Organization,which continual improvement of energy performance,including continually maintains and updates the IPMVP. energy efficiency,energy use,and consumption. The IPMVP provides a framework and definitions that can — ISO 50001:2011 specifies requirements applicable to help practitioners develop M&V plans for their projects. It energy use and consumption,including measurement, includes guidance on best practices for determining savings documentation and reporting,design,and procurement from efficiency projects.It is not a"cookbook"of how to practices for equipment,systems, processes,and personnel perform M&V for specific projects;rather,it provides guid- that contribute to energy performance. ance and key concepts that are used in the United States and — ISO 50001:2011 applies to all variables affecting energy internationally.The IPMVP is probably best known for defining performance that can be monitored and influenced by the four M&V options for energy efficiency projects.These options organization. ISO 50001:2011 does not prescribe specific (A,B,C,and D),presented in Chapter 4,differentiate the most performance criteria with respect to energy. common approaches for M&V. Reference:Efficiency Valuation — Subsequent to the release of 50001, ISO created a Organization(2010).International Performance Measurement technical committee(TC242)to support the deployment of and Verification Protocol.www.evo-world.org. ISO 50001. In January 2011,the ISO Technical Management • FEMP M&V Guidelines.The purpose of this document is to Board announced the creation of another committee to provide guidelines and methods for documenting and verify- create standards on measuring and verifying savings—TC ing the savings associated with federal agency performance 257 Energy Savings.Any standards originating from TC 242, contracts. It contains procedures and guidelines for quantifying TC 257,or the joint working group are expected to provide the savings resulting from energy efficiency equipment,water broad guidance on M&V principles in 2012 or 2013. conservation,improved operations and maintenance, renew- able energy,and cogeneration projects.Along with the FEMP M&V Guidelines are several useful companion documents. http://mnv.lbl.gov/keyMnVDocs. • ASHRAE Guideline 14 Measurement of Energy and Demand Savings.ASHRAE is the professional engineering society that has been the most involved in writing guidelines and standards associated with energy efficiency.Compared with the FEMP M&V Guidelines and the IPMVP,Guideline 14 is a more detailed technical document that addresses the analyses,statistics,and physical measurement of energy use for determining energy savings. Reference:American Society of Heating,Refrigerating, and Air-Conditioning Engineers(2002).A new version is expected to be available in 2013.www.ashrae.org. December 2012 www.seeaction.energy.gov C-5 References "Acid Rain Program."(2012).U.S.Environmental Protection Agency. California Public Utilities Commission(CPUC).(2007).Guidelines for www.epa.gov/airmarkets/progsregs/arp. Estimating Net-To-Gross Ratios Using the Self-Report Approaches. CPUC Energy Division Master Evaluation Contractor Team. "Air Markets Program Data."(2012).U.S. Environmental Protection ftp://ftp.cpuc.ca.gov/puc/energy/electric/energy+efficiency/ Agency. http://ampd.epa.gov/ampd/. ee+workshops/self repo rtguidelinesd eta iled_v20.pdf. American Council for an Energy-Efficient Economy(ACEEE). California Public Utilities Commission(CPUC).(April 9,2010). (October 2011).State Energy Efficiency Resource Standard(EERS) Codes&Standards(C&S)Programs Impact Evaluation: Activity. http://aceee.org/files/pdf/policy-brief/.State EERS California Investor Owned Utilities'Codes and Standards Program Summary October 2011.pdf. Evaluation for Program Years 2006-2008. CALMAC Study ID: CP00030.06 Final Evaluation Report.Prepared by KEMA, Inc.; American Society of Heating, Refrigerating and Air-Conditioning The Cadmus Group,Inc.; Itron, Inc.;Nexus Market Research, Engineers(ASHRAE).(2002).Guideline 14:Measurement of Inc.;and ENRG,Inc.for the CPUC Energy Division. Demand and Energy Savings.www.ashrae.org. www.calmac.org/publications/Codes_Standards—Vol-111— Fi na I Eva I uation ReportU pdated-04122010.pdf. American Society of Heating, Refrigerating and Air-Conditioning Engineers(ASHRAE).(2010).Performance Measurement "Clean Air Interstate Rule."(2005). U.S. Environmental Protection Protocols for Commercial Buildings.www.ashrae.org/ Agency.www.epa.gov/cair. standards-research--technology/special--project-activities. "Compilation of Air Pollutant Emission Factors."(2011). U.S. Barbose,G.; Billingsley; M.;Goldman,C.;Hoffman,I.;Schlegel, Environmental Protection Agency. EPA AP-42.www.epa.gov/ttn/ J.(August 2012)."On a Rising Tide:The Future of U.S. Utility chief/ap42. Customer-Funded Energy Efficiency Programs."2012 ACEEE Summer Study Proceedings,August 12-17,2012, Pacific Grove, Consensus Report on the EmPower Maryland EM&V Process.(June California.Washington,D.C.:American Council for an Energy- 24,2009).www.raponline.org/document/download/id/551. Efficient Economy(ACEEE).LBNL-5755E. http://www.aceee.org/ files/proceedings/2012/data/papers/0193-000173.pdf. Efficiency Valuation Organization(EVO). International Performance Measurement and Verification Protocol(IPMVP).(multiple "Building Energy Codes Program."(2012). U.S. Department of Energy. dates).www.evo-world.org. www.energycodes.gov. "eGRID."(2012). U.S.Environmental Protection Agency. "California Greenhouse Gas(GHG)Cap-and-Trade Program."(2012). www.epa.gov/cleanenergy/energy-resources/egrid/index.html. California Air Resources Board(CARB).www.arb.ca.gov/cc/ capandtrade/capandtrade.htm. Energy&Resources Solutions,Inc.(ERS);Left Fork Energy;West Hill Energy&Computing; Megdal&Associates,LLC.; California Public Utilities Commission(CPUC).(2001).California Opinion Dynamics Corporation.(June 2010).Report Standard Practice Manual:Economic Analysis of Demand-Side on Impact Assessments of Two NYSERDA Natural Gas Programs and Projects.www.energy.ca.gov/greenbuilding/ Efficiency Programs:Con Edison Natural Gas Efficiency documents/background/07-J_CPUC_STANDARD_PRACTICE_ Program&National Grid Low Income Natural Gas MANUAL.PDF.See also http://www.cpuc.ca.gov/PUC/energy/ Efficiency Program. Prepared for the New York State Energy+Efficiency/EM+and+V/. Energy Research and Development Authority(NYSERDA). http://documents.dps.ny.gov/public/Common/ViewDoc. California Public Utilities Commission(CPUC).(April 2006). aspx?DocRefld={CFCD7DE1-8B68-46CC-B64B-83ACD072A4FD). California Energy Efficiency Evaluation Protocols:Technical, Methodological,and Reporting Requirements for Evaluation Eto,J.; Prahl,R.;Schegel,J.(1996).A Scoping Study on Energy- Professionals.www.calmac.org/events/EvaluatorsProtocols_ Efficiency Market Transformation by California Utility DSM Final_AdoptedviaRuling_06-19-2006.pdf. Programs.Lawrence Berkeley National Laboratory. http://eetd.lbl.gov/ea/ems/reports/39058.pdf. D-1 www.seeaction.energy.gov December 2012 Federal Energy Management Program(FEMP).(April 2008).M&V Intergovernmental Panel on Climate Change(IPCC).(2007). Guidelines:Measurement and Verification for Federal Energy Pathways to a Low-Carbon Economy:Version 2 of the Global Projects.wwwl.eere.energy.gov/femp/pdfs/mv_guidelines.pdf. Greenhouse Gas Abatement Cost Curve. Prepared by McKinsey &Company.www.epa.gov/statelocalclimate/documents/pdf/ Federal Energy Regulatory Commission(FERC).(June 2010). mckinsey_summary_11-19-09.pdf. National Action Plan on Demand Response. Docket No.AD09-10. Prepared by FERC staff.www.ferc.gov/legal/staff-reports/ Kushler,M.; Nowak,S.;Witte,P.(February 2012).A National 06-17-10-demand-response.pdf. Survey of State Policies and Practices for the Evaluation of Ratepayer-Funded Energy Efficiency Programs.American "Global Warming Solutions Act:Assembly Bill 32."(2006).California Council for an Energy-Efficient Economy(ACEEE). Report Air Resources Board(CARB).www.arb.ca.gov/cc/ab32/ab32.htm. Number U122.www.aceee.org/research-report/ul22. Gottstein,M.;Schwartz,L.(2010).The Role of Forward Capacity Lazar,J.;Baldwin,X.(August 2011). Valuing the Contribution of Markets in Increasing Demand-Side and Other Low-Carbon Energy Efficiency to Avoided Marginal Line Losses and Reserve Resources Experience and Prospects. Regulatory Assistance Requirements. Regulatory Assistance Project(RAP). Project(RAP).www.raponline.org/docs/RAP_Gottstein_ www.raponline.org/document/download/id/4537. Schwa rtz_RoleofFCM_Experienceand Prospects2_2010_05_04.pdf. Mahone,D.; Hall,N.(August 2010)."Pursuit of Aggressive Energy Haeri, H.; Khawaja,M.S.(March 2012)."The Trouble With Savings Targets—Aligning Program Design and Evaluation." Freeriders."Public Utilities Fortnightly.www.fortnightly.com/ 2010 ACEEE Summer Study Proceedings;August 15-20,2010, fortnightly/2012/03/trouble-freeriders. Pacific Grove,California.Washington, D.C.:American Council for an Energy-Efficient Economy. http://neep.org/uploads/ Hibbard,P.;Tierney,S.;Okie,A.; Darling,P.(November 2011). The EMV%20Forum/Steering%20Committee%20Notes/DMahone_ Economic Impacts of the Regional Greenhouse Gas Initiative on ACEEE_Paper_for_SC_Notes.pdf. Ten Northeast and Mid-Atlantic States,Review of the Use of RGGI Auction Proceeds from the First Three-Year Compliance Period. Mahone,D.(August 2011)."Free-Ridership as a Way to Kill The Analysis Group.www.analysisgroup.com/uploadedfiles/ Programs—How Evaluation Policies Can Frustrate Efficiency publishing/articles/economic_impact_rggi_report.pdf. Goals." Boston:International Energy Program Evaluation Conference.www.iepec.org/20llPapersTOC/papers/054.pdf. Hirst,E.;White, D.(1985).Indoor Temperature Changes After Retrofit:Inferences Based on Electricity Billing Data for Messenger,M.; Bharvirkar, R.;Golemboski,B.;Goldman,C.A.; Nonparticipants and Participants in the BPA Residential Schiller,S.R.(April 2010).Review of Evaluation,Measurement Weatherization Program.Oak Ridge National Laboratory. and Verification Approaches Used to Estimate the Load Impacts www.ornl.gov/info/reports/1985/3445600444954.pdf. and Effectiveness of Energy Efficiency Programs. Lawrence Berkeley National Laboratory. Report LBNL-3277E. Intergovernmental Panel on Climate Change(IPCC).(2006).2006 http://emp.lbl.gov/sites/all/files/Ibnl-3277e.pdf. Guidelines for National Greenhouse Gas Inventories. www.ipcc-nggip.iges.or.jp/public/2006gl/index.html. Mid-Atlantic Regional Air Management Association(MARAMA). www.marama.org. Intergovernmental Panel on Climate Change(IPCC).(2007). Climate Change 2007:Synthesis Report.www.ipcc.ch/pdf/ National Action Plan for Energy Efficiency.(2007).Guide to Resource assessment-report/ar4/syr/ar4_syr.pdf. Planning with Energy Efficiency. Prepared by Snuller Price et al., Energy and Environmental Economics,Inc.(E3).www.epa.gov/ Intergovernmental Panel on Climate Change(IPCC).(2007). cleanenergy/documents/suca/resource_planning.pdf. Contribution of Working Group III to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change,2007. National Action Plan for Energy Efficiency.(2007).Model Energy www.ipcc.ch/publications_and_data/ar4/wg3/en/contents.html. Efficiency Program Impact Evaluation Guide. Prepared by Steven R.Schiller,Schiller Consulting,Inc.www.epa.gov/eeactionplan. December 2012 www.seeaction.energy.gov D-2 National Action Plan for Energy Efficiency.(2008). Understanding Schiller,S.R.(August 2011). "Documenting Greenhouse Gas Emission Cost-Effectiveness of Energy Efficiency Programs:Best Practices, Reductions from End-Use Efficiency Activities in the Kyoto Technical Methods,and Emerging Issues for Policy-Makers. Protocol's Clean Development Mechanism Offset Program." Prepared by Energy and Environmental Economics,Inc.(E3); 2011 International Energy Program Evaluation Conference Regulatory Assistance Project(RAP).www.epa.gov/cleanenergy/ Proceedings,August 16-18,2011,Boston,Massachusetts. documents/suca/cost-effectiveness.pdf. Madison,Wisconsin: International Energy Program Evaluation Conference(IEPEC).www.iepec.org/20llPapersTOC/papers/ NMR Group, Inc.; Research Into Action,Inc.(November 2010)."Net 115.pdf#page=l. Savings Scoping Paper."Submitted to Northeast Energy Efficiency Partnerships:Evaluation, Measurement,and Verification Forum. Schiller,S.R.;Goldman,C.A.(August 2011)."Developing State http://neep.org/uploads/EMV%20Forum/EMV%20Products/ and National Evaluation Infrastructures—Guidance for the FINAL%20Net%20Savings%20Scoping%20Paper%2011-13-10.pdf. Challenges and Opportunities of EM&V."2011 International Energy Program Evaluation Conference Proceedings,August Nowak,S.;Kushler,M.;Sciortino,M.;York,D.;Witte,R(June 2011). 16-18,2011, Boston,Massachusetts. Madison,Wisconsin: Energy Efficiency Resource Standards:State and Utility Strategies International Energy Program Evaluation Conference(IEPEC). for Higher Energy Savings.American Council for an Energy-Efficient www.iepec.org/20llPapersTOC/papers/073.pdf#page=l. Economy(ACEEE).Report Number U113.www.aceee.org/ research-report/u113. Schiller,S.R.; Prindle, B.;Cowart, R.;Rosenfeld,A.H.(2008)."Energy Efficiency and Climate Change Mitigation Policy."2008ACEEE Peters,J.;McRae,M.(August.2008)."Free-Ridership Measurement Summer Study Proceedings.August 17-22,2008,Pacific Grove, Is Out of Sync With Program Logic...or,We've Got the California.Washington, DC:American Council for an Energy- Structure Built, but What's Its Foundation?"2008ACEEE Efficient Economy(ACEEE).www.aceee.org/files/proceed- Summer5tudyProceedings.August 17-22,2008,Pacific ings/2008/data/papers/8_306.pdf. Grove,California.Washington, DC:American Council for an Energy-Efficient Economy(ACEEE).www.aceee.org/files/ Skumatz, L.A.(1997)."Recognizing All Program Benefits: Estimating proceed ings/2008/data/papers/5_491.pdf. the Non-Energy Benefits of PG&E's Venture Partners Pilot Program(VPP)."Proceedings of the 1997 Energy Evaluation "Promoting Air Quality and Climate Protection with Clean Energy Conference.August 27-29,1997,Chicago,Illinois.Madison, and Energy Efficiency in the Metropolitan Washington Area." Wisconsin: International Energy Program Evaluation Conference Metropolitan Washington Council of Governments(MWCOG). (IEPEC).www.iepec.org/1997PapersTOC/papers/033.pdf. www.mwcog.org/environment/air/EERE/default.asp. Skumatz, L.A.; Khawaya,M.;Colby,J.(2009).Lessons Learned and Rathbun,P.;Sabo,C.;Zent, B.(2003).Standardized Methods for Next Steps in Energy Efficiency Measurement and Attribution: Free-Ridership and Spillover Evaluation—Task 5 Final Report Energy Savings,Net to Gross,Non-Energy Benefits,and (Revised). Prepared for National Grid, NSTAR Electric, Northeast Persistence of Energy Efficiency Behavior. Prepared for the Utilities, Unitil,and Cape Light Compact.www.ceel.org/eval/ California Institute for Energy and Environment(CIEE)Behavior db_pdf/297.pdf. and Energy Program. http://uc-ciee.org/downloads/EEM_A.pdf. "Regional Greenhouse Gas Initiative."(2012).http://rggi.org/. Skumatz, L.A.; Khawaja,M.S.; Krop, R.(May 2010).Non-Energy Benefits:Status,Findings,Next Steps,and Implications for Low Rosenberg, M.;Hoefgen,L.(March 2009).Market Effects and Market Income Program Analyses in California. Revised Report.Prepared Transformation:Their Role in Energy Efficiency Program Design for Brenda Gettig,Sempra Utilities.www.liob.org/docs/LIEE Non- and Evaluation. Prepared for California Institute for Energy and Energy Benefits Revised Report.pdf. Environment(CIEE)by KEMA, Inc.;Nexus Market Research. http://uc-ciee.org/downloads/mrkt_effts—wp.pdf. D-3 www.seeaction.energy.gov December 2012 State and Local Energy Efficiency Action Network.(April 2011). U.S. Environmental Protection Agency.(August 2011).Federal National Energy Efficiency Evaluation,Measurement and Implementation Plans:Interstate Transport of Fine Particulate Verification(EM&V)Standard:Scoping Study of Issues and Matter and Ozone and Correction of SIP Approvals.www.gao.gov/ Implementation Requirements. Prepared by Schiller,S.R.; prod ucts/C00127.40 CFR Parts 51,52,72,78,and 97[EPA— Goldman,C.A.;Galawish,E.;LBNL Environmental Energy HQ—OAR-2009-0491;FRL-9436-8] RIN 2060—AP50.ACTION: Technologies Division.wwwl.eere.energy.gov/seeaction/pdfs/ Final rule. emvsta nda rd_scopingstudy.pdf. Vine, E.; Hall,N.;Keating, K.M.;Kushler,M.; Prahl, R.(August 2011). State and Local Energy Efficiency Action Network.(June 2011). "Emerging Evaluation Issues Revisited."2011 International Energy Scoping Study to Evaluate Feasibility of National Databases for Program Evaluation Conference Proceedings,August 16-18, EM&V Documents and Measure Savings. Prepared by Jayaweera, 2011,Boston,Massachusetts. Madison,Wisconsin:International T.;Haeri, H.;Lee,A.; Bergen,S.;Kan,C.;Velonis,A.;Gurin,C.; Energy Program Evaluation Conference(IEPEC).www.iepec.org/ Visser, M.;Grant,A.;Buckman,A.;The Cadmus Group Inc. 2011PapersTOC/papers/096.pdf#page=l. wwwl.eere.energy.gov/seeaction/pdfs/emvscoping_ databasefeasibility.pdf. Wallace,R;Forster,H.J.(2012).State of the Efficiency Program Industry Budgets, Expenditures,and Impacts 2011.Consortium State and Local Energy Efficiency Action Network.(May 2012). for Energy Efficiency.www.ceel.org/files/2011 CEE Annual Evaluation,Measurement,and Verification(EM&V)of Residential Industry Report.pdf. Behavior-Based Energy Efficiency Programs:Issues and Recommendations. Prepared by Todd,A.;Stuart, E.;Schiller,S.; Woolf,T.;Malone, E.;Takahashi,K.;Steinhurst,W.(July 23,2012). Goldman,C.; Lawrence Berkeley National Laboratory. Best Practices in Energy Efficiency Program Screening:How to wwwl.eere.energy.gov/seeaction/pdfs/emv_behaviorbased_ Ensure that the Value of Energy Efficiency is Properly Accounted eeprograms.pdf. For.Synapse Energy Economics, Inc. Prepared for National Home Stewart,J.;Haeri, M.H.(July 2011).Critical Review and Performance Council.www.synapse-energy.com/Downloads/ Recommendations on Top-Down Evaluation. White paper. Syn a pse Re po rt.2012-07.N H PC.E E-Program-Sc re e n i n g.12-040.pd f. The Cadmus Group,Inc. Prepared for the California Public World Resources Institute(WRI);World Business Council for Utilities Commission(CPUC). Sustainable Development(WBCSD).(December 2005). "Uniform Methods Project."(2012). U.S. Department of Energy. GHG Protocol for Project Accounting(Project Protocol). wwwl.eere.energy.gov/deployment/ump.html. www•ghgprotocol.org/files/ghgp/ghg_project_protocol.pdf. U.S. Environmental Protection Agency.(2004).Guidance on State World Resources Institute(WRI);World Business Council for Implementation Plan(SIP)Credits for Emission Reductions from Sustainable Development(WBCSD).(August 2007).Protocol Electric-Sector Energy Efficiency and Renewable Energy Measures. Guidelines for Quantifying GHG Reductions from Grid-Connected www.epa.gov/ttn/oarpg/t1/memoranda/ereseerem_gd.pdf. Electricity Projects.www.wri.org/publication/guidelines-quantify- ing-ghg-reductions-grid-connected-electricity-projects. U.S. Environmental Protection Agency.(September 2005).(Draft). State Set-Aside Programs for Energy Efficiency and Renewable Zabin,C.(2011).California Workforce Education and Training Needs Energy Projects Under the NOx Budget Trading Program: Assessment.Center for Labor Research&Education, Donald A Review of Programs in Indiana,Maryland,Massachusetts, Vial Center on Employment in the Green Economy. Missouri,New Jersey,New York,and Ohio. http://epa.gov/ www.irle.berkeley.edu/vial. state loca Id i m ate/d ocu me nts/pdf/ee re_rpt.pdf. U.S. Environmental Protection Agency.(February 2010).Assessing the Multiple Benefits of Clean Energy:A Resource for States. www.epa.gov/statelocalclimate/resources/benefits.html. December 2012 www.seeaction.energy.gov D-4 This document was developed as a product of the State and Local Energy Efficiency Action Network(SEE Action),facilitated by the U.S. Department of Energy/U.S. Environmental Protection Agency.Content does not imply an endorsement by the individuals or organizations that are part of SEE Action working groups,or reflect the views,policies,or otherwise of the federal government. SEE Action STATE&LOCAL ENERGY EFFICIENCY ACTION NETWORK