Evaluation 101 Energy Efficiency Program Evaluation By Nick Hall - - PowerPoint PPT Presentation

evaluation 101 energy efficiency program evaluation by
SMART_READER_LITE
LIVE PREVIEW

Evaluation 101 Energy Efficiency Program Evaluation By Nick Hall - - PowerPoint PPT Presentation

Delaware Webinar Evaluation 101 Energy Efficiency Program Evaluation By Nick Hall TecMarket Works February 8, 2012 Delaware Evaluation Webinar February 2012 Page 1 of 54 Workshop Objectives Create a greater understanding of evaluation,


slide-1
SLIDE 1

Delaware Evaluation Webinar February 2012

Page 1 of 54

Delaware Webinar Evaluation 101 Energy Efficiency Program Evaluation

By Nick Hall TecMarket Works February 8, 2012

slide-2
SLIDE 2

Create a greater understanding of evaluation, evaluation issues and the evaluation process Address Delaware evaluation-related questions, issues, concerns, needs.

Workshop Objectives

slide-3
SLIDE 3

What we will cover in 3 hours

1. History of evaluation 2. Key definitions 3. Evaluation Framework (why needed) 4. Evaluation as portfolio management tool 5. General what, when and why of evaluation 6. Impact evaluation 7. What is EM&V 8. Net to Gross and Attribution 9. Process Evaluation

  • 10. Market Effects Evaluation
  • 11. Cost effectiveness
  • 12. Evaluation plans and planning

3

slide-4
SLIDE 4

Delaware Evaluation Webinar February 2012

Page 4 of 54

History of Energy Program Evaluation

  • USDOE formed in 1970s – implementing wide range of

information programs

  • Early evaluation was 100% ex ante and conducted by the

program administrators

  • These results were very unreliable – not field based, didn’t

capture the actual results

  • Solution was to create EE program evaluation building upon the

broader (non energy) evaluation field, applying those same evaluation definitions and standards to EE program evaluation.

  • Overtime the approaches have improved to specifically address

the unique issues associated with EE/RE/DR/ME

  • State approaches have evolved independently with the

introduction of utility programs – creating a need for Frameworks and protocols

slide-5
SLIDE 5

Delaware Evaluation Webinar February 2012

Page 5 of 54

Key Definitions

  • Ex ante: projected (pre-program estimated) savings to be

achieved

  • Ex post: measured (evaluated) savings achieved
  • EM&V: evaluation, measurement and verification.
  • Framework: Evaluation policy and operational

systems/structures and definitions

  • Protocol: prescribed ways of conducting evaluation efforts
  • Gross savings: unadjusted savings achieved by all

program participant for a program-covered intervention

  • Verified gross savings: savings achieved by all program

participant for a program-covered intervention adjusted to account for verified instillations

slide-6
SLIDE 6

Delaware Evaluation Webinar February 2012

Page 6 of 54

Key Definitions

  • Net savings: total savings achieved as a result of a

program or portfolio effort

  • Freeriders: participants who would have taken the same

action at the same time without the program intervention

  • Freedrivers/spillover: Non-participants that took actions as

a result of the program’s interventions but who did not participate in any of the programs offerings

  • Participant spillover: Participants who repeat the same

actions but did not receive another incentive or program service.

slide-7
SLIDE 7

Delaware Evaluation Webinar February 2012

Page 7 of 54

Evaluation Topics of Interest

  • Evaluation and its role in understanding

the adequacy of a portfolio

– Typically covered at the program level – Not typically covered at portfolio level – The New York Approach

slide-8
SLIDE 8

Delaware Evaluation Webinar February 2012

Page 8 of 54

Evaluation Topics of Interest

  • Why is a Framework Needed…

– Guides all evaluation efforts with regard to:

  • Who, When, Why, How, Under what conditions
  • Covers most everything related to evaluation
  • Makes sure everyone playing by the same rule book

– Avoids the oops factor:

slide-9
SLIDE 9

Delaware Evaluation Webinar February 2012

Page 9 of 54

Framework Topics

  • Framework can cover

– Approaches to use – Objectives and metrics on which to focus – Ethics, standards and principles – Planning and approval processes – Content roles & schedules for TRMs – Policies (baselines, net, gross, IPMVP, sampling, timing) – Data security and management – Customer contact and data collection – Planning and budgeting – Reporting and report contents – Cost effectiveness approach

slide-10
SLIDE 10

General What, When, Why of Evaluation

slide-11
SLIDE 11

What is Evaluation?

  • Evaluation is an objective systematic

process for assessing an organization’s activities in order to quantify the effectiveness, efficiency or effects of those activities for the purpose of documenting performance or making improvements.

11

slide-12
SLIDE 12

Why Evaluate?

Evaluation results can benefit stakeholders by ensuring better and more cost-effective programs!

– Ensure that the program is delivering the benefits that it was designed to produce – Unbiased independent assessment that supports regulatory process – including cost recovery, administrator compensation, etc. – Optimize energy and non-energy benefits – Provide valuable information about program operations

12

slide-13
SLIDE 13

Evaluation Types

– Process evaluation (documents and improves) – Impact evaluation (short term impacts) – Market effects (longer term impacts)

Process + Impact + Market effects = a well rounded

evaluation

13

slide-14
SLIDE 14

What are we measuring?

– Energy savings – Demand Savings – Environmental impacts – Economic impacts – Customer satisfaction – Non-energy benefits – Technology penetration – Other program specific research issues

14

slide-15
SLIDE 15

When to Evaluate

  • Early enough to be of use!

– Evaluation create a feedback loop that informs:

  • Program design
  • Program implementation.
  • But not too early!

–Process evaluation (after 6 months) –Impact evaluation

  • When there is something to structure into a plan
  • When pre-data is needed
  • Regularly within systematic process!

– The cycle is continuous – When a need is identified

15

Program Design Program Implement ation

Program Evaluation Process Evaluation

The feedback loop

slide-16
SLIDE 16

Data Collection Primary Methods

– Surveys (Phone, Mail, Internet, email) – Focus Groups – Observation Visits – Mystery Shopping – In-depth Interviews – Site Inspections – Metering

16

slide-17
SLIDE 17

Sample Design

– Strategy varies by research question and study objectives – When designing a sampling plan, consider:

  • Population size and distribution
  • Presence of the characteristic being measured

and conditions affecting that characteristic

  • Confidence level
  • Precision level
  • Coefficient of variation
  • Effect size

17

slide-18
SLIDE 18

Delaware Evaluation Webinar February 2012

Page 18 of 54

Precision and Bias

Precise Imprecise Biased/ Inaccurate Unbiased/ Accurate

slide-19
SLIDE 19

Delaware Evaluation Webinar February 2012

Page 19 of 54

Impact Evaluation, Measurement and Verification

slide-20
SLIDE 20

Delaware Evaluation Webinar February 2012

Page 20 of 54 20

The Evaluation Challenge Evaluation attempts to measure what did not happen. Measuring invisible energy!

Savings: The difference between energy use after the program and what the energy use would have been without the program

  • Not an easy question to answer; we need a
  • baseline…
  • Nutshell: Impact = Actual post – Actual pre ± Adjustment
slide-21
SLIDE 21

Delaware Evaluation Webinar February 2012

Page 21 of 54

What Do You Measure?

21 Gross & Net Energy & Demand Savings Participation and Market Effects Environmental Impacts Economic Impacts

slide-22
SLIDE 22

Delaware Evaluation Webinar February 2012

Page 22 of 54

How Do You Measure Impacts?

– Engineering calculations/algorithms – Billing analysis (utility meter) – Metered data analysis (evaluation meter) – Load shape analysis – Building energy simulation modeling

  • DOE-II

22

slide-23
SLIDE 23

Delaware Evaluation Webinar February 2012

Page 23 of 54

Engineering Approaches

– Engineering calculations use formulas or algorithms to estimate the energy use of equipment before and after installation. – These approaches are good for projects that do not have a variance in equipment use patterns – There are many on-line calculators that can be used including EIA DOE, Energy Star, and other web sites

  • Lighting equipment replacements
  • Prescriptive measures such as high efficiency

packaged air conditioning

  • Computer and plug-load savings

23

slide-24
SLIDE 24

Delaware Evaluation Webinar February 2012

Page 24 of 54

Billing Analysis

– Uses customer or facility billing data – May be simple pre-post comparison

  • If pre-measure installation data are not available,

such as for new homes, a comparison group is needed.

– May be complex statistical billing analysis

  • Including engineering estimates for installed

measures can improve the statistical billing estimates (Statistically Adjusted Engineering (SAE)

24

slide-25
SLIDE 25

Delaware Evaluation Webinar February 2012

Page 25 of 54

Billing Analysis Use

– Use billing analysis when:

  • There is a sufficient number of sample points
  • Sufficient historical data
  • The expected energy savings is “5” percent or

more of electric bill

  • There is good data on dates measures were

installed and information on specific measures

  • Billing data is relatively clean

25

slide-26
SLIDE 26

Delaware Evaluation Webinar February 2012

Page 26 of 54

Metered Data Analysis

– Metering end use loads can be the most direct and accurate method for measuring changes in energy consumption

  • Used selectively due to the cost of the

equipment and the labor to install/remove the meters

  • Not easily transferred from other climates

and service areas

26

slide-27
SLIDE 27

Delaware Evaluation Webinar February 2012

Page 27 of 54

Load Shape Analysis

– Load shape analysis may rely on secondary, as well as primary metered data, to develop end-use load shapes to estimate peak demand or energy savings – Critical to the evaluation of programs designed to reduce demand or shift loads (demand response programs)

27

slide-28
SLIDE 28

Delaware Evaluation Webinar February 2012

Page 28 of 54 28

8760 hours (annual) load example Air conditioning load

slide-29
SLIDE 29

Delaware Evaluation Webinar February 2012

Page 29 of 54

Building Simulation Tools

– Engineering estimates of savings may also be based on building simulation models or energy analysis software such as DOE-2, FEDS, EZ Sim, PowerDOE, eQuest, Trace 700, TREAT, and many others – Billing data useful to calibrate models for specific buildings/typical homes, etc.

29

slide-30
SLIDE 30

Delaware Evaluation Webinar February 2012

Page 30 of 54

So….How does Impact Evaluation Differ from Measurement and Verification?

slide-31
SLIDE 31

Delaware Evaluation Webinar February 2012

Page 31 of 54

The M is Guided by the IPMVP

– International Performance Measurement and Verification Protocol (IPMVP) – Framework to determine energy and water savings resulting from the implementation of an energy efficiency project

31

slide-32
SLIDE 32

Delaware Evaluation Webinar February 2012

Page 32 of 54

Definition of M&V

– Measurement and Verification (M&V) – The M is for measurement: the process of collecting on- site measured or observed data from an individual project or group of projects. The data is used to feed an energy impact evaluation. (a function of physics or engineering) – The V is for Verification: the process of verifying that measures are installed, configured and used in a way that reflects the assumptions of the program. The data is used to feed an energy impact evaluation. (a confirmation function) – The E in EM&V is for Evaluation, the analysis conducted to determine the amount of energy impacts. This analysis uses the results of the M&V. (an evaluation function)

32

slide-33
SLIDE 33

Delaware Evaluation Webinar February 2012

Page 33 of 54

Program Attribution (or Net to Gross)

– Program attribution refers to energy impacts that can confidently be attributed to program efforts – A net-to-gross factor that reflects program attribution is applied to gross program energy savings to get net energy savings

Evaluation Consultants

slide-34
SLIDE 34

Delaware Evaluation Webinar February 2012

Page 34 of 54

Components of NTG

  • Net Savings adjust for:

– Free riders – Participant Spillover – Market effects Net = (Gross – Freeriders + Participant Spillover + Market Effects)

Caveat: It has to be measured / documented in an independent program evaluation

with NTG.

slide-35
SLIDE 35

Delaware Evaluation Webinar February 2012

Page 35 of 54

Gross vs. Net Savings

Why all the fuss about gross vs. net

  • Gross savings are easier to measure but may not

reflect the actual savings induced by the program - policy makers not sure of actual impacts

  • Net savings are difficult (and sometimes)

impossible to accurately measure but in theory the provide a better measure of what effect the program had (as opposed to what would have been naturally occurring)

slide-36
SLIDE 36

Delaware Evaluation Webinar February 2012

Page 36 of 54

NTG Approaches

– How do you measure program attribution?

  • Deemed, or stipulated, net-to-gross ratios (least

expensive/least precise)

  • Self-reporting surveys (moderately expensive)
  • Enhanced self-reporting surveys (moderately

expensive, but more than self-report alone)

  • Econometric methods (more expensive)
  • Market based analysis (data intensive, market

cooperation needed) – Best practices continually evolve, but typically include multiple methods or data sources

slide-37
SLIDE 37

Delaware Evaluation Webinar February 2012

Page 37 of 54

– Are net assessment approaches reliable?

  • Deemed, or stipulated, net-to-gross ratios

– least expensive – can be adjusted over time and based

  • n best information available
  • Self-reporting surveys

– Self selection / false response / positive outcome / bias

  • Enhanced self-reporting surveys

– does not solve the bias issue, just brings more of it in

  • Econometric methods

– Very difficult and expensive to do well

  • Market based analysis

– Often not granular enough and data can be difficult to get.

– Why climb this slippery slope? Additionality!

slide-38
SLIDE 38

Delaware Evaluation Webinar February 2012

Page 38 of 54

A process evaluation…

– Systematically reviews a single program or portfolio of programs – Regularly provides feedback on program progress and performance – Openly recognizes what is working well with a program and identifies program design issues and barriers to delivery – Clearly lays out actionable recommendations for program improvements or changes in program goals

  • And let’s not forget, it often fulfills a regulatory requirement and is
  • ften directly or indirectly tied to recovery
slide-39
SLIDE 39

Delaware Evaluation Webinar February 2012

Page 39 of 54

What is Assessed

– Program design and processes – Program administrative activities – Program delivery and implementation activities – Customer response – Internal and external program barriers – Market response – Program impacts

slide-40
SLIDE 40

Delaware Evaluation Webinar February 2012

Page 40 of 54

Elements of a Good Process Evaluation

A good process evaluation…

– Is objective and unbiased – Is systematic and timely – Identifies what is working well and opportunities for improvement – Maintains regular communication with stakeholders

  • Feedback should be provided after each major activity
  • Regular status updates should be communicated
  • Provide the option of a presentation to stakeholders
  • Share the findings with program managers and discuss

implications and improvements in advance of next program planning cycles

slide-41
SLIDE 41

41

Market Effects Evaluation

  • Market transformation is present when a

program design are effective at overcoming barriers to adoption within the marketplace.

  • Market transformation can have big effects.
  • Savings can be much larger than standard

programs when successful (using the market expands reach sometimes without expanding program administrative costs)

slide-42
SLIDE 42

42

ME Goal – early movement along the S-curve

Laggards

Time Without the program

MT Strategy A MT Strategy B

slide-43
SLIDE 43

43

Cost Effectiveness

C/E Policy decisions drive everything The TRC is the EE energy supply valve. How cost effectiveness test are set up sets the limits of what can be achieved.

slide-44
SLIDE 44

Delaware Evaluation Webinar February 2012

Page 44 of 54

The TRC is the EE Supply Valve

44

Harder

TRC opens and closes the EE accomplishment gate

Barriers to EE

TRC Gate

Low est cost EE Low er cost EE Lower cost EE Energy Efficiency Potential Med cost EE High cost EE Easy Harder Hardest Harder

slide-45
SLIDE 45

Delaware Evaluation Webinar February 2012

Page 45 of 54

The Control Valves are…

45

Low High

Avoided Cost

Low EE - - - - - - - - - - - - - - - - - - - - - - High EE

Low High

Discounting

High EE - - - - - - - - - - - - - - - - - - - - - - Low EE

Low High

Carbon Value

Low EE - - - - - - - - - - - - - - - - - - - - - - High EE

Low High

Effective Useful Life

Low EE - - - - - - - - - - - - - - - - - - - - - - High EE

These values are set is policy decisions. How they are set limits the amount of energy and carbon that can be saved.

slide-46
SLIDE 46

Delaware Evaluation Webinar February 2012

Page 46 of 54

Cost Effectiveness Tests

Test Acronym Key questions answered Summary approach

Participant cost test PCT Will customers choose to participate? Comparison of costs and benefits

  • f the customer installing the
  • measure. Ignores impact on utility,

non-participants and society of making or not making the investment Program administrator cost test / (Utility cost test) PACT Will utility bills in aggregate be lower? Comparison of program administrator costs to supply-side resource costs Ratepayer impact measure RIM Will energy efficiency contribute to utility rates increase? Comparison of administrator costs and utility bill reductions to supply-side resource costs Total resource cost test TRC Will the total costs of energy decrease? Comparison of program administrator and customer efficiency costs to utility resource savings Societal cost test SCT Is society better off as a whole? Comparison of society’s costs of energy efficiency to resource savings in addition to other societal costs and benefits

slide-47
SLIDE 47

Delaware Evaluation Webinar February 2012

Page 47 of 54

Different tests do different things

Cost Benefit Elements

TRC RIM UCT PCT SCT

Benefits

Avoided Power Supply Costs X X X X Avoided T&D Costs X X X X Bill Reductions X Non Energy Benefits X

Costs

Direct Utility DSM Costs X X X X Direct Customer DSM Costs X X X Utility Program Administration X X X X Lost Revenues X

slide-48
SLIDE 48

Introducing Evaluation Plans

slide-49
SLIDE 49

It is the Evaluation Road Map

– Provide overview of program by

  • Confirming we have an accurate understanding of

program and its goals

– Documents research objectives, program metrics, and researchable issues – Lays out research methodology, assumptions, and activities – Ensures pressing research questions are included in activity – Establishes a timeline and key deliverables

49

slide-50
SLIDE 50

Evaluation Plan – Possible Contents

– Evaluation Goals – Program Description – Program Logic Model – Key Researchable Questions – Metrics to Be Measured – Sampling/Data Collection Plan – Cross-Cutting Evaluation Activities – Impact Evaluation Approach – Process Evaluation Approach – Market Effects Assessment – Budget and Schedule – Report Content and Structure.

50

slide-51
SLIDE 51

Evaluation Costs

– Generally, total evaluation costs range from three percent to six percent of program costs – Most states set @ 5% for evaluation – Budget drives reliability

  • What reliability do you want – need?

– Evaluation costs can vary depending upon the following elements:

  • Size of program
  • Maturity of program & past evaluation efforts
  • Purpose of the evaluation
  • Type of evaluation
  • Audience
  • Required level of statistical precision
  • Timing – how often you evaluate each program

51

slide-52
SLIDE 52

Delaware Evaluation Webinar February 2012

Page 52 of 54

Cost-Quality-Time Relationship Quality – Time – Cost

Choose two!

slide-53
SLIDE 53

Reporting Results

Standard format includes following pieces: – Executive summary – Introduction

  • Includes program description, researchable issues,

and methodology.

– Key findings – Conclusions and recommendations – Appendices with documentation on sample design, response rate, data collection instruments, technical data, etc.

53

slide-54
SLIDE 54

Wrap-up and Questions