Protocols and Regulatory Guidance

The purpose of this document is to establish minimum requirements for load impact estimation for DR resources and to provide guidance concerning issues that must be addressed and methods that can be used to develop load impact estimates for use in long term resource planning. The minimum requirements indicate that uncertainty adjusted, hourly load impact estimates be provided for selected day types and that certain statistics be reported that will allow reviewers to assess the validity of the analysis that underlies the estimates.

 

Event Based Resources

Non-Event Based Resources

Day Types

Event Driven Pricing

Direct Load Control

Callable DR

Non-event Driven Pricing

Scheduled DR

Permanent Load Reductions

Ex Post Day Types

 

 

 

 

 

 

Each Event Day

X

X

X

 

 

 

Average Event Day

X

X

X

 

 

 

Average Weekday Each Month

 

 

 

X

X

X

Monthly System Peak Day

 

 

 

X

X

X

Ex Ante Day Types

 

 

 

 

 

 

Typical Event Day

X

X

X

 

 

 

Average Weekday Each Month (1-in-2 and 1-in-10 Weather Year)

     

X

X

X

Monthly System Peak Day (1-in-2 and 1-in-10 Weather Year)

X

X

X

X

X

X

Table 2-2. SCE Demand Response Resources

Table 2-3. SDG&E Demand Response Resources

The purpose of this document is to establish minimum requirements for load impact estimation for DR resources and to provide guidance concerning issues that must be addressed and methods that can be used to develop load impact estimates for use in long term resource planning. The minimum requirements indicate that uncertainty adjusted, hourly load impact estimates be provided for selected day types and that certain statistics be reported that will allow reviewers to assess the validity of the analysis that underlies the estimates.

Protocol 1:

Prior to conducting a load impact evaluation for a demand response (DR) resource option, an evaluation plan must be produced. The plan must meet the requirements delineated in Protocols 2 and 3. The plan must also include a budget estimate and timeline.17

Protocol 2:

Protocols 4 through 27 establish the minimum requirements for load impact estimation for long term resource planning. There are other potential applications for load impact estimates that may have additional requirements. These include, but are not necessarily limited to:

· Forecasting DR resource impacts for resource adequacy;

· Forecasting DR resource impacts for operational dispatch by the CAISO;

· Ex post estimation of DR resource impacts for use in customer settlement; and

· Monthly reporting of progress towards DR resource goals.

The evaluation plan required by Protocol 1 must delineate whether the proposed DR resource impact methods and estimates are intended to also meet the requirements associated with the above applications or others that might arise and, if so, delineate what those requirements are.

Protocol 3:

The evaluation plan must delineate whether the following issues are to be addressed during the impact estimation process and, if not, why not:

· The target level of confidence and precision in the impact estimates that is being sought from the evaluation effort;

· Whether the evaluation activity is focused exclusively on producing ex post impact estimates or will also be used to produce ex ante estimates;

· If ex ante estimates are needed, whether changes are anticipated to occur over the forecast horizon in the characteristics of the DR offer or in the magnitude or characteristics of the participant population;

· Whether it is the intent to explicitly incorporate impact persistence into the analysis and, if so, the types of persistence that will be explicitly addressed (e.g., persistence beyond the funded life of the DR resource; changes in average impacts over time due to changes in customer behavior; changes in average impacts over time due to technology degradation, etc.);

· Whether it is the intent to develop impact estimates for geographic sub-regions and, if so, what those regions are;

· Whether it is the intent to develop impact estimates for sub-hourly intervals and, if so, what those intervals are;

· Whether it is the intent to develop impact estimates for specific sub- segments of the participant population and, if so, what those sub-segments are;

· Whether it is the intent to develop impact estimates for event-based resources for specific days (e.g., the day before and/or day after an event) or day types (e.g., hotter or cooler days) in addition to the minimum day types delineated in protocols 8, 15 and 22;

· Whether it is the intent to determine not just what the DR resource impacts are, but to also investigate why the estimates are what they are and, if so, the extent to which Measurement and Verification activities will be used to inform this understanding ;

· Whether free riders and/or structural benefiters are likely to be present among DR resource participants and, if so, whether it is the intent to estimate the number and/or percent of DR resource participants who are structural benefiters or free riders;

· Whether a non-participant control group is appropriate for impact estimation and, if so, what steps will be taken to ensure that use of such a control group will not introduce bias into the impact estimates; and

· Whether it is the intent to use a common methodology or to pool data across utilities when multiple utilities have implemented the same DR resource option.

Methodology

Ex Post Event Based Resources

Ex Post Non-Event Based Resources

Ex Ante Estimation

Participants Similar to the Past

Participants Different from the Past

Day-matching

-Hourly usage for event and reference value days

-Customer type19

Not Applicable

Not Applicable

Not Applicable

Regression

-Hourly usage for all days

-Weather20

-Hourly usage for participants

-Hourly usage for participants prior to participation and/or for control group

-Weather

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Same as prior columns

-Survey data on participant characteristics

-Projections of participant characteristics

Demand Modeling

-Same as above

-Prices

-Same as above

-Prices

-Same as prior columns & above row

-Same as prior columns & above row

Engineering

-Detailed information on equipment and/or building characteristics

-Weather (for weather-sensitive loads)

-Same as prior column

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Projections of participant characteristics

Sub-metering

-Hourly usage for sub-metered loads

-Weather for weather sensitive loads

Hourly usage for sub-metered loads for participants prior to participation and/or for control group

-Weather for weather sensitive loads

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Projections of participant characteristics

Experimentation

-Hourly usage for control & treatment customers

-Weather

-Hourly usage for control & treatment customers for pretreatment & treatment periods

-Weather

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Same as prior columns

-Weather for ex ante day types

-Other conditions for ex ante scenarios

-Projections of participant characteristics

Additional Research Needs

Additional Input Data Requirements

What is the required level of statistical precision?

-Ceteris paribus, greater precision requires larger sample sizes.

Are ex ante estimates required and, if so, what is expected to change?

-Incremental data needs will depend on what is expected to change in the future (see Table 3-1)

Are estimates of impact persistence needed?

-Estimating changes in behavioral response over time should be based on multiple years of data for the same participant population.

-Estimates of equipment decay could be based on data on projected equipment lifetimes, manufacturer's studies, laboratory studies, etc.

-If multiple years of data are not available, examination of impact estimates over time from other utilities that have had similar resources in place for a number of years can be used.

Are impacts needed for geographic sub-regions?

-Data needs vary with methodology.

-Could require data on much larger samples of customers (with sampling done at the geographic sub-region level).

-Could require survey data on customers to reflect cross-sectional variation in key drivers.

Are estimates needed for sub-hourly time periods?

-Requires sub-hourly measurement of energy use. If existing meters are not capable of this, could require meter replacement for sample of customers.

Are estimates needed for specific customer segments?

-Could require data on much larger samples of customers, segmented by characteristics of interest.

-Additional survey data on customer characteristics is needed.

Do you need to know why the impacts are what they are?

-Could add extensively to the data requirements, possibly requiring survey data on customer behavior and/or on-site inspection of equipment.

Do you need to know the number of structural benefiters?

-Could require larger sample sizes and/or additional survey data.

Is an external control group needed?

-Requires usage data on control group.

-Survey data needed to ensure control is good match for participant population.

Is a common methodology and joint estimation being done for common resource options across utilities?

-Will likely require smaller samples compared with doing multiple evaluations separately.

-May require additional survey data to control for differences across utilities.

Protocol 4:

The mean change in energy use per hour (kWh/hr) for each hour of the day shall be estimated for each day type and level of aggregation defined in the following Protocol 8. Protocol also calls for the mean change in energy use for the day must also be reported for each day type.

Protocol 5:

The mean change in energy use per year shall be reported for the average across all participants and for the sum of all participants on a DR resource option for each year over which the evaluation is conducted.

Protocol 6:

Estimates shall be provided for the 10th, 30th, 50th, 70th and 90th percentiles of the change in energy use in each hour, day and year, as described in Protocols 4 and 5, for each day-type and level of aggregation described in Protocol 8.

An application of protocol 6 to the production of the information required by the reporting templates (Table 4-1, below) is presented in "Day-Matching Analysis - An Example" on page 54.

Protocol 7:

Impact estimates shall be reported in the format depicted in Table 4-1 for all required day types and levels of aggregation, as delineated in Protocol 8.

Protocol 8:

The information shown in Table 4-1 shall be provided for each of the following day types and levels of aggregation:

    · Each day on which an event was called;

    · The average event day over the evaluation period;

    · For the average across all participants notified on each day on which an event was called;

    · For the total of all participants notified on each day on which an event was called; and

    · For the average across all participants notified on the average event day over the evaluation period.

An average event day is calculated as a day-weighted average of all event days.29 The number of event days that apply to each hour may vary for resource options that have variable length event periods.30 As such, for the average event day, the following information must be provided:

    · The number of actual event days included in the calculation for each hour of the average day;

    · Average number of customers enrolled in the resource option over the year31; and

    · Average number of customers notified across all event days in the year.

In addition to the information contained in Table 4-1, the following information must be provided for each event day:

    · Event start and stop time;

    · Notification lead time;

    · The number of customers who were enrolled in the resource option on the event day;

    · The number of customers who were notified on the event day; and

    · Any other factors that vary across event days that are considered by the evaluator to be important for understanding and interpreting the impacts and why they vary across events.

Protocol 9:

This statistical measures protocol is specific to Day-matching methods. A different protocol (e.g., protocol 10) is appropriate for regression methods. These calculations should be based on a suitable and sufficiently large number of proxy days. From this process, the following statistics should be calculated and reported for day-matching reference value methods:

    · The number of proxy days used in the calculations below and an explanation of how the proxy days were selected.

    · Average error across customers and proxy days for each hour for the entire day. This is calculated as follows:

                  (4-1)

where:

        i = the cross-sectional unit or customer

        j = the event-like day

        · = the hour of the day

        = the actual load for the customer on the proxy day of interest for the hour of interest

        = the predicted load for the customer on the proxy day of interest for the hour of interest

        = the total number of customers in the observation group

        = the total number of days in the observation group

    · Median error across customers and proxy days for each hour for the entire day. The median error is the error corresponding to the exact center of the distribution of errors when all the errors under consideration are arranged in order of magnitude. It is calculated as follows:

        a. Calculate the error for each customer and proxy day for the hour of interest:

        b. Sort the resulting distribution of errors by magnitude for each hour of interest.

        c. If the number of errors is odd, the median is the error associated with the observation.

        d. If the number of errors is even, the median is the average of the errors associated with observations and .

    · The relative average error for each hour. This is calculated as the ratio of the average error to the average actual load that occurred in the hour:

        REL (4-2)

where:

        = the average error across customers and proxy days for the hour of interest

    · The relative median error for each hour. This is calculated as follows:

        REL (4-3)

where:

        = the median error across customers and proxy days for each hour for the entire day, as calculated above

        = the median load for the customer on the proxy day of interest

    · The Coefficient of Alienation32, which describes the percentage of the variation in actual load for each hour that is not explained by variation in the predicted load. This is calculated as follows:

                (4-4)

      where:

        i = the cross-sectional unit or customer

        j = the event-like day

        k = the hour of the day

        = the actual load for the customer on the proxy day of interest for the hour of interest

        = the predicted load for the customer on the proxy day of interest for the hour of interest

        = the average load on the proxy day of interest for the hour of interest

        = the total number of hours being observed on the proxy day

    · Theil's U, calculated as follows:

              (4-5)

      where:

        = the number of periods

        k = the period of interest

        = the actual observed load for the period of interest

        = the predicted load for the period of interest

Protocol 10:

For regression based methods, the following statistics and information shall be reported:

    · Adjusted R-squared or, if R-squared is not provided for the estimation procedure, the log-likelihood of the model;34

    · Total observations, number of cross-sectional units and number of time periods;

    · Coefficients for each of the parameters of the model;

    · Standard errors for each of the parameter estimates;

    · The variance-covariance matrix for the parameters;35

    · The tests conducted and the specific corrections conducted, if any, to ensure robust standard errors; and

    · How the evaluation assessed the accuracy and stability of the coefficient(s) that represent the load impact.

Date

Day Type

SDG&E Daily Peak (MW)

Avg. Load During Peak Period (11am-6pm)

Daily Peak

Rank for 2005

Proxy Day

           

Friday, July 22, 2005

Event-day

4,057.2

3,916.4

1

 

Monday, August 29, 2005

Non-event weekday

4,031.5

3,869.3

2

_

Friday, August 26, 2005

Event-day

3,995.3

3,834.4

3

 

Thursday, July 21, 2005

Event-day

3,985.0

3,848.5

4

 

Thursday, August 25, 2005

Non-event weekday

3,947.2

3,748.2

5

_

Wednesday, July 20, 2005

Non-event weekday

3,821.3

3,508.9

6

_

Saturday, August 27, 2005

Weekend or holiday

3,799.3

3,679.0

7

 

Tuesday, August 30, 2005

Non-event weekday

3,753.3

3,571.7

8

_

Thursday, September 29, 2005

Event-day

3,734.8

3,632.5

9

 

Sunday, August 28, 2005

Weekend or holiday

3,712.9

3,597.3

10

 

 

 

 

 

 

 

Day-Matching Method

Coefficient of Alienation

Theil's U

3-day average with day-of adjustment

3.740%

0.12104

5 day average with prior-day adjustment

3.736%

0.18109

Prior day, no adjustment

3.740%

0.19428

"It is important to recognize that energy savings estimates depend not on the predictive power of the model on energy use, but on the accuracy, stability, and precision of the coefficient that represents energy savings."

Problems that potentially bias estimates

Problems that lead to incorrect standard errors

1. Omitted Variable: This is a type of specification error. Omitted variables that are related to the dependent variable are picked up in the error term. If correlated with explanatory variables representing the load impacts, they will bias the parameter estimates.

1. Serial-Correlation: Also known as auto-correlation, this occurs when the error term for an observation is correlated with the error term in another observation. This can occur in any study where the order of the observations has some meaning. Although it occurs most frequently with time-series data, it can also be due to spatial factors and clustering (i.e., the error terms of individual customers are correlated).

3. Improper functional form: This occurs when the relationship of an explanatory variable to the dependent variable is incorrectly specified. For example, the function may be treating the variable as linear when, in fact, it is logarithmic. This type of error can lead to incorrect predictions of load impacts.

2. Heteroscedasticity: This occurs when the variance is not constant but is related to a continuous variable. Depending on the model, if unaccounted for, it can lead to incorrect inferences of the uncertainty of the estimates

4. Simultaneity: Otherwise known as endogeneity, this occurs when the dependent variable influences an explanatory variable. This is unlikely to be a problem in modeling load impacts.

3. Irrelevant Variables: When irrelevant variables are introduced into a model, they generally weaken the standard errors of the explanatory variables related to the dependent variable. This leads to overstating the uncertainty associated with the impacts of other explanatory variables.

5. Errors in Variables: Explanatory variables that contain measurement error can create bias if the measurement error is correlated with explanatory variables(s).

 

6. Influential data: A data point is considered influential if deleting it changes the parameter estimates. Influential variables are typically outliers with leverage. These are more of an issue with large C&I customers.

 

 

 

Per participant load impacts

 

 

 

Percentiles

Hour Ending

Temp (F)

Mean (kW)

10%

30%NC

50%

70% NC

90%

1

71.2

 -0.009

0.011 

 

-0.010 

 

-0.030 

2

70.0

 -0.033

-0.008 

 

-0.032 

 

-0.057 

3

68.8

 -0.064

-0.039 

 

-0.064 

 

-0.090 

4

67.8

 -0.074

-0.041 

 

-0.074 

 

-0.101 

5

66.9

 -0.057

-0.030 

 

-0.057 

 

-0.084 

6

66.1

 -0.047

-0.019 

 

-0.047 

 

-0.074 

7

65.9

 -0.034

-0.007 

 

-0.034 

 

-0.061 

8

67.2

 -0.033

-0.005 

 

-0.032 

 

-0.060 

9

70.1

 -0.017

0.011 

 

-0.017 

 

-0.044 

10

74.4

 -0.041

-0.013 

 

-0.041 

 

-0.068 

11

78.7

 -0.022

0.005 

 

-0.022 

 

-0.049 

12

82.9

 -0.023

0.004 

 

-0.023 

 

-0.051 

13

86.4

 0.030

0.058 

 

0.030 

 

0.002 

14

89.1

0.040

0.067 

 

0.040 

 

0.013 

15

90.8

 -0.185

-0.158 

 

-0.185 

 

-0.212 

16

91.7

 -0.160

-0.132 

 

-0.160 

 

-0.188 

17

91.6

 -0.131

-0.104 

 

-0.131 

 

-0.159 

18

90.5

 -0.090

-0.062 

 

-0.090 

 

-0.117 

19

88.2

 -0.057

-0.030 

 

-0.057 

 

-0.085 

20

84.5

 0.222

 0.249

 

0.222 

 

0.195 

21

80.2

 0.294

 0.320

 

0.294 

 

0.267 

22

76.7

 0.250

 0.275

 

0.250 

 

0.225 

23

74.3

 0.186

 0.210

 

0.186 

 

0.162 

24

72.6

 0.097

 0.118

 

0.097 

 

0.077 

· Non-event based pricing-This resource category includes TOU, RTP and related pricing variants that are not based on a called event-that is, they are in place for a season or a year.

· Scheduled DR-There are some loads that can be scheduled to be reduced at a regular time period. For example, a group of irrigation customers could be divided into five segments, with each segment agreeing to not irrigate/pump on a different selected weekday.

· Permanent load reductions and load shifting-Permanent load reductions are often associated with energy efficiency activities, but there are some technologies such as demand controllers that can result in permanent load reductions or load shifting. Examples of load shifting technologies include ice storage air conditioning, timers and energy management systems.

Protocol 11:

The mean change in energy use per hour (kWh/hr) for each hour of the day shall be estimated for each day type and level of aggregation defined in Protocol 15. The mean change in energy use for the day shall also be reported for each day type.

Protocol 12:

The mean change in energy use per month and per year shall be reported for the average across all participants and for the sum of all participants in a DR resource option in each year over which the evaluation is conducted.

Protocol 13:

Estimates of the 10th,30th, 50th, 70th, and 90th percentiles of the change in energy use in each hour, day and year, as described in Protocols 11 and 12, for each day-type and level of aggregation described in Protocol 15, shall to be provided.

Protocol 14:

Impact estimates shall be reported in the format depicted in Table 4-1 for all required day types, as delineated in Protocol 15.

Protocol 15:

The information shown in Table 4-1 shall be provided for each of the following day types for the average across all participants sum of all participants:

    · For the average weekday for each month in which the DR resource is in effect53

    · For the monthly system peak day for each month in which the DR resource is in effect.

Day type definitions and additional reporting requirements for each day type are summarized below:

Average Week Day for Each Month: The average across all weekdays in each month during which the DR resource is in effect. In addition to the information contained in Table 4-1, the following information shall be provided:

    · Average temperature54 for each hour for a typical week day for each month.

    · Average degree hours for the typical week day for each month.

    · Average number of customers participating in the DR resource option each month

Monthly System Peak Day for Each Month: The day with the highest system load in each month. In addition to reporting all of the information shown in Table 4-1, the following information shall be provided:

    · Temperature for each hour on the system peak day for each month

    · Average degree hours on the system peak day for each month.

    · Average number of customers participating in the DR resource option on the system peak day for each month.

Protocol 16:

For regression based methods, the following statistics and information shall be reported:

    · Adjusted R-squared or, if R-squared is not provided for the estimation procedure, the log-likelihood of the model

    · Total observations, number of cross-sectional units and number of time periods

    · Coefficients for each of the parameters of the model

    · Standard errors for each of the parameter estimates

    · The variance-covariance matrix for the parameters

    · The tests conducted and the specific corrections conducted, if any, to ensure robust standard errors

    · How the evaluation assessed the accuracy and stability of the coefficient(s) that represent the load impact.

· Ex ante estimation may require developing estimates for values of key drivers that are outside the boundaries of historical experience (e.g., for extremely hot days that might not have occurred over the historical period) where the relationship of demand response and the variable of interest may differ from the relationship that exists within a narrower range of values;

· Ex ante estimation may require determining how demand response might evolve over time as participants become better educated about how to modify behavior in response to demand response stimuli or, alternatively, lose interest in modifying their behavior. The persistence of demand response impacts over time may also be impacted by degradation of or improvement in enabling technology, which may also need to be factored into ex ante estimates.

· Ex ante estimates are subject not only to the uncertainty associated with ex post impact estimates (e.g., due to sample selection, model specification and the like), but also to the additional uncertainty associated with the exogenous factors that drive demand response (e.g., uncertainty in weather, participation levels and customer characteristics, etc.).

Protocol 17:

Whenever possible, ex ante estimates of DR impacts should be informed by ex post empirical evidence from existing or prior DR resource options. Evidence from resource options and customer segments most relevant to the ex ante conditions being modeled should be used, regardless of whether they come from the host utility or some other utility. If ex post estimates or models are not used as the basis for ex ante estimation, an explanation as to why this is the case shall be provided.

Protocol 18:

The mean change in energy use per hour (kWh/hr) for each hour of the day shall be estimated for each day type and level of aggregation defined in Protocol 22. The mean change in energy use for the day shall also be estimated for each day type.

Protocol 19:

The mean change in energy use per month shall be estimated for non-event based resources and the mean change in energy use per year shall be estimated for both event and non-event based resources for the average across all participants and for the sum of all participants on a DR resource option for each year over the forecast horizon.

Protocol 20:

Estimates of the 10th, 30th, 50th, 70th and 90th percentiles of the change in energy use in each hour, day and year, as described in Protocols 17 and 18, and for each day-type described in Protocol 22, shall be provided.

Protocol 21:

Impact estimates shall be reported in the format depicted in Table 6-1 for all required day types and levels of aggregation, as delineated in Protocol 22.

Protocol 22:

The information shown in Table 6-1 shall be provided for each of the following day types using 1-in2 and 1-in-10 weather conditions for the average across participants and for the sum of all participants for each forecast year:

    · For a typical event day for a 1-in-2 and for a 1-in-10 weather year for event-based resource options.

    · For the average weekday for each month in which the resource option is in effect for a 1-in-2 and for a 1-in-10 weather year for non-event based resource options 58

    · For the monthly system peak day for each month in which the resource option is in effect, for a 1-in-2 and for a 1-in-10 weather year for event-based and non-event based resources.

Day type definitions and additional reporting requirements for each day type are summarized below:

Typical Event Day for a 1-in-2 and 1-in-10 Weather Year: This day type requirement applies primarily to event-based resources. It is meant to capture both the exogenous factors such as weather and the event characteristics for a day on which an event is likely to be called. The relevant characteristics can be defined by the evaluator. At a minimum, the following information shall be provided:

    · An explanation of how the weather and any other relevant day-type characteristics were chosen

    · Detailed information on the timing and duration of the event or any other factors (e.g., notification lead time) that were explicitly factored into the impact estimates (e.g., factors that, if different than those reported, would change the estimated impacts)

    · The number of notified consumers included in the aggregate impact estimate

    · Any other factors that have been explicitly incorporated into the impact estimate, such as prices for price based resource options and population characteristics (e.g., air conditioning saturation, business type, etc.).

Average Week Day for Each Month In A 1-in-2 and for a 1-in-10 Weather Year: This day type applies primarily to non-event based resources. It is meant to capture the weather conditions and other relevant factors for an average weekday. In addition to the information contained in Table 6-1, the following information must be provided:

    · An explanation of how the weather and any other relevant day-type characteristics were chosen for the typical weekday in each month

    · The number of enrolled customers included in the aggregate impact estimate

    · Any other factors that have been explicitly incorporated into the impact estimate, such as prices for price based resource options and population characteristics (e.g., air conditioning saturation, business type, etc.).

Monthly System Peak Day for Each Month In a 1-in-2 and for a 1-in-10 Weather Year: This day type applies to event- based and non-event based59 resources. It is meant to capture impacts for the day with the highest system load in each month. In addition to reporting all of the information shown in Table 6-1, the following information must be provided:

    · An explanation of how the weather and any other relevant day-type characteristics were chosen for the typical monthly system peak day

    · The number of enrolled customers included in the aggregate impact estimate

    · Any other factors that have been explicitly incorporated into the impact estimate, such as prices for price based resources and population characteristics (e.g., air conditioning saturation, business type, etc.).

Protocol 23:

All ex ante estimates based on regression methodologies shall report the same statistical measures as delineated in Protocols 10 and 16.

Protocol 25:

If sampling is required, evaluators shall use the following procedures to ensure that sampling bias is minimized and that its existence is detected and documented.

1. The population(s) under study must be clearly identified and described - this must be done for both participants and control groups to the extent that these are used;

2. The sample frame(s) (i.e., the list(s) from which samples are drawn) used to identify the population(s) under study must be carefully and accurately described and if the sample frame(s) do not perfectly overlap with the population(s) under study, the evaluator must describe the measures they have taken to adjust the results for the sample frame so that it reflects the characteristics in the population of interest - this would include the use of weighting, matching or regression analysis;

3. The sample design used in the study must be described in detail including the distributions of population and sample points across sampling strata (if any);

4. A digital snapshot of the population and initial sample from the sample frame must be preserved - this involves making a digital copy of the sample frame at the time at which the sample was drawn as well as a clean digital copy of the sample that was drawn including any descriptors needed to determine the sampling cells into which the sampled observations fall;

5. The "fate" of all sampled observations must be tracked and documented throughout the data collection process (from initial recruitment to study conclusion) so that it is possible to describe the extent to which the distribution of the sample(s) may depart from the distribution of the population(s) of interest throughout the course of the study;

6. If significant sample attrition is found to exist at any stage of the research process (i.e., recruitment, installation, operation), a study of its impact must be undertaken. This study should focus on discovering and describing any sampling bias that may have occurred as a result of selection. This should be done by comparing the known characteristics of the observed sample with the known characteristics of the population. Known characteristics would include such variables as historical energy use, time in residence, geographical location, reason for attrition from sample, and any other information that may be available for the population and sample.

7. If selection bias is suspected, the evaluator must describe it as well as any efforts made to control for it.65

 

Event Based Resources

Non-Event Based Resources

Day Types

Event Driven Pricing

Direct Load Control

Callable DR

Non-event Driven Pricing

Scheduled DR

Permanent Load Reductions

Ex Post Day Types

 

 

 

 

 

 

Each Event Day

X

X

X

 

 

 

Average Event Day

X

X

X

 

 

 

Average Weekday Each Month

 

 

 

X

X

X

Monthly System Peak Day

 

 

 

X

X

X

Ex Ante Day Types

 

 

 

 

 

 

Typical Event Day

X

X

X

 

 

 

Average Weekday Each Month (1-in-2 and 1-in-10 Weather Year)

X

X

X

X

X

X

Monthly System Peak Day (1-in-2 and 1-in-10 Weather Year)

X

X

X

X

X

X

1 R07-01-041, p.1.

2 Assigned Commissioner and Administrative Law Judge's Scoping Memo and Ruling, April 18, 2007

3 CPUC/CEC. Staff Guidance for Straw Proposals On: Load Impact Estimation from DR and Cost-Effectiveness Methods for DR. May 24, 2007. p.10.

4 Stephen George, Michael Sullivan and Josh Bode. Joint IOU Straw Proposal on Load Impact Estimation for Demand Response. Prepared on behalf of Pacific Gas & Electric Co., Southern California Edison Co., and San Diego Gas & Electric Co. July 16, 2007.

5 EnerNOC, Inc., Energy Connect, Comverge, Inc., Ancillary Services Coalition, and California Large Energy Consumers Association.

6 The Joint IOUs filed a motion on August 7th to obtain permission to file a revised proposal incorporating agreements reached at the August 1st workshop and to modify the original schedule to allow for this submission to made and for comments to be provided prior to the Commission's ruling. The presiding administrative law judge granted the Joint IOU request in a ruling on August 13, 2007.

7 The following parties filed comments on the Staff Report: Comverge, EnerNOC, and Energy Connect (jointly), the IOUs(jointly), CAISO, DRA, TURN, KM, and Wal-Mart.

8 R07-01-041, p.2.

9 R07-01-041, p.1.

10 Ibid. p.2

11 CPUC/CEC. Staff Guidance for Straw Proposals On: Load Impact Estimation from DR and Cost-Effectiveness Methods for DR. May 24, 2007, with assistance by Summit Blue Consulting, p.10.

12 Pacific Gas & Electric Co., Southern California Edison Co., and San Diego Gas & Electric Co.

13 EnerNOC, Inc., Energy Connect, Comverge, Inc., Ancillary Services Coalition, and California Large Energy Consumers Association.

14 The Joint IOUs filed a motion on August 7th to obtain permission to file a revised proposal incorporating agreements reached at the August 1st workshop and to modify the original schedule to allow for this submission to made and for comments to be provided prior to the Commission's ruling. The presiding ALJ granted the Joint IOU motion on August 13, 2007.

15 The following parties filed comments on the Staff Report: Comverge, EnerNOC, and Energy Connect (jointly), the IOUs(jointly), CAISO, DRA, TURN, KM, and Wal-Mart.

16 The original intent was to include summaries of many more studies in the appendix but there was not sufficient time to complete this work. The studies contained in the appendix are by no means the only examples of exemplary or interesting work in this area.

17 The final budget and timeline may differ from the planned budget and timeline as a result of the contractor selection process.

18 The various methodologies and applications contained in the table are discussed at length in subsequent sections.

19 The best day-matching method may vary across customer segments.

20 In all cases, weather data must be mapped to the locations of customers in the estimation sample.

21 The DRMEC was established by the CPUC in Decision # D.06-11-049 as an informal group charged with developing evaluation plans for demand response resources and reviewing interim evaluations of ongoing demand response programs.

22 Some of the reasons why day-matching methods are not viewed as robust as regression approaches include the need to produce estimates in a short time frame for settlements, as a result most day-matching methods are designed to produce estimates within a few days after an event to allow for prompt payments to participants. This limits the amount of data that is used, e.g., regression methods can use an entire season's data and data across multiple events to improve on the accuracy of impact estimates. Forecasting future impacts of DR events is limited with day-matching methods as they usually do not collect data on influential variables that would cause impacts for vary in the future. However, day-matching methods can be combined with regression and other statistical approaches to develop forecasts of impacts if day-matching estimates are available for several years and can be combined specific customer data as well as event-day data such as temperature, and system data.

23 This could occur if load control is used in combination with a CPP tariff, for example.

24 For example, one can imagine a DR resource option that automatically switches off pumps that otherwise are always running and pretty much drawing the same load at all times. In this situation, sub-metering the pumps would provide a highly precise estimate of what the load would have been on the event day if they had not been switched off. However, this is not the typical situation faced by DR impact evaluators.

25 As discussed in Section 6, with ex ante estimation, uncertainty can also result from the inherent uncertainty associated with key drivers of DR impacts such as weather. If a user wants to know what impacts are likely to occur tomorrow or on a day with a specific weather profile, it is important to recognize that the temperature at 2 pm on the day of interest, for example, is not knowable. It may have a high probability of equaling 92 degrees, say, but it is more realistic to base impact estimates on some distribution of temperatures (preferably derived from historical weather data) with a mean of 92 degrees and a distribution that would indicate, for example, that the temperature has a 90 percent probability of being between 90 and 94 degrees.

26 Other methods include a comparison of means between control and treatment groups, engineering analysis, sub-metering, etc.

27 Given the significant variation in temperature across a day in many climate zones within California, often rising from the 60s to the 90s or higher between early morning and late afternoon, degree hours may be more informative for comparison purposes across locations than are maximum daily temperature or average temperature. Degree hours are typically better predictors of daily air conditioning load than is average or maximum temperature for a day.

28 There is at least one type of DR resource where enrollment is more difficult to define, namely a peak-time rebate program such as the one outlined by SDG&E in its AMI application. The program concept in that application was that all customers would be eligible to respond to a peak time rebate offering and some subset of the entire customer base would be aware of the offer through promotional schemes. Only customers who were aware would be in a position to respond. Thus, it is difficult to determine whether the number of enrolled customers for such a resource is all customers or just those who are aware and, if the latter, how to measure awareness.

29 Put another way, it is the sum of the impacts in each hour for each event day divided by the number of event days. The reason to think of this as a day-weighted average is because the weights to use when calculating the standard errors are squared.

30 For example, if there were 10 event days, and the event was triggered from 3 pm to 5 pm on all days and between 5 pm and 6 pm on 5 event days, the average for each hour between 3 pm and 5 pm would be based on all 10 days but the average from 5 pm to 6 pm would be based on the 5 event days on which the event was triggered for that hour.

31 Since enrollment will change over time, a day-weighted average should be calculated (e.g., if there were 2 event days in the year and there were 100 customers enrolled on the first event day and 200 on the second, the day-weighted average would be 150).

32 The Coefficient of Alienation is a measure of the error in a prediction algorithm (of any kind) relative to the variation about the mean of the variable being predicted. It is related to the Coefficient of Determination by the function k = (1-R2). The Coefficient of Determination is a measure of the goodness of fit of a statistical function to the variation in the dependent variable of interest. Correspondingly, the Coefficient of Alienation is a measure of the "badness of fit" or the amount of variation in the dependent variable that is not accounted for by the prediction function. The R2 obtained from regression analysis is a special case of the Coefficient of Determination in which the regression function is used to predict the value of the dependent variable. Coefficients of Determination and Alienation can be calculated for virtually any algorithm that makes a prediction of a dependent variable.

33 For examples of how Theil's U can be applied, see KEMA-XENERGY (Miriam L. Goldberg and G. Kennedy Agnew). Protocol Development for Demand Response Calculation-Findings and Recommendations. Prepared for the California Energy Commission, February 2003.

34 The log-likelihood is a standard output whenever a maximum likelihood method (vs. OLS) is employed. Most statistical packages produce the log-likelihood (or do so by default) when a maximum likelihood estimation is used. Many statistical packages will show the changes to the log-likelihood as the computer goes through the iterative process of finding the best fittings set of parameters. The log-likelihood may be expressed as a pseudo R-square as that may be more familiar to some researchers. The protocols request for the R-square or, if the R-squared is not available, the log-likelihood. The log-likelihood is often used for equations where the dependent variable is a takes on discrete values. This Logit or Tobit type models do not typically produce R-squared values. For example, an A/C cycling evaluation that relies on directly metered A/C units should be, theoretically, analyzed with Tobit regression because for many hours the A/C unit will have zero usage due to either low temperature or no one at home. In other words, it is a dependent variable (e.g., energy usage) is truncated at a value of zero. The Tobit output will likely not produce an R-squared in which case the log-likelihood is the standard output. Peter Kennedy, A Guide to Econometrics, Fifth Edition, MIT Press, 2003 on p. 23-24 and 42-46) discusses maximum likelihood estimation. Another source is Woolridge, Econometric Analysis of Cross Section and Panel Data, Chapter 13; and W Green's textbook on Econometric Analysis, Chapter 17. SAGE publications has published a booklet titled Maximum Likelihood Estimation. Any of the above references will illustrate the use of the log-likelihood of a model. (Source: Communication with Mr. Josh Bode, Freeman, Sullivan & Co.)

35 The variance-covariance matrix is needed in order to calculate the correlations between the model parameters for use in determining forecast precision and uncertainty bands.

36 This reference method is discussed in a recent LBNL report, Estimating DR Load Impacts: Evaluation of Baseline Load Models for Commercial Buildings in California, July 2, 2007.

37 This discussion is based on information in KEMA-XENERGY (Miriam L. Goldberg and G. Kennedy Agnew). Protocol Development for Demand Response Calculation-Findings and Recommendations. Prepared for the California Energy Commission, February 2003. p. 2-12. This report uses the term baseline for what we call reference value. Hereafter, we refer to this report as the KEMA/CEC study.

38 SDG&E AMI Proceeding (A.05-03-015). DRA Exhibit 109.

39 There are several ways to approach this calculation. Three are outlined below:

40 In this second approach, these standard errors come from the selected proxy days rather than from actual event days, as a result the standard errors from the proxy day analysis in protocol 9 are used as the best information on the likely standard errors for the event days. Actual standard errors for the event days can not be calculated as the true reference loads for those days are never known.

41 Some model specifications use ratios of energy use in different time periods as a dependent variable.

42 The reader is referred to the KEMA/CEC (2003) report for a useful comparison of the relative accuracy and other attributes of a variety of regression models and day-matching methods.

43 TecMarket Works. The California Evaluation Framework, June 2004. pp. 105 - 120.

44 Page 274-276 of J. Woolridge's textbook, Econometric Analysis of Cross-section and Panel Data provides an excellent discussion on serial correlation and the robust variance matrix estimator.

45 In this instance, separate output tables should be reported for each market segment.

46 There are situations in which an external control group might still be needed. For example, if an event is only called on the hottest days of the year, and the relationship between energy use on those days is different from what it is on other days, the model may not be able to accurately estimate resource impacts on event days. In this instance, it may be necessary to have a control group in order to accurately model the relationship between weather and energy use on the hottest days in order to obtain an unbiased estimate of the impact of the resource on those day types.

47 There may still be some interest in knowing how participants differ from non-participants if there is a need to extrapolate the impact estimates to a population of customers who are unlikely to volunteer (which may differ from those who have not yet volunteered). If so, an external control group may be needed. A more in depth discussion of control groups is contained in Section 5.2.

48 Charles River Associates. "Impact Evaluation of the California Statewide Pricing Pilot," Final Report. March 16, 2005, p. 66. See CEC Website: http://www.energy.ca.gov/demandresponse/documents/index.html#group2

49 Peter Kennedy. A Guide to Econometrics, Fifth Edition, MIT Press, 2003. This book provides an excellent discussion of some of the advantages of having repeated measures across a cross-section of customers in the introduction to Chapter 17. Kennedy (2003) is also a good general reference for the regression methods and issues discussed in this chapter.

50 Charles River Associates. Op. Cit. 2005. CEC web http://www.energy.ca.gov/demandresponse/documents/index.html#group2

51 Quantum Consulting Inc. The Air Conditioner Cycling Summer Discount Program Evaluation Study. January 2006.

52 The definition of M&V used here differs from how the term is sometimes used elsewhere. In some instances, M&V is defined much more broadly and essentially is synonymous with impact estimation. It is important to keep the narrower definition in mind when reviewing this section and when encountering the term elsewhere in this document.

53 If a resource is seasonal, only the months in which the resource is in effect needs to be reported.

54 As noted in Section 4, when reporting temperatures and degree days, it is intended that the temperature be reasonably representative of the population of participants associated with the impact estimates. If participation in a resource option is concentrated in a very hot climate zone, for example, reporting population-weighted average temperature across an entire utility service territory may not be very useful if a substantial number of customers are located in cooler climate zones. Some sort of customer or load-weighted average temperature across weather stations close to participant locations would be much more accurate and useful.

55 pp. 142-145.

56 The remainder of this discussion consists mainly of selected text from The California Evaluation Framework, pp. 120 - 129.

57 For more information on building energy simulation models, see State-of-the-Art Review: Whole Building, Building Envelope and HVAC Component and System Simulation and Design Tools. (Jacobs and Henderson 2002).

58 If a resource is seasonal, only the months in which the resource is in effect must be reported.

59 Nonevent-based resources may have impacts vary from day to day, and may be quite different on monthly peak days. Some nonevent-based resources will have impacts that are dependent on weather and, therefore, will vary across event-type days and on monthly peak days. As an example, one resource that falls into the nonevent category is ice storage that can be used to displace cooling loads on hot days. On the hottest days of the year, ice storage may have greater impacts since there is likely to be a greater demand for cooling that can be displaced. Using a baseline taken from AC loads that would otherwise have been utilized, ice storage may have larger impacts on hot days and monthly system peak days that are driven by higher electricity loads due to hot weather.

60 The threshold temperature above which most or all air conditioners will be running will vary depending upon the typical unit sizing practices for a location. It may be that many air conditioners will still be cycling above 100 degrees in some locations but most will be on in other locations.

61 CRA International. Residential Hourly Load Response to Critical Peak Pricing in the Statewide Pricing Pilot. May 18, 2006. CEC website: http://www.energy.ca.gov/demandresponse/documents/index.html#group2

62 Monte Carlo simulation is a straightforward, widely used approach for reflecting uncertainty in key model parameters, but there may be other approaches that can be used to accomplish the same objective.

63 Section 7.2.3 provides a detailed example of how failure to account for correlations can distort uncertainty estimates.

64 In theory, the convolutions of the underlying distributions of load impacts from different DR resources could be accomplished with calculus, but it is much easier to do so with Monte Carlo simulation.

65 The problem of controlling for selection bias has been discussed at great length in the literature on econometrics. The seminal articles on this topic are by James Heckman "The common structure of statistical models of truncation, sample selection and limited dependent variables and a simple estimator for such models", in The Annals of Economic and Social Measurement 5: 475-492 1976; and Sample selection bias as a specification error" in Econometrica, 47: 153-161

66 A level of precision that is quite high may be inappropriate for programs that are expected to have smaller impacts either due to the design of the program, or due to the program not yet attaining its target level of participation. If the DR impacts are small, achieving increasing high precision levels may likely to cost more than achieving the same levels of precision for programs with sizeable impacts and a large number of participants.

67 Classic textbooks useful in survey sampling include:

68 The actual equation for calculating sample size includes a correction for the size of the population called the finite population correction. This adjustment has been left off of the equation for ease of exposition. In general, its effect on the sample size calculation is de minimus when the population of interest is large (e.g., more than a few thousand).

69 Ibid.

70 See "Minimum Variance Stratification" Dalenius T. and Hodges J. L., Journal of the American Statistical Association, 1959, 4, pp. 88-101

71 See "On the two different aspects of the representational method: the method of stratified sampling and the method of purposive selection", Jerzy Neyman, Journal of the Royal Statistical Society, 1934, 97, pp 558-625.

72 Using Multivariate Statistics (3rd ed.), Tabachnick, B. G., & Fidell, L. S. New York: Harper Collins (1996).

73 See Frison and Pocock (1992) "Repeated measures in clinical trials: An analysis using mean summary statistics and its implications for design", in Statistics in Medicine 11: 1685-1704 for a technical discussion of the method used to estimate the impacts of repeated measures on sampling precision and sample size.

74 The DRMEC was established by the CPUC in Decision # D.06-11-049 as an informal group charged with developing evaluation plans for demand response resources and reviewing interim evaluations of ongoing demand response programs. Here is an excerpt from that decision: "In D.06-03-024, we authorized the Working Group 2 Measurement and Evaluation subcommittee to continue its work in providing oversight of demand response evaluation, and we continue that authorization for the program augmentations we approve here under the more appropriate name of the Demand Response Measurement and Evaluation Committee. Due to the importance of monitoring and assessing the progress of these programs, the IOUs will provide all data and background information used in monitoring and evaluation projects to Energy Division and the CEC, subject to appropriate confidentiality protections."

Previous PageTop Of PageGo To First Page