PRISM MARS
Comparison of PRISM and MARS
A comprehensive technical summary that compares two LOLE assessment
tools: PJM’s Probabilistic Reliability Index Model (PRISM) and General
Electric International, Inc.’s Multi-Area Reliability Simulation (MARS).
PJM Resource Adequacy Analysis Subcommittee (RAAS)
February 17, 2011
DOCs # 599848
© PJM Interconnection 2011. All rights reserved
Table of Contents
Executive Summary ...................................................................................................................................... 3
Introduction ................................................................................................................................................... 6
Background ................................................................................................................................................... 7
Reliability Criterion … 1 day in 10 years ...................................................................................... 7
Early Reserve Study Models: GEBGE, PTI’s MAREL and GE-MARL / GE-MARS ........................... 8
Comparing GEBGE with MAREL ................................................................................................... 9
Current Reserve study Models .................................................................................................................. 10
Two–Area Model: PJM PRISM ................................................................................................... 10
Multi-Area Model: GE-MARS ..................................................................................................... 11
Complementary Aspects of PRISM and MARS ........................................................................... 14
Ongoing LOLE Assessment Work ............................................................................................... 15
Model Calculation Processes .................................................................................................................... 17
PRISM Calculation Processes ..................................................................................................... 17
MARS Calculation Processes ...................................................................................................... 23
Model Calculation Comparisons: Observations ......................................................................... 33
PRISM’s ability to automatically solve to a specified LOLE level ............................................... 34
Historical Continuity .................................................................................................................. 34
PJM’s Recent Comparison Efforts ............................................................................................................ 36
Access to models: ARC screen display and GUI layout .............................................................. 38
#1 Comparing the Models ......................................................................................................... 41
#2 Comparing the Calculations Performed................................................................................ 41
#3 Comparing the Available Standard Output .......................................................................... 41
Comparison of Attributes Documentation ................................................................................................ 42
Table 1 – Model Attributes / Application Comparison Matrix .................................................. 42
Table 2 – OUTPUT Comparison Matrix ...................................................................................... 50
Table 3 – DATABASE Modeling Relationship Matrix ................................................................. 56
Summary of Tables 1, 2, & 3 Comparisons ................................................................................ 62
Resource & Cost Assessment ................................................................................................................... 64
Frequently Asked Questions (FAQs) ........................................................................................................ 73
Interregional Assessments ........................................................................................................................ 83
Glossary....................................................................................................................................................... 85
References................................................................................................................................................... 94
© PJM Interconnection 2011. All rights reserved
Page 1 of 136
APPENDICES ............................................................................................................................................... 97
Appendix A – MARS Model vs. PRISM Model Overview PJM Staff presentation ..................... 98
Appendix B – Translation from PLOTS to LOD-UNCY table ..................................................... 103
Appendix C – Load Modeling Comparison Issues .................................................................... 105
Appendix C1 – Load distribution granularity .......................................................................... 105
Appendix C2 – Baldwin Paper................................................................................................. 107
Appendix C3 – ISO-NE’s Comparison of Westinghouse Model and MARS ............................. 110
Appendix D – MARS public solution techniques ..................................................................... 115
Appendix E – Transparency of process to perform LOLE calculations .................................... 122
Appendix F – Details of Input Parameter Requirements......................................................... 127
Appendix G – Items for possible future investigation ............................................................. 136
© PJM Interconnection 2011. All rights reserved
Page 2 of 136
Executive Summary
For many years, PJM has conducted its annual resource adequacy modeling and Installed Reserve Margin
(IRM) studies using their in-house, two-area model: the Probabilistic Reliability Index Model (PRISM)
program. By comparison, PJM’s neighboring control areas (most notably the New York Independent
System Operator (NYISO), the Independent System Operator of New England (ISO-NE) and the Midwest
Independent System Operator (MISO)) utilize a multiple-area program, the Multi-Area Reliability Simulation
(MARS) that was developed by General Electric International, Inc. (GEII).
Resource adequacy models apply both deterministic and probabilistic methods with varying degrees of data
requirements and various input / output. By nature of its inherent complexity, multi-area assessments
involve more operational specific and rigorous data requirements than does a two-area model. Multi-area
assessments also require more time, computing horsepower and resources to evaluate and interpret the
results.
The PJM Capacity Market is a $6+ billion annual enterprise. As such, the determination, characterization
and quantification of resource adequacy is important to all PJM Stakeholders. Over the past four decades,
resource adequacy modeling has typically involved the modeling of interconnected areas. Such programs
can employ either a one- or two-area modeling approach or more expansive multiple-area modeling.
Several members on the PJM Reserve Adequacy Analysis Subcommittee (RAAS) have long encouraged
PJM to adopt MARS in place of PRISM while other members believe that PRISM already fills the bill with
respect to modeling resource adequacy.
Despite the fact that both PRISM and MARS have been known and proven in the electric industry for
decades, there is a lingering question: Which is the better model for resource assessments?
Evaluation of this question began years ago – at the RAAS and its predecessor groups. By early 2010, it
was agreed that a detailed comparison was in order. The genesis of the report began with a simple
comparison table (similar to that of Table 1). Subsequent feedback from the RAAS participants thereby
prompted more comprehensive evaluation and the opportunity to document the history and evolution of
resource adequacy modeling.
PJM has used both PRISM and MARS since 2005, taking advantage of the complementary features offered
by both. This “blended” approach using both programs is regarded by PJM Staff as more beneficial than
using just one of the programs exclusively. Realizing that there is somewhat more intensive and rigorous
database requirements for MARS, PJM has relied more heavily on PRISM for its resource modeling needs.
This decision enables PJM to fully utilize their existing staff resources, structures and tools to achieve the
best technical results at the least cost.
This evaluation summarizes the models, calculations and output with detailed comparison of various
attributes, strengths and weaknesses. Cost estimates were also developed to provide a comparative
framework for needed resources.
The purpose of this report is to deliver an objective technical evaluation – not to endorse one
modeling method over another (per unanimous RAAS directive). In some cases, PRISM and MARS
provide complementary information that enhances the overall reliability study. This report is designed to
stimulate technical discussion and lay the groundwork for possible further study and future action items.
The intent of this evaluation is to provide an objective analysis and technical comparison of both PRISM and
MARS. This comparative analysis attempts to evaluate the consistency of: A) the models, B) the data and
calculations and C) the standard output. Key findings are summarized as follows:
A. Comparing the Models:
PRISM

PRISM is a two-area simulation model.

PRISM uses a probabilistic distribution to model load.
© PJM Interconnection 2011. All rights reserved
Page 3 of 136

PRISM uses probabilistic distributions for capacity modeling (as does MARS).

PRISM uses deterministic distributions for transmission system modeling (as does
MARS).

PRISM uses statistical input data requirements. This is a distinct advantage as PRISM
has relatively few statistical parameters that incorporate the full model.

PRISM uses a probabilistic forecast load model.

PRISM features a comparatively fast solution time; this advantage thereby allows more
time for additional sensitivity cases to more fully assess system impacts.

Running PRISM requires no additional PJM staff to complete a resource adequacy
assessment.

Unlike MARS, PRISM cannot perform Hourly assessments or determine metrics regarding
either Loss-of-Load Hours (LOLH) or Expected Unserved Energy (EUE).
MARS

MARS is a multiple-area model; this is a tremendous advantage.

MARS uses a deterministic load distribution.

MARS uses probabilistic distributions for capacity modeling (as does PRISM).

MARS uses deterministic distributions for transmission system modeling (as does
PRISM).

MARS can perform hourly calculations and include more direct Operational parameters.

Compared to PRISM, data collection and maintenance of inputs for MARS is relatively
time-intensive.

Being a Monte Carlo simulator, MARS requires longer solution times.

MARS is ably supports and continuously refined and updated by GE technical staff.

MARS has gained industry-wide acceptance and usage throughout North America.
MARS will therefore, be used to comply with any forthcoming North American Reliability
Corporation (NERC) Planning Committee (NERC-PC) reporting for new metrics, LOLH
and EUE.
B. Comparing the Data & Calculations:
The significance of the mantra “It is all about the data!” cannot be overlooked in Loss of Load
Expectation (LOLE) assessment work: Having high-quality data allows enables greater confidence,
correct interpretation of reported results, appropriate decision making and high accuracy of final
reported values. PJM’s database management conforms to standards and processes governed by
third-party audits and assurance that Best Practices are used in the underlying data systems.

PJM uses the same main graphical user interface (GUI) and underlying database for both
PRISM and MARS. Both programs are included in the Applications for Reliability Calculations
(ARC) process used by the PJM staff. Many data relationships are established to assist
automating a consistent model between both tools. (See Table 3 for further details.)

MARS contains more data input categories than does PRISM (see Appendix F). This allows
MARS more flexibility and the potential to perform assessments that PRISM cannot perform.
© PJM Interconnection 2011. All rights reserved
Page 4 of 136
But it also requires increased coordination among PJM Staff and more time to maintain and
update the data.

While PJM Staff is able to perform most Adequacy work in-house, increased MARS efforts
would likely increase member technical representatives’ responsibilities and efforts. This
dynamic can be witnessed by comparing MARS assessment work that is done by neighboring
RTOs, ISOs and EROs.
C. Comparing the Output:

MARS contains several output summaries not found in PRISM, including: load level,
Emergency Operating Procedure level, and Interface flows.

PRISM uses a database schema both for its input and output results. This allows a mapping of
the relationships between these data and is defined in an OLAP metadata process that allows
assessment and reporting of several complex summaries.

Both PRISM and MARS offer many detailed outputs to perform LOLE assessments.
© PJM Interconnection 2011. All rights reserved
Page 5 of 136
Introduction
Resource capacity and reliability planning assessments are performed each year by PJM and its
neighboring RTOs, ISOs and EROs to ensure the reliability of the interregional bulk power system. These
comprehensive, time-intensive evaluations are critical to the planning process. Each region and reliability
entity that performs such resource adequacy studies does so with various tools and models at their disposal.
Modeling techniques generally utilize combinations of deterministic and probabilistic systems (such as
Monte Carlo simulation methods), depending on the modeling parameters, assumptions and inputs.
PJM Interconnection (PJM) currently uses the two-area Probabilistic Reliability Index Study Model (PRISM)
for reliability and resource adequacy modeling. Developed by PJM, PRISM uses cumulative probability
functions involving numerical techniques that combine (convolve) a probabilistic load distribution with a
probabilistic resource distribution.
Several of PJM’s neighbors, most notably the New York Independent System Operator (NYISO), the
Independent System Operator of New England (ISO-NE), the Midwest Independent System Operator
(MISO), and the SERC Reliability Corporation, use the General Electric International, Inc. (GEII) Multi-Area
Regional Simulator (MARS) program. MARS’ core calculations are based on sampling using the proven
Monte Carlo solution technique. MARS analyzes multi-area generation systems with explicit recognition of
unit and system operating considerations, rules, and constraints that influence system reliability indices.
MARS utilizes a random number generator to facilitate calculations with many replications performed to
satisfy a chosen solution criteria of standard error.
PJM’s Resource Adequacy Analysis Subcommittee (RAAS) reviews the modeling and analysis techniques
used in the reserve requirement annual study and Capacity Emergency Transfer Objective (CETO).
Reporting to PJM’s Planning Committee (PC), the RAAS charter also includes supporting capacity model
review. Several members expressed interest in using MARS as PJM’s primary modeling tool.
PJM’s use of MARS is increasing due to the capabilities inherent in a multi-area Monte Carlo solution tool.
These efforts address Industry topics such as Wind unit modeling, Demand Response (DR) assessments,
and collaboration with neighboring regions staffs in performing intraregional assessments.
While PRISM remains PJM’s primary resource adequacy evaluation tool, PJM is also licensed to use MARS
– and does so in the conduct of the overall Resource Requirements Study (RRS) report. In recent years,
PJM conducted comparative analysis of PRISM and MARS but until now, these efforts were not welldocumented. With the recent expansion of PJM’s regional footprint and increasing emphasis on
interregional planning activities, the RAAS determined it was worthwhile and timely to provide a more
comprehensive, fully-documented comparison.
The intent of this report is to provide comparative information of PRISM and MARS with respect to inputs,
assumptions, system requirements, required resources, outputs and results. This analysis strives to present
viewpoints useful for evaluating the technical aspects of the tools so that objective comparisons can be
made.
© PJM Interconnection 2011. All rights reserved
Page 6 of 136
Background
Reliability Criterion … 1 day in 10 years
Although it is difficult to determine when the first published material appeared describing reliability criterion,
the first notable group of papers on this topic appeared in 1947. In that year, several key papers were
presented, namely those by Calabrese1, Loane & Watchorn2, Lyman, Seelye and others. These documents
proposed the basic concepts upon which some of the methods in use at the present time are based.
Concepts such as the “loss of load method” and the “frequency and duration” approach were explored at
3
that time. Until that time, reserve capacity (planning reserve) was deterministically established to be in the
range of 18 to 24 percent.
In particular, the analysis by Loane & Watchorn used forced outage rates of generators in conjunction with
outage probabilities (valid under the assumption of exponential failure density and other approximations)
that resulted in a computed loss of load probability (probability of curtailment) of 0.32e-03.
The Calabrese and Watchorn papers, when taken together, presented similar ideas and philosophies. As
such, the resulting daily expectation of load loss can be combined, using a simple average, which yields an
approximate probability of curtailment of 0.4e-03.
Since using loads of working weekdays during the entire year in their model (~260 days/year), the
mathematical expectation of this risk became 0.4e-03 x 260 = 0.1 days per year … or 1 day in 10 years. As
other contemporary analyses eventually arrived at similar conclusions, the “1 in 10” evolved as an industry
standard for probabilistic bulk system planning, as the Loss of Load Expectation (LOLE).
Different load models gave different reserve margins, although they all used the same risk criterion, namely
one day in 10 years. There is considerable debate regarding risk level in technical journals and regulatory
proceedings. Some economists have long questioned the Bulk Electric System’s usage of the one day-inten year LOLE criterion. Notwithstanding such discussions, this risk criterion is ubiquitous and has gained
acceptance as an acceptable index for probabilistic resource adequacy planning. This acceptance ia also in
line with business and governmental officials’ expectations and requirements for the resource adequacy of
the bulk power grid.
From this historical backdrop, the bulk electric grid industry and related organizations have adopted the one
day (on average) in 10 years requirement – used for LOLE analysis in planning the Adequacy of the bulk
system.
The Northeast Power Coordinating Council (NPCC)4, the Mid-Continent Area Power Pool (MAPP)5 and the
6
previous NERC regions of the Mid-Atlantic Area Council (MAAC) and the Mid-America Interconnected
7
Network (MAIN) all conducted LOLE analysis to comply with this requirement. NPCC conducts LOLE
analysis to demonstrate its Member’s compliance with this requirement, as specified by the NPCC Bylaws.
1
Giuseppe Calabrese, “Generating Reserve Capacity Determined by the Probability Method” (March 25, 1947),
presented at the AIEE Midwest meeting of November 3, 1947 – see page 21 (which cites a 0.00046 probability of loss
of load)
2
E.S. Loane and C.W. Watchorn, “Probability Methods Applied to Generating Capacity Problems of a Combined Hydro
and Steam System” (August 14, 1947), presented at the AIEE Midwest meeting of November 3, 1947 – see page
1651 (which cites a 0.00032 probability of failure to carry peak daily load)
3
Roy Billinton and Ronald Allan, “Reliability Evaluation of Power Systems” (1984), Plenum Press, New York – page 7
4
NPCC Regional Reliability Reference Directory #1 “Design and Operation of the Bulk Power System (December 1,
2009) - section 5.2, page 9
5
MAPP Loss of Load Expectation (LOLE) Study, MAPPCOR and MAPP Composite System Reliability Working Group
(December 30, 2009); http://www.mapp.org/ReturnBinary.aspx?Params=584e5b5f405c567900000002cb
6
7
Mid-Atlantic Area Council (MAAC) Document A-1, MAAC Reliability Principles and Standards (March 30, 1990)
MAIN Guide #6, Generation Reliability Study 20045-2014 (September 27, 2005)
© PJM Interconnection 2011. All rights reserved
Page 7 of 136
Currently the ReliabilityFirst Corporation (RFC)8 has this criterion specified in their region’s Standard BAL502-RFC-02 as well as Midwest Reliability Organization (MRO)9 in their Standard RES-501-MRO-01.
Early Reserve Study Models: GEBGE, PTI’s MAREL and GE-MARL / GE-MARS
With the 1 day in 10 years LOLE reliability criterion having gained broad acceptance, engineers and
reliability professionals sought to develop simulation models. In the decades since, various modeling tools
were developed for one-area, two-area and multi-area applications. Although using the same risk criteria
these different models also yield different reserve margins 10.
One such early model to gain recognition and industry acceptance was the GEBGE. Developed by General
Electric International, Inc. (GEII) in conjunction with staff from Baltimore Gas & Electric (BGE), “GEBGE” is a
two-area, probabilistic reliability model. Starting sometime in the mid-1960s, GEBGE became the primary
reliability (R) model used by PJM – until replaced by the later-generation PRISM program. A 1969 paper by
Baldwin11 provides the general specifics for how GEBGE was created in line with the philosophies
expressed by both the Calabrese and Loane & Watchorn studies. Around this same time, the Westinghouse
two-area model also gained recognition and industry acceptance (used by ISO-NE Staff and is similar in
nature to GEBGE).
GEBGE uses a probabilistic load model based on a daily peak distribution, aggregated on a weekly basis,
derived from historic loads and a forecasted growth rate. The daily peak distribution is represented by a 21point normal curve. GEBGE’s capacity model uses each generating unit’s capacity, forced outage rate and
maintenance requirements to develop a cumulative capacity outage probability table for each week. On a
daily basis, GEBGE calculates probability of every possible load level simultaneous with every possible
generation availability level. Individual week LOLEs are summed over the entire year to determine the
annual LOLE. The GEBGE calculations and solutions techniques served as the basis in the development of
PJM’s PRISM model.
From 1988-1991, the former New York Power Pool (NYPP) Resource Planning Advisory Subcommittee
(RPAS) initiated a project to develop a multi-area computer program using two different technical solution
techniques: 1) the analytical method (convolution) and 2) the Monte Carlo method.
In response, power systems software developer Power Technologies Inc. (PTI) developed a prototype multiarea analytical method called the Multi-Area Reliability (MAREL) program. PTI’s approach imposed no
theoretical limitations on the number of areas or topology of the interconnected network. MAREL considered
various types of generation, transmission interfaces, firm interchanges (contracts) and emergency operating
procedures (EOPs).
GE, meanwhile, took up the Monte Carlo method and developed the Multi-Area Reliability (GE-MARL). GEMARL used a sequential Monte Carlo simulation for modeling the effects of random events such as forced
outages. GE-MARL considered various types of generation, transmission interfaces, firm interchanges
(contracts) and emergency operating procedures (EOPs). GE-MARL was later renamed GE-MARS (MultiArea Reliability Simulation) once the program went into production.
8
9
ReliabilityFirst Corporation (RFC) region planning resource Adequacy Analysis, Assessment and Documentation,
Standard BAL-502-RFC-02 (December 4, 2008): http://www.nerc.com/files/BAL-502-RFC-02.pdf
Midwest Reliability Organization (MRO) Standard RES-501-MRO-01, planned Resource Adequacy Assessment
(December 29, 2007); http://www.midwestreliability.org/04_standards/approved_standards/mro_standards/RES-501MRO-01_Final_20071229_Clean.pdf
10
11
“PJM Generation Adequacy Analysis: Technical methods” (October 2003) available at:
http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/res-adeq/20040621-white-papersections12.ashx.
C.J. Baldwin, “Probability Calculation of Generation Reserves” (March 1969), published by The Westinghouse
Engineer, This paper is copyright protected but can be purchased online at Infotrieve; article information accession
number 00434361 (800-422-4633; www.infotrieve.com).
© PJM Interconnection 2011. All rights reserved
Page 8 of 136
In 1991, ESEERCO evaluated the PTI-MAREL and GE-MARL programs.12 The NYPP ran various
simulations for single and multiple area and six-area regions. While MAREL and GE-MARL produced similar
results for the single and six-area representations, the GE-MARL program was able to solve the eleven and
fifteen-area simulations. NYPP therefore concluded that GE-MARL was the superior product.
Comparing GEBGE with MAREL
In 1993, the PJM Load and Capacity Working Group (L&CWG) conducted a comparison of the PTI-MAREL
application and GEBGE.13 Concerns over the possible shortcomings of the two-area model prompted this
comparison. From the outset, it was PJM’s intent that MAREL was not intended to replace GEBGE but
rather to complement GEBGE. The LCWG report made the following key observations and conclusions:

12
13

It was determined that the PJM Reserve Requirement should be calculated with GEBGE. The
multi-area MAREL results indicated that no accuracy was sacrificed by modeling a single-area
World as was done by GEBGE.

The benefit of modeling a multi-area World became apparent only when reserves were reduced
from their forecasted level of 26.6% down to 10%. Under this condition, load diversity and
transmission constraints within the World were judged to affect the reserve levels to meet a 1 day
in 10 years LOLE.

Since PJM installed capacity (ICAP) accounting is predicated on 1 day in 10.00 (hundredths
accuracy), the multi-area complexity of MAREL prevented it from obtaining a solution to this level of
accuracy.

Unlike GEBGE, MAREL lacked the capability to automatically adjust load levels to solve to a userspecified Reliability Index (RI).

LCWG recommended to purchase MAREL pointing out some compelling rationale:
o
It was thought that future conditions in the electric utility industry (such as increased wheeling
and Non-Utility Generation (NUG) activity) would require multi-area reliability assessments.
o
MAREL included several features not found in GEBGE, including the modeling of firm
contracts, demand side management and modeling of emergency operating procedures
(EOPs).
o
MAREL could be used (the only tool at that time) to verify GEBGE results.
o
Unlike GEBGE, MAREL calculated EUE in addition to LOLE.
Ultimately, PJM elected to not purchase MAREL. At some point thereafter, PTI discontinued further
development of the MAREL software.
NYPP Reliability Evaluation Task Force, “Multi-Area Generation Reliability – Phase I Testing Report” (October 22,
1991), ESEERCO Project EP87-28.
PJM Load & Capacity Working Group, “Evaluation of PTI’s Multi-Area Reliability Program MAREL” (November 1993)
© PJM Interconnection 2011. All rights reserved
Page 9 of 136
Current Reserve study Models
Two–Area Model: PJM PRISM
Developed by PJM, the Probabilistic Reliability Index Study Model (PRISM) is owned and maintained by
PJM – with PJM Members having the rights and ability to review its technical attributes. PRISM is the
primary modeling tool used for conducting PJM’s resource adequacy studies. PRISM is classified as a “twoarea” model that simulates the PJM RTO and areas adjacent to PJM’s footprint, called the “World”.
The modeling, core calculations, and underlying techniques used by PRISM are based on the GEBGE
model. PRISM was developed on a Web-based platform. GEBGE remains as a legacy system based on a
FORTRAN platform.
PRISM calculations use a cumulative probability function involving numerical techniques that combine
(convolve) a probabilistic load distribution with a probabilistic resource distribution. A given load distribution
point is “looked up” in the available generation array, the cumulative probability distribution table, resulting in
a loss-of-load value that – when greater than zero (10-11), is weighted by the appropriate probability for this
numerical load occurrence.
PRISM can model either a single region or two regions can be modeled together linked by a chosen tie size.
The interface tie size (or “pipe” size) represents aggregate transmission facilities capabilities at peak
conditions between the study regions.
PRISM requires a forecast probabilistic representation of the daily peak load and generation resources. The
tie size can be considered probabilistically, with PRISM enhancements needed to make this an automated
process. The numerical methods used to implement the LOLE calculations using the cumulative probability
distributions are considered not practical for more than six areas14.
Due to its use of statistical parameters and application of probability theory using numerical methods,
PRISM is considered a planning tool – and not well-suited for use in operational assessments.
The ISO-NE planning staff has used the similar, single-area Westinghouse Model for LOLE assessments in
the ISO-NE region. (Appendix C-3 includes a description of ISO-NE’s Westinghouse model.)
Probabilistic resource adequacy criteria require models to determine compliance with the criteria, while
deterministic criteria require simple event tracking and measuring. Loss-of-load expectation measures the
probability that the load will exceed the available generating capacity. Load, used in resource adequacy
assessments, can be modeled in several ways:

Model all hourly loads deterministically, or 8,760 hours per year. The resultant metric is termed
loss-of-load expectation (LOLE), and is expressed in either hours/year, or in days per year. A
days-per-year metric can be calculated with a Monte Carlo model. In such a model, a Monte Carlo
replication results in one of two outcomes: either load lost in a day or it is not lost. The sum of the
days per year that load is lost over all 8,760 hours is represented by LOLE in days per year.

Models that compute LOLE with all hourly loads, typically a deterministic set of values, can also
compute a system’s EUE which represents the summation of the expected amount of load lost for
each hour of the year.

Model only daily peak loads. Usually just week-day peak loads are modeled, or 260 hours per year
(52 weeks x 5 days per week x one peak hour per day).

All daily peak loads (365 days per year) can be modeled, if there is non-trivial risk on the
weekends. Assessments are needed to determine if weekday risk is appropriate for modeling the
entire week’s risk.

LOLE is related to the Loss of Load Probability (LOLP) but is different. The LOLE, a mathematical
expectation term, is explained further in reference number 22, showing the relationship of LOLE to
14
NYPP Reliability Evaluation Task Force, “Multi-Area Generation Reliability – Phase I Testing Report” (October 22,
1991), ESEERCO Project EP87-28 – See pages S4-S6.
© PJM Interconnection 2011. All rights reserved
Page 10 of 136
LOLP.15 However in several technical discussions these two terms are used interchangeably. Care
should be taken in comparing this paper’s use of these technical terms with a more general use in
other papers.

The term Loss of Load Hour (LOLH) is the probability that the peak load will exceed the capacity in
a single hour.

The Peak Load Ordered Time Series (PLOTS) load model is based on the hourly loads for the
regions in the Study, typically using about 12 years of history. Hourly loads are required for the
entire region. The historic hourly loads are transformed into a set of forecast, probabilistic values
having a weekly mean and standard deviation.16 These forecast, probabilistic values are used for
each week in the Study.
Multi-Area Model: GE-MARS
The Multi-Area Reliability Simulation (MARS) program is owned and licensed by the General Electric
International, Inc. (GEII), and is sometimes referred to as “GE-MARS”. MARS’ core calculations are based
on sampling using the proven Monte-Carlo technique. In this, a random number is used to seed the
calculations with many replications performed to satisfy a chosen solution criteria of standard error (based
on individual replications).
MARS performs a chronological hourly simulation of the system, comparing the hourly load demand in each
area to the total available generation in the area, which has been adjusted to account for planned
maintenance and randomly occurring forced outages. Typically a solution option is chosen to consider only
the daily peak load (Maximum Hourly integrated value). Areas with excess capacity will provide emergency
assistance to those areas that are deficient, subject to the transfer limits between the areas.
MARS utilizes many tables (50) for discrete entries involving many modeling specifics for various categories
including: general solutions, interfaces, loads, maintenance for both load and generation, modifications to
hourly load and generation shapes, generation resources, emergency operating levels, and reserve sharing
among areas.
MARS’ generation resource modeling can use either a transition state matrix or forced outage rate. The load
model uses a discrete 8,760 hourly load pattern, typically chosen by assessment of historic patterns. MARS
can calculate LOLE, LOLH and EUE. .
MARS modeling can also be utilized with operational assessments – but it is still primarily regarded as a
planning tool that can assess adherence to probabilistic criteria (outage event one day, on average, in 10
years)
MARS calculates, on an area and pool basis, the standard reliability indices of daily and hourly LOLE,
expected unserved energy and time-correlated indices such as frequency and duration of outage. The
program can also calculate the expected number of days per year that various emergency operating
procedures would need to be implemented – by those staff operating the grid.
Typical MARS applications include:






Resource adequacy assessments
Installed capacity (ICAP) requirements
Locational capacity requirements
Benefits of load diversity assessment
Benefits of reserve sharing
Tie-line effectiveness
15
Power System Reliability Evaluation, Mathematical Expectation - page 12, Gordon and Beach, Science Publishers, –
1970 – by Roy Billinton and Reinvent Legacy Software with SAS, the Web, and OLAP Reporting, Appendix B –2007 –
available at: http://www.qlx.com/whitepapers/25-2008.pdf .
16
Re-invent a Legacy System with SAS, the Web and OLAP reporting, November 13, 2007, 2007 Northeast SAS user
group conference – Applications Big and Small, Paper Number 10. http://www.lexjansen.com/nesug/ , Appendix:
Download the code (2369 KB).
© PJM Interconnection 2011. All rights reserved
Page 11 of 136


Expected need for implementing emergency operating procedures (EOPs)
Reliability impact and capacity value of intermittent or variable resources such as wind, hydro,
thermal and solar energy.
A sequential Monte Carlo simulation forms the basis for MARS software. The Monte Carlo method provides
a convenient input specifications, versatility, and easily expandable program that can be used to fully model
varied generation and reserve sharing options. MARS software makes the following reliability indices
available on both an isolated (zero ties between areas) and interconnected (using the input transfer limits
between areas) basis:






Daily LOLE (days/year)
Hourly LOLE (hours/year)
Loss of Energy Expectation (LOEE) (MWh/year)
Frequency of outage (outages/year)
Duration of outage (hours/outage)
Need for initiating Emergency Operating Procedures (EOPs, as days/year and hours/year)
PJM is licensed to use MARS and uses this program to enhance the analytical capabilities of PRISM. PJM
uses the MARS program for developing case sensitivities, coordinating study models and results with
neighboring regions, participating in interregional studies, and performing the Winter Weekly Reserve Target
analysis.
PRISM is used to calculate the LOLE of up to two interconnected systems with a single transfer link. PRISM
is used to calculate the Forecast Pool Requirement (FPR) and the Demand Resource (DR) factor (DR
Factor).
Comparative Strengths
The strengths of PRISM as a two-area convolution model include:

PRISM’s relatively fast speed of calculations

PRISM provides historic benchmarking with Operational experience for the PJM footprint

PRISM has reduced input requirements due to using statistical parameters

PRISM is consistent with PJM Agreements and Operational practices.

PRISM provides confidence in accuracy of technical results directly applied in the
Marketplace.

PRISM promotes efficiency and effectiveness of PJM user staff and PJM support staff.
Mitigates tedious repetitive tasks.
17
The strengths of MARS as a multi-area Monte-Carlo model include:

MARS can model Operation specific actions and events.

MARS can model many areas and incorporate different modeling for each area such as 1)
reserve sharing, 2) invocation of Emergency Operation events, 3) load patterns and
distributions, and 4) transmission path flow coordination.


MARS can determine energy related indices such as LOLH and EUE values.
MARS can model various impacts using hourly assessment calculations and methods,
such as intermittent resources, demand side management programs, and load shedding.
17
PJM Generation Adequacy Analysis: Technical Methods - October 2003 – developed by PJM Staff; available at:
http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/res-adeq/20040621-white-papersections12.ashx. - See Section 2 of this White Paper.
© PJM Interconnection 2011. All rights reserved
Page 12 of 136
Comparative Strengths
The strengths of PRISM as a two-area convolution model include:

PRISM’s relatively fast speed of calculations

PRISM provides historic benchmarking with Operational experience for the PJM footprint 18

PRISM has reduced input requirements due to using statistical parameters

PRISM is consistent with PJM Agreements and Operational practices.

PRISM provides confidence in accuracy of technical results directly applied in the
Marketplace.

PRISM promotes efficiency and effectiveness of PJM user staff and PJM support staff.
Mitigates tedious repetitive tasks.
The strengths of MARS as a multi-area Monte-Carlo model include:

MARS can model Operation specific actions and events.

MARS can model many areas and incorporate different modeling for each area such as 1)
reserve sharing, 2) invocation of Emergency Operation events, 3) load patterns and
distributions, and 4) transmission path flow coordination.

MARS can determine energy related indices such as LOLH and EUE values.

MARS can model various impacts using hourly assessment calculations and methods,
such as intermittent resources, demand side management programs, and load shedding.
18
PJM Generation Adequacy Analysis: Technical Methods - October 2003 – developed by PJM Staff; available at:
http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/res-adeq/20040621-white-papersections12.ashx. - See Section 2 of this White Paper.
© PJM Interconnection 2011. All rights reserved
Page 13 of 136
Complementary Aspects of PRISM and MARS
Throughout this report, the term “complementary” is used to characterize usage of both PRISM and MARS
in reliability studies. As was mentioned in the Background section, the 1993 assessment by PJM’s L&CWG
(the predecessor to RAAS) clearly stated its intention to use a multi-area tool in concert with a two-area
convolution tool. Since that time, the PJM Agreements, Operational procedures and LOLE technical
methods have not significantly changed to alter that over-arching philosophy.
Below is a recap of “complementary” tasks that each program provides to the other:

Complementary tasks that PRISM provides to MARS:
1.
Since MARS can require significant time to reach a result, PRISM can be used to quickly
determine LOLE results. This enables more time for fine-tuning and other analytical work.
2.
The variability of MARS’ Monte Carlo calculation results can be minimized but not eliminated19.
Conversely, PRISM’s non-variable, direct table lookup method can be used to verify the
MARS’ results.
The PJM Agreements and Operational procedures currently in place reflect PRISM results –
and does not require event-driven detailed, operational and transmission values (as required
by MARS). PRISM’s use in planning assessments can reduce the resources required to reach
the desired result (i.e. FPR, DR Factor and CETO values in load deliverability tests).
3.
4.
MARS results are based on operational input data parameters. PRISM can help evaluate,
measure, and assist coordination with operational experience and MARS input parameters.20
5.
The MARS load model parameters are critical in CETO assessment work. Since PRISM uses
a forecast probabilistic load model, it can assist MARS in evaluating the choice in: 1)
deterministic load model parameters, 2) distributions used in the MARS Table, 3) in02 file, 4)
LOD-UNCY and 5) LOD-MTAR. These deterministic distributions, especially for smaller PJM
RTO sub regions, have been identified as additional development efforts for any MARS CETO
assessment work.
6.
PRISM can independently verify and help measure the MARS results. This can be done by
evaluating the solutions on a very detailed basis, and at every solution point used (See
Appendix E).
7.
MARS is the tool used by our neighboring regions and control areas. PRISM can assist these
interregional study efforts by translating MARS modeling and results into the parameters used
in the PJM Planning division assessments.
19
Refer to NYPP Reliability Evaluation Task Force - Phase I Testing Report (October 22, 1991) – See page S-8.
PJM Generation Adequacy Analysis: Technical Methods - October 2003 – developed by PJM Staff; available at:
http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/res-adeq/20040621-white-papersections12.ashx. - See Section 2 of this White Paper.
20
© PJM Interconnection 2011. All rights reserved
Page 14 of 136

Complementary tasks that MARS provides to PRISM:
1.
MARS can assess the single-area World model used in PRISM. (It is important to note that
single-area modeling was tested and evaluated to be appropriate, in the comparing GEBGE
with MAREL section.21)
2.
MARS can assist assessment of intermittent resources such as Wind, Solar and Hydro due to
its various tables and hourly patterns. Such assessments can also include Demand
Resources (DRs). More development work would be needed for these efforts.
3.
Operational issues such as transmission constraints, reserve sharing events, and assessment
of load management events can be performed.
4.
MARS can provide weekly distributions and resulting LOLE for the winter period that is
consistent with the annual PRISM patterns. This function can be used in setting the Winter
Weekly Reserve Target used by the Operation division staff for scheduling unit planned
outages over the 13-week winter period (December through February).
5.
MARS can independently verify and benchmark PRISM results. This can be done by
evaluating the solutions on a very detailed basis, and at every solution point used. See
Appendix E.
6.
MARS is used extensively by PJM’s neighboring regions. Using MARS therefore helps
facilitate interregional coordination and planning efforts.
7.
MARS can assist in evaluation of the impact of high ambient conditions on generation
resources by using its UNT-DERT table.
8.
MARS can be used to comply with any new, forth coming requirements, per the NERC
Planning Committee directives. MARS can report the identified new metrics, LOLH and EUE.
In reporting for EUE, the forecast energy, per the annual PJM load forecast report and other
specifics will need attention. These two metrics have been historically not used and reported
within the electric Industry for bulk power systems.
Ongoing LOLE Assessment Work

Reserve Requirement Study (RRS) – PJM Staff annually conducts a Reserve Requirement Study
(RRS). These efforts involve seven or more staff and several professional disciplines for a
successful completion. These efforts are required to:
o
Establish the needed Reliability Pricing Model (RPM) Market parameters: the FPR,
and associated Demand Resource DR Factor (DR Factor)
o
Perform load deliverability tests for PJM’s Regional Transmission Expansion Plan
(RTEP)
o
Demonstrate compliance with RFC Standard BAL-502-RFC-02.
The RRS also determines the Winter Weekly reserve target used in Operations. These
assessment activities serve to coordinate efforts between the Planning, Operations and Market
Divisions. The efforts for the RRS include:
o
21
A data completion effort that is developed for each study showing resources,
estimated hours, duration, and target completion dates (Using a tracking
spreadsheet).
See the related discussion in the “Evaluation of PTI’s Multi-Area Reliability Program (MAREL)” by the Load and
Capacity Working Group (November 1993).
© PJM Interconnection 2011. All rights reserved
Page 15 of 136


22
o
A quality data source is established before any analysis can be started. The data
sources involve both the internal and external regions in the model, have more than
20 specifics tasks, involve the LOLE modeling Team, the load forecasting Team, and
the GADS administrators.
o
PJM’s Information Technology Services (SAS, Oracle, Web (Java), Database
Administrators, and Network systems) and Operations staff assist supporting
compilation of the generation model with involvement by the Generator Owners to
review their units’ data.
o
These efforts can take approximately 480 Full Time Equivalent (FTE) hours to
complete and typically span over three to four months. Experienced PJM Resource
Adequacy and Planning Department staff oversees and perform the primary review of
these efforts.
LOLE Analysis – The RRS analysis to be performed, after the data has been finalized, is
discussed and assignments agreed to among the LOLE analysis Team.
o
PJM’s Resource Adequacy Planning (RAP) Department staff performs these studies,
coordinating with other Planning, Markets and Operation Division staff as results and
observations become apparent.
o
The RAAS and PC are part of this initial review. The intent is to quickly identify any
issues that are seen as significant drivers so that a thorough assessment and
reporting can be made.
o
Typically there are over fifty separate assessment items involved in the analysis,
resulting in a one hundred page report (see 2010 RRS reference).
o
These analysis efforts and related reporting can take approximately 680 hours and
span over three to four months, primarily performed by four RAP staff members.
Coordination and review with the RAAS and PC are important steps in this process.
Capacity Emergency Transfer Objective (CETO) Assessments - Related analysis efforts also
involved CETO assessments for the 24 Locational Deliverability Areas (LDAs)22.
o
CETO assessments use a given RRS model as the basis for its assessment work.
These efforts involve the LOLE Team, Load Analysis Team (LAT), Transmission
Planning Staff, and Generation Interconnection Queue staffs.
o
The CETO efforts are performed for two primary efforts: 1) To address the
requirements of the RPM Market Place and 2) To address the needs of the RTEP
process overseen by the PJM Transmission Expansion Advisory Committee (TEAC)
stakeholder group.
o
The amount of CETO related assessment work has increased six fold in recent years
and continues to increase.
o
The CETO assessments currently are performed during the entire calendar year, with
concentrated efforts in the December to February time periods, and recently the June
to August time periods.
o
There are at minimum 96 required CETO assessments in each year, for various
delivery years. There can also be ad-hoc assessment work based on generation
termination requests and state commission inquiries.
o
The approximate total efforts are 730 Hours, for the LOLE Team, but other
department staffs are affected as well. The coordination with the other departments
and divisions is critical in the LOLE Team’s ability to perform the required
As discussed in the Reliability assurance Agreement (RAA), Schedule 15 and listed in PJM Manual 14B, Study Area
definitions – Zonal and Global (pages 49-50).
© PJM Interconnection 2011. All rights reserved
Page 16 of 136
assessments with high data quality and organization being a significant impact in the
process.
The Resource and Cost Assessment section of this paper provides further details concerning the LOLE
assessment efforts.
Model Calculation Processes
PRISM Calculation Processes
The calculations used in PRISM use forecast statistical parameters for both the load and generation
distributions. PRISM simulates “the joining” of a given load distribution point with the associated generation
distribution point (Click here for joining calculation reference). This joining is performed by a “lookup” table
approach using the cumulate probability (Cum Prob) distribution of the generation distribution.
The distributions used are both forecast and probabilistic. The first step is to build the cumulative probability
23
array, based on the individual generation unit forced outage rates. This calculation is widely documented .
The creation of this Cum Prob table is the most calculation-intensive aspect of the method used in PRISM,
done for each week in the model. The table is created using the binominal expansion method11. Once the
Cum Prob distribution pattern is determined for the generation model, the load model distribution is used to
look up the associated Loss-of-Load Probability (LOLP). This lookup is when the actual load distribution is
“convolved” with the generation distribution.
A graphical depiction of the numerical look-up method is shown in Figure 1, as an illustration only (again the
“joining” reference). The red bars of Figure 1 depict when there is risk of load exceeding the available
generation; this results in a loss-of-load event with an associated probability. Each load level has a defined
probability of occurrence and the red bars occur at significantly high loads, but with a low probability of
occurrence.
In the context of PJM Adequacy calculations the term “convolution” is used to depict the relationship
between two input distributions of load and capacity and the resulting loss of load distribution. In terms of
formal Mathematics, “convolution” (as used in Digital Signal Processing) does not have the same meaning
as convolution used in this report. However, science and engineering problems approach this single concept
from two different directions: 1) by considering a system in terms of what its impulse response looks like or
2) considering the system as a set of weighting coefficients. Familiarity with both views allows one to toggle
between directions one and two24.
Convolution plays a major role in both the capacity table approach and the load approach. The convolution
method illustrated for both approaches is called the recursive method.25 In this paper convolution is used to
convey to a technical audience, but yet non-mathematician audience, the relationship between two input
distributions of load and capacity and the resulting loss of load distribution. The load model derived from the
hourly load (Daily Peak) curve is thereby “convolved” (joined) with the generation system model for
computing the LOLE26.
23
Refer to Roy Billinton’s book, “Power System Reliability Evaluation”, Gordon and Beach, Science Publishers, or a
simple example shown in Appendix B of the Paper titled “Reinvent Legacy Software with SAS, the Web, and OLAP
Reporting” available at the following link: http://www.qlx.com/whitepapers/25-2008.pdf .
24
The Scientist and Engineer's Guide to Digital Signal Processing , By Steven W. Smith, PhD, copyright © 1997-2006
by California Technical Publishing – http://www.dspguide.com/pdfbook.htm
25
Dr. James McCalley’s course notes, module PE.PASU19.5 on Generation adequacy evaluation, Convolution
techniques, item U19.7.3 on page 43. http://www.ee.iastate.edu/~jdm/ee653/ee653schedule.htm
26
Dr. Chanan Singh, course notes: Electrical Power System Reliability, part3 Discrete Convolution Method, page 30,
copyright 1995, http://www.ece.tamu.edu/People/bios/singh/coursenotes/part3.pdf
© PJM Interconnection 2011. All rights reserved
Page 17 of 136
Figure 1 – Load distribution and Cumulative Probability Capacity Distribution
depicting PRISM calculations
Illustration of PJM RTO Convolution - Daily (during Peak Week)
Load Probability & Load LOLE
0.18
1.00
14
0.90
13
15
0.80
0.14
12
0.12
0.70
16
0.60
0.1
0.50
11
0.08
17
0.40
0.06
0.30
10
18
0.04
Cumulative Probability of Generatioon Availability
Daily Peak Load Probability of Occurance
0.16
0.20
09
0.02
19
0.10
08
01
02
03
04
06
05
20
07
21
0
Megawatts ( Load and Generation ) Zero to
Daily Load Probability
Load LOLE
22
23
24
25
26
0.00
∞
Generation Cumumulative Probability (Right Y Axis)
Figure 1 shows that as the load (in megawatts) increases, there is higher likelihood for loss of load (i.e. loss
of load occurs at much higher values than the mean). For each level of probable load (shown as the blue
bars) there is an associated cumulative probability of available generation (black dashed line). The
calculations mirror a continuous distribution (probable load), approximated by the blue bars.
Figure 2 – Extreme high loads: Detail of green oval
Illustration of PJM RTO Convolution - Daily (during Peak Week)
Detail: Load Probability & Load LOLE for "Extreme High" Loads
0.20
0.18
Daily Peak Load Probability of Occurance
0.16
0.14
19
0.015
0.12
0.10
0.01
0.08
20
Bars 21 - 26 reflect Equal
values for LOLE and Load
Probability.
0.06
18
0.005
0.04
21
15
16
0
0.02
22
17
23
24
25
Megawatts ( Load and Generation ) Zero to ∞
Daily Load Probability
Load LOLE
© PJM Interconnection 2011. All rights reserved
Generation Cumumulative Probability (Right Y Axis)
Page 18 of 136
26
0.00
Cumulative Probability of Generatioon Availability
0.02
Figures 1 and 2 are for illustration, but 21 points are used in the calculations. The number of points used
was due to practical considerations of speed and accuracy. Therefore twenty one points are used for each
daily peak lookup.
The red bars indicate when a loss of load state (LOLE) occurs –when the load is excessively large (which
rarely occurs and shown in green oval –see bars 19 and beyond). The cumulative probability of available
generation is low at these excessively large loads. The daily peak load probability of occurrence scale is
shown on the left Y axis. Example calculations used to determine the load model lookup value (into the
cumulative probability array), is shown in Table 1.
For the extremely high-load levels encircled by the green oval, the red LOLE bars increase because of the
higher cumulative probability of unavailable generation –at least until a certain higher load value is reached.
Figure 2 shows a magnified view of the green oval area, extreme load that have LOLE states. After
reaching an LOLE peak (in 19th bar from left), the red bars taper off due to the diminishing likelihood of
higher loads. (I.e. Such very high loads do not often occur while the generation probability (of unavailability)
saturates at a value of 1.0 –generation probability (of availability – dashed line) declines as megawatts
increase.
As generation unavailability saturates (approaching a value of 1), generation resources will not be able to
serve that load level (generation availability approaches zero). Even though there is a much greater risk of
generation unavailability as load increases, the LOLE is reduced because of the very small chance of that
higher load occurring.
It is important to note that the blue bars of this graph include PJM RTO loads up approximately 200,000
27
MW. There is a higher risk of generation unavailability (until saturation when a total generation is less than
load) as the load increases, but a lower likelihood that a higher load will occur.
Figure 3’s probability of per-unit load column (last column on right) correlates with the left Y axis in Figure 1
(Probability of occurrence). Figure 3’s Daily Peak Load distribution (MWs) column is analogous to the X
axis of Figure 1. Figure 3’s values are used to perform a table look-up into the generation’s cumulative
probability array to determine the probability of the available generation. This probability of generation
unavailability (reciprocal of availability –right Y axis), is multiplied by that load’s probability of occurrence (left
Y axis) to calculate the final LOLE for that day’s load.
For a given daily peak in a given week, using two summer months’ peak week (July and August) for these
examples, the following data can be derived as Load Model Lookup Values (Figure 3A & 3B):
Figure 3’s 21-point calculated load values use the equation shown at the bottom of Figure 3. Each Daily
Peak Load distribution value is used to “lookup” the associated cumulative probability in the generation
distribution (i.e. the load is thus “joined” with the generation). By multiplying the generation’s cumulative
probability value by the associated load probability the LOLE value is determined. If an LOLE value is less
than 10 -11 it is assumed to be zero. The daily LOLE, from the 21 points (points 4 through 24 in Figures 3A &
3B), are summed for all weeks in a delivery year to determine the annual LOLE.
The calculation of the available generation cumulative probability is performed using a binomial expansion
function. An example of the binomial expansion is shown on the second page of reference number 12 , and
Appendix B of reference number 15 .
27
PJM 2010 Load Forecast Report shows a 50/50 peak for the 2014 forecast delivery year of
163,093 MW.
© PJM Interconnection 2011. All rights reserved
Page 19 of 136
Figure 3A – Load Model Lookup Values for Daily Peak distribution- JULY
Key Parameters:
Final Week Peak Frequency (WKPKFQ) Mean =
1.0000
Final WKPKFQ Standard Deviation =
Total Sigma =
Forecast Error Factor (FEF) =
0.0703
0.0710
0.01
Daily EWM (July Peak Week) =
Daily Per-Unit EWM -of annual peak (July Peak Week) =
Daily Per-Unit Mean (on annual peak basis -Bar # 14) of EWM =
1.08173
1.000000
0.924448
Annual Peak Load (MW) =
EWM (for each day) =
Daily Mean Load (Bar # 14) =
163,093
1.000000
0.924448
1.0000
MW
163,093
150,771
EWM = Expected Weekly Maximum : Used for Daily Peak in that week.
Bar #
Probability of
Per-Unit Load
(given)*
# Sigma
from Mean
(given)
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
0.000033
0.000145
0.000638
0.002351
0.007273
0.018940
0.041400
0.076080
0.117490
0.152480
0.166340
0.152480
0.117490
0.076080
0.041400
0.018940
0.007273
0.002351
0.000638
0.000145
0.000033
-4.20
-3.78
-3.36
-2.94
-2.52
-2.10
-1.68
-1.26
-0.84
-0.42
0.00
0.42
0.84
1.26
1.68
2.10
2.52
2.94
3.36
3.78
4.20
Annual
Daily Per-Unit
Peak Load
Mean of EWM
(MW)
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
0.924448
<= Original
WKPKFQ value
Total
Sigma
Per-Unit Load
Daily Peak Load
Distribution (MW)
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.0710
0.648844082
0.676404499
0.703964917
0.731525334
0.759085751
0.786646168
0.814206585
0.841767003
0.869327420
0.896887837
0.924448254
0.952008671
0.979569089
1.007129506
1.034689923
1.062250340
1.089810758
1.117371175
1.144931592
1.172492009
1.200052426
105,811
110,307
114,803
119,299
123,795
128,291
132,787
137,283
141,779
146,275
150,771
155,267
159,763
164,259
168,755
173,251
177,747
182,243
186,739
191,235
195,731
Sigma from Mean Load Calculation:
Daily LOAD = Annual Peak Load (MW) x Per-Unit Mean EWM x (1 + (# Sigma from Mean) x Total Sigma)
Per-Unit Load = (Mean of EWM * (1 + (# of Sigma from Mean) X Total Sigma)
* Probability of Per-Unit Load values are given and taken from the 21-point Standard Normal distribution table.
The values in this example (Figure 3A & 3B) are the final adjusted mean and standard deviation
values, after adjustment to match the forecast monthly peak shape. The fifteen step calculation
process for the original WKPKFQ mean and standard deviation values is shown in Appendix A of
reference number 15.
© PJM Interconnection 2011. All rights reserved
Page 20 of 136
Figure 3B – Load Model Lookup Values for Daily Peak distribution- AUGUST
Key Parameters:
Final Week Peak Frequency (WKPKFQ) Mean =
0.9529
Final WKPKFQ Standard Deviation =
Total Sigma =
Forecast Error Factor (FEF) =
0.0719
0.0726
0.01
Daily EWM (August Peak Week) =
Daily Per-Unit EWM -of annual peak (August Peak Week) =
Daily Per-Unit Mean (on annual peak basis -Bar # 14) of EWM =
1.03250
0.954494
0.880876
Annual Peak Load (MW) =
EWM (for each day) =
Daily Mean Load (Bar # 14) =
163,093
0.954494
0.880876
0.968141
MW
155,671
143,665
EWM = Expected Weekly Maximum : Used for Daily Peak in that week.
Bar #
Probability of
Per-Unit Load
(given)*
# Sigma
from Mean
(given)
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
0.000033
0.000145
0.000638
0.002351
0.007273
0.018940
0.041400
0.076080
0.117490
0.152480
0.166340
0.152480
0.117490
0.076080
0.041400
0.018940
0.007273
0.002351
0.000638
0.000145
0.000033
-4.20
-3.78
-3.36
-2.94
-2.52
-2.10
-1.68
-1.26
-0.84
-0.42
0.00
0.42
0.84
1.26
1.68
2.10
2.52
2.94
3.36
3.78
4.20
Annual
Daily Per-Unit
Peak Load
Mean of EWM
(MW)
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
163,093
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
0.880876
<= Original
WKPKFQ value
Total
Sigma
Per-Unit Load
Daily Peak Load
Distribution (MW)
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.0726
0.612444339
0.639287529
0.666130718
0.692973908
0.719817097
0.746660287
0.773503476
0.800346666
0.827189856
0.854033045
0.880876235
0.907719424
0.934562614
0.961405803
0.988248993
1.015092182
1.041935372
1.068778562
1.095621751
1.122464941
1.149308130
99,858
104,239
108,620
113,000
117,381
121,762
126,142
130,523
134,903
139,284
143,665
148,045
152,426
156,807
161,187
165,568
169,949
174,329
178,710
183,090
187,471
Sigma from Mean Load Calculation:
Daily LOAD = Annual Peak Load (MW) x Per-Unit Mean EWM x (1 + (# Sigma from Mean) x Total Sigma)
Per-Unit Load = (Mean of EWM * (1 + (# of Sigma from Mean) X Total Sigma)
* Probability of Per-Unit Load values are given and taken from the 21-point Standard Normal distribution table.
The final week peak frequency (WKPKFQ) mean and standard deviation values are determined
in a separate process. These final values are an input into the calculation example shown in
Figure 3. Part of the determination of these weekly values, is an adjustment to match a
forecasted monthly peak shape, given in the PJM load forecast report. The original WKPKFQ
mean and standard deviation values28 are adjusted so that the final WKPKFQ monthly peaks
match the forecast monthly peak shape. The 52 week values, for the non monthly peaks, keep
their original ratio relative to their adjusted monthly peak. The final and original WKPKFQ values
are shown at the top of Figure 3A for comparison purposes.
28
2010 PJM Reserve Requirement study (RRS), Table II-2. Week number 12 is the August Peak.
Week number 10 is the July peak (annual peak).
© PJM Interconnection 2011. All rights reserved
Page 21 of 136
The Total Sigma calculation from PJM Manual 20 is:
Total Sigma = (Standard Deviation 2 + Forecast Error Factor 2)0.5
The process for the calculations to determine an IRM for compliance to the Standard, are shown in Figure 4.
PRISM uses a proxy for adjusting the capacity distribution, assuming that an adjustment in the load level is a
good proxy for individual unit adjustments to the capacity. Assessment in the NPCC CP-8 WG demonstrated
that adjustments to load gave almost identical changes as equal adjustments to capacity.29
The load shape is adjusted for the entire deliver year in this process until an LOLE (sum of individual daily
peak LOLPs) is at a Reliability Index (RI) of 10 (RI = 1 / LOLE).
Figure 4 – Installed Reserve Margin (IRM) Weekly Load Profile
LOLE Studies
Installed Reserve Margin
The Peak Load line is statistically manipulated until the
1-day-in-10-years criterion is met.
3/28/07
©2007 PJM
©2008 PJM
29
www.pjm.com
41
1
Northeast Power Coordinating Council Tie Benefits Methodology, by Glenn Haringa and Philip Fedora, November 5-6,
2008, Best LOLE Practices meeting held at California ISO offices, Agenda item 8. See slide 12, last bullet.
© PJM Interconnection 2011. All rights reserved
Page 22 of 136
MARS Calculation Processes
The publicly available solution method and techniques are summarized in Appendix D, as documented by
GE Power Systems. The solution information is given further details from the complete information in
Appendix D – and is based on the RAAS drafting team’s interpretation, observation and feedback from GE
Power Systems Staff, and participation in other LOLE Industry groups’ efforts (See Interregional
Assessments section).
MARS is a Multi-area generation system reliability model based on sequential Monte Carlo simulation. It was
developed in the late 1980’s by a project initiated by the New York Power Pool (NYPP) (refer to the Early
Reserve Study Models section of this report.) MARS uses a transportation algorithm to model flows
between areas. A transportation model requires separate assessment of the interconnected facilities and
operational practices to determine the interface transfer (pipe) size limits between the areas in the model.
The reliability calculations are performed in MARS by determining the capacity available each hour from
each unit on the system based on three primary model characteristics:
1.
2.
3.
Unit rating and capacity states
Scheduled planned outages
Random forced outages
The objective is to determine the area margins, details of which can be listed in output file 10, for each hour
which can be expressed as:
Margin = Capacity - Load
This algorithm’s solution order is:
1)
2)
3)
4)
5)
Accumulate statistics for isolated indices
Calculate flows between areas (if needed) and resulting area margins
Accumulate statistics for interconnected indices
Proceed to the next hour
Proceed to next replication
A further overview of the algorithm’s order for solution is shown in Figures 5A, 5B and 5C, a flow chart of the
MARS processing of the Master Input File (MIF). Note that the typically solution option for large system
studies in step 4, is to proceed to the next daily peak.
The solution step to calculate emergency assistance and resulting area margins, in Figure 5c, in the solution
process is the most computational intensive step. If dynamic interface limits are specified they are evaluated
in this step with a potential to significantly increase the solution time.
The MARS solution process is performed as shown in Figures 5A - 5C, on a given model basis. Typically a
chosen seed number is given for the solution results to be repeatable for auditing and verification purposes.
In this system- wide solution for large regional models, the default is to share assistance between areas on
an equitable basis. This approach mitigates individual areas’ LOLE states, without priority. This avoids
deciding which areas have priority over others in receiving assistance. MARS allows specifying a priority list
for allocating known assistance priorities among the areas that comprise the large regional model (Table
RES- PRIO).
The solution is unique for each model, based on the Master Input File’s Table entries and specifications.
Figures 8a- 8C show how the resulting Standard Error (SE) reduces by running more replications. The LOLE
metric result, for any specified Margin state, is consistent with how the SE improves with more replications.
The right (secondary) axis on Figures 8A-8C show the LOLE metric after replications of 1, 50, 100, 200, 500,
1000, 1500, and 2000 –demonstrating how it “converges” to a final unique solution.
© PJM Interconnection 2011. All rights reserved
Page 23 of 136
Figure 5A – Overview MARS solution
method
© PJM Interconnection 2011. All rights reserved
Figure 5B – Monthly
data preparation loop in solution method
Page 24 of 136
Figure 5C – Overview of Monte Carlo simulation
The Monte Carlo simulations usually specify a “random number seed” to start the solution process – with
defined convergence criteria for the Standard Error. The seed number allows for repeating the exact solution
in future sensitivity assessments.
The Standard Error is the average difference of all simulations performed. By making many simulations, this
solution method tends to saturate at a given level of Standard Error. If the given level of Standard Error is at
or below the convergence criteria, the simulation is deemed fully solved and all output reporting is generated
for the simulations.
A graphical depiction of the solution process is shown in Figures 5A-5C with an example resulting
Probability Density Function (PDF) of this process shown in two diagrams in Figure 6A and 6B. The
Standard Error of the individual replication results, sum of square average from mean, is shown in Figures
8A-8C.
© PJM Interconnection 2011. All rights reserved
Page 25 of 136
Figure 6A – Sample Probability Density Function (PDF)
Figure 6B – Example PDF and associated Cumulative Density Function (CDF)
Histogram
100
120.00%
90
100.00%
80
Frequency
70
80.00%
60
50
60.00%
40
40.00%
30
20
20.00%
10
0
0.00%
Bin
Frequency
Cumulative %
Figures 6A and 6B depict the PDF of the model in the Monte Carlo Simulations. They are a function of a
continuous variable, the generation outage set, such that the integral of the function over a specific region
yields the probability that its value will fall within the region.
© PJM Interconnection 2011. All rights reserved
Page 26 of 136
MARS models and simulates Load Forecast Uncertainty (LFU) by use of the LOD-UNCY table. This table
allows for each daily peak, 7 peaks each week, and 10 values to represent the distribution of the daily peaks
so that all possible loads, with the applicable probability of occurrence, assessed in the calculations. This
allows the assignment of probabilistic characteristic to the 8760 deterministic hourly load values defined in
the in02 file.
Figure 7 shows an example of this LFU and the associated calculated LOLE. Industry groups have identified
that the high probability risk, for models in North America, are in the high load values that have a low
probability of occurrence. (See Appendix C3 for further details concerning discussion of this topic.)
Figure 7 – Impact on Load Forecast Uncertainty (LFU) on LOLE
Figures 8A through 8C show the Standard Error (SE) achieved over a 2000 replications. These results are
from an Inter-regional Industry model that had over 25 areas, which included dynamic interface transfer
limits, thermal unit capacity derations, and seven load forecast uncertainty probability values, and six
Emergency Operation Procedure levels.
© PJM Interconnection 2011. All rights reserved
Page 27 of 136
Figure 8A – Standard Error (SE) of Monte Carlo Simulation
0.500
0.006
0.450
Standard Error (SE)
0.350
0.004
Trial 38
0.300
Trial 46
Trial 70
0.250
0.003
Tolerance
Trial 74
0.200
Trial 45
LOLE-Trial 38
0.002
0.150
0.100
Loss-of-Load Expectation (LOLE)
0.005
0.400
0.001
0.050
0.000
0.000
0
200
400
600
800
1000
1200
1400
1600
1800
2000
Number of Replications
The vertical axis(Y axis) is the standard error while the horizontal (X axis) is the number of replications.
Figure 8B – Standard Error of Monte Carlo Simulation – Low SE
0.09
1.800
0.085
0.08
1.750
0.07
Standard Error (SE)
0.065
0.06
0.055
1.700
Trial 51
0.05
Trial 52
0.045
Trial 72
0.04
Trial 75
0.035
Trial 76
1.650
Tolerance
0.03
LOLE -Trial 75
0.025
0.02
1.600
0.015
0.01
0.005
0
1.550
0
200
400
600
800
1000
1200
Number of Replications
© PJM Interconnection 2011. All rights reserved
Page 28 of 136
1400
1600
1800
2000
Loss-of-Load Expectation (LOLE)
0.075
Figure 8C – Standard Error of Monte Carlo Simulation – High SE
1
0.001
0.001
Standard Error (SE)
0.8
0.7
0.001
0.6
Trial 47
LOLE
0.5
0.001
0.4
0.000
0.3
0.2
0.000
Loss-of-Load Expectation (LOLE)
0.9
0.1
0
0.000
0
200
400
600
800
1000
1200
1400
1600
1800
2000
Number of Replications
The Standard Error from the mean is defined as the standard deviation of the distribution of sample
averages. From the Central Limit Theorem it is known that sample averages are approximately normally
distributed. The Mean of the normal distribution of sample averages is µ, which is also the mean of the
population. The standard deviation of the distribution of sample averages is [σ/n]1/2, where σ
is the standard deviation of the population. All iterations are compared to this average mean value, with the
mean changing as the simulations are performed until the Standard Error value is at or below the define
criteria. MARS defines a default Standard Error value of 0.05 with some Industry studies using a value of
0.025.
Note that even after 2000 replications, the Standard Error is still on a trend downward for Figure 8C. In this
model there were 76 various trials or area summaries that could be assessed concerning the solution criteria
for the best value to use for Standard Error. There can be a large variation between the various areas’
Standard Error, compare the standard Error values between Figures 8A-8C. This indicates that proper
assessment and Engineering judgment needs to be used to specify the solution criteria and to choose which
area should be used to define a successful final solution.
It is evident from observing Figures 8A – 8C that there is a “law of diminishing returns” for running many
MARS replications. The accuracy of the LOLE results may not be improved by performing thousands of
replications –in this case more than 2000 replications. Engineering judgment and experience must be used
to choose the best trade off of computer resources, resource cost and accuracy of results.
The right (secondary) axis on Figures 8A-8C show the LOLE metric after replications of 1, 50, 100, 200, 500,
1000, 1500, and 2000 –demonstrating how it “converges” to a final unique solution. An orange line was
generated, for the LOLE values of Figures 8A-8c, between these points for illustration purposes only.
© PJM Interconnection 2011. All rights reserved
Page 29 of 136
Figure 9 – Confidence Interval
Figure 9 is a modified version of a Confidence Interval graph that regularly appears in the NYSRC’s Annual
IRM Study30. (This graph represents a Standard Error of 0.025 occurring after 2,173 iterations.)
The Monte-Carlo Simulations can be summarized by Figures 8A-8C. This is based on a given Standard
Error to define the convergence criteria and typically requires many thousands of replications to reach a
solution. Figure 9 depicts the Confidence Interval of the thousands of replications performed. Typically
LOLE results are measured against a probabilistic measure, in this case 0.1 outage/delivery year. This
outage rate is reported at a defined Emergency Outage Procedure level.
The 0.1 outage per delivery year is viewed as a practical equivalent of the value of 1 loss of load event, on
average, every 10 years. It comes from a defined Industry Standard (BAL-502-RFC-02) that is viewed as a
requirement to demonstrate compliance. Further details of the source of this 1 day in ten years are
discussed in the background section of this paper.
An advantage of this calculation approach is that by changing the assumptions used for the discrete event
driven parameters, in the various input tables, one can observe the change in the confidence level and
LOLE results. The confidence level for simulations is generally a known technical measure and understood
for reporting purposes. The many replications are needed as the mix of generation resources are varied in
each replication, to value the loss of load probability for the value of load chosen in the input tables (LODDATA, LOD-MTAR, LOD-UNCY). These replications are a function of the continuous variable, the
generation outage set. The deterministic nature of the load necessitates the use of the Monte-Carlo
simulation method to choose the generation outage set, iterating into the solution.
30
NYSRC-ICS; “New York Control Area Installed Capacity Requirements for the Period May 2010 through April 2011 Technical Study” Report (December 4, 2009) - New York State Reliability Council, LLC (NYSRC) Installed Capacity
Subcommittee (ICS) – see page 25
© PJM Interconnection 2011. All rights reserved
Page 30 of 136
Figure 10 – Area Diagram
MRO-USA
Independent Electric System
Operator
1570
1200
New York Independent
System Operator
750
7500
Western
MA
6500
RFC
2660
PJM
7500
Eastern
MA
300
5700
PJM
7500
810
Rest of PJM
3000
1015
1500
PJM
Central
MA
8400
8400
MA = Mid-Atlantic
2000
1700
2330
200
3000
ReliabilityFirst
Other ECAR &
MAIN
1500
1350
1200
PJM
5500
1000
Non-PJM
1525
NYISO
140
Midwest Reliability
Organization
1000
Independent System
Operator of New
England
IESO (Ontario)
90
550
ISO-NE
1650
1300
1700
1500
1500
1600
ENTERGY
TVA
Tennessee Valley
Authority
2500
1000
3200
3500
© PJM Interconnection 2011. All rights reserved
Page 31 of 136
2500
Southern
2300
Non-PJM
VACAR
Virginia Carolinas
1100
550
February 16, 2009
2000
3500
2500
Figure 10 shows an example diagram of possible individual areas that could be modeled. Note that the New
York (NYISO), New England (ISO-NE) and Ontario (IESO) areas can be represented in further detail (at
least 10 subareas comprise each region shown). The calculations are done on an area basis, for each load
level specified, at each EOP level on an hourly basis. If this determines that an area needs assistance to
avoid an LOLE state for this hour, assistance is considered from the other areas in the model. See Figures
5A- 5C for further details on the solution process.
The assistance, determined in the calculation process to mitigate LOLE states, which the surrounding areas
can give, is a function of: 1) the individual given area of interest’s LOLE state (need for assistance) and 2)
the resource allocation assigned to each of the surrounding areas. The LOLE calculations attempt to
mitigate any area of interest’s LOLE states, using the defined resource allocation assistance (i.e. RES-PRIO
and RES-POOL). See Appendix F for further discussion concerning MARS input Tables.
Figure 11 – Expected need for Emergency Operating Procedures (EOP)
Figure 11 shows possible invocations of six EOP levels as the IRM is varied. As IRM increases, the number
of times a given EOP is invoked decreases. These are representative values only, not based on any model
or assessment performed. However this type of reporting is typical for MARS LOLE calculations and
assessments. The reduction in invocation values, from Step 1 (Operating Reserves) to Step 6 (Appeals for
Public Curtailment), is representative of how the system is controlled in the day to day operation of the bulk
power grid by the Control Room Dispatcher Staff. There are many more Control Room Dispatcher actions
than the limited number and aggregated steps shown for LOLE simulation.
© PJM Interconnection 2011. All rights reserved
Page 32 of 136
Model Calculation Comparisons: Observations
31

Both PRISM and MARS can perform their calculations on daily peaks. They use the hourly load
data for this dally peak, the maximum of the hourly integrated value for that day, used in the LOLE
calculations.

MARS can perform calculations, for each of the EOP levels, for each hour in its model. This is a
recent option, available in the current MARS version 3.00.

During 2002-2004, PJM compared the underlying generation distribution patterns used in the
calculations. MARS uses a transition state matrix distribution while PRISM uses a two-state
distribution. It was shown that the transition state matrix distribution were essentially identical to the
two state distribution as used in the calculations. This assessment was for the Mid-Atlantic region.

On a weekly basis, PRISM typically evaluates 105 values (21 points for each of 5 week days) while
MARS can evaluate up to 70 values (10 points in LOD-UNCY table for 7 days). Typical Industry
models for MARS use 7 values in the LOD-UNCY table. A difference in this is that PRISM uses the
same 21 values across all 5 weekdays during a given week while MARS uses different values for
each of the 7 peak demands (assuming that a given week does not have the same peak for
multiple days).

The evaluation of the Expected Weekly Maximum (EWM) and its consistent nature with the load
forecast has been documented in the 2009 RRS (pages 17 and 18). However additional effort is
pending to document and compare the use of MARS’ load model with the load forecast model.
Once this documentation is completed a comparison of PRISM and MARS LOLE load modeling
could be further investigated and assessed.

The Load Model in PRISM is a forecast probabilistic model, aggregating daily peak values for each
week of a delivery year (from June 1 through May 31). MARS uses a deterministic 8760 hourly
loads, which are typically historic in nature. The PRISM solution method requires a forecast
probabilistic daily peak load model. Both load models can be adjusted to be consistent with
forecast demands (See Appendix C3).

The PRISM calculation can be considered a direct solution; i.e. – there is no Confidence Interval
around a given IRM value with the calculated value the exact solution. The IRM value represents
the exact value for a given model’s assumptions and distribution characteristics.

MARS is useful for seasonal, monthly, and weekly assessment work. Due to PRISM’s use of a
probabilistic load model, it is useful for delivery year and seasonal assessments.

PRISM can automatically iterate into the determination of the IRM to comply with the BAL-502RFC-02 Standard. MARS requires additional manual intervention for this determination. MARS,
used in the NYSRC study, uses the Unified Method approach which adjusts the generation
resources applying a “TAN 45 criteria” for determining the compliance IRM value.31

Please see the section in this report concerning the early reserve study model about historic testing
of computational engines. Those reports, by the NYPP and PJM LCTF, performed a technical and
comprehensive evaluation of the solution methods that still apply today. The full reports are
available to the RAAS members, on a restricted private SharePoint site basis, so that individual
members can review and dialog concerning application of these reports on the current Industry
issues of interest.
New York State Reliability Council LLC NYSRC Policy, No 5-3, procedure for establishing New York Area Installed
Capacity Requirements, November 16, 2009)
© PJM Interconnection 2011. All rights reserved
Page 33 of 136
PRISM’s ability to automatically solve to a specified LOLE level
An advantage that PRISM has over other Industry computer applications is that it can automatically iterate to
a user-specified reliability index (RI). This can greatly aid the efficiency of PJM’s RAP staff, allowing
members to focus on assessment and analysis of model impacts rather than tedious, repetitive manual
tasks. This advantage has been known for years, and was cited in the 1993 report that compared MAREL
to GEBGE. This specific comparison and discussion also applies to the comparison of MARS to PRISM.
The following was cited in the Conclusion and Recommendations section of the November 1993 Load and
Capacity Working Group (L&CWG) report comparing the multi-area application MAREL to GEBGE. (The full
details of this report are posted on the private RAAS SharePoint site.)
“Unlike GEBGE, MAREL lacks the capability to automatically adjust load levels to solve the case to
a user-specified reliability index. Manually solving to a 1 in 10 reliability index is a tedious process
that requires numerous iterations with various load levels. The LCWG was forced to accept a
reliability level between 1 in 9 and 1 in 11 (due to resource limitations). Conducting reliability
studies with MAREL would, therefore, require much more time and effort than conducting the
studies with GEBGE.”
Typically, assessment work desires to solve to either a 1 day in 10 criteria … or a more stringent 1 day in 25
criteria. The automatic Reliability Index (RI) solution ability allows efficient assessments to any LOLE criteria
of interest. This is such an important aspect of the calculation method that the letters RI in the name,
PRISM, point to this automatic RI solution ability.
PRISM uses an analytical solution technique similar to PTI’s MAREL program. However it is purposely
limited to only two areas, so that its computational speed advantage is not compromised. Also the
requirement of many areas modeled in a study, does not apply to the PJM RTO due to the Agreements in
place for the PJM Membership.
The assessment need is for establishing a single RTO-wide reserve requirement, currently measured at the
unforced capacity (UCAP) level by the FPR. This single RTO-wide reserve value is critical in the Operational
methods used by the PJM dispatchers. In addition, again due to the structure of the Agreements in place,
the PJM assessment makes an assumption that any megawatt of generation can serve any megawatt of
load; i.e., there are no transmission system bottlenecks or constraints.
Through the TEAC Stakeholder group and the RTEP process, this assumption that any megawatt of
generation can serve any megawatt of load, is tested and satisfied in the load deliverability tests – the CETO
and the Capacity Emergency Transfer Objective Limit (CETL) – performed to satisfy the RPM capacity
market place. Therefore, one needs to ascertain any differences in the Agreement and Operation structure
in place for any RTO, State, or Control Area to completely understand why the Monte Carlo or analytical
(convolution) method makes the best choice.
Historical Continuity
The calculation algorithms used and deployed in PRISM are the same as those used in the GEBGE
program. The difference between PRISM and GEBGE is that GEBGE is based on text files and historic
FORTRAN coding methods (Not modular) while PRISM uses current SAS modular coding methods, data
base schemas, and java based GUIs.
While PRISM and GEBGE arrive at the same calculated results, the maintenance and ability to
accommodate changing Industry needs (including those to address Operations and Markets issues) is much
more efficient and less costly for PRISM. Because the underlying calculation methods and application of
statistics and probability theory are essentially the same between GEBGE and PRISM, it seems apparent
that the conclusions and overall intent of the comparison between GEBGE and MAREL are applicable to
PRISM as well. It is interesting to note in the 1993 detailed report that,”The evaluation was a result of an
investigation of the SRTF in September of 1987 which assessed general reliability methodologies.”
This indicates that a thorough and comprehensive evaluation needs sufficient time to reach conclusions and
that various phases can provide insights and information to the parties involved to evaluate and decide on
support for identified next steps (i.e.; it took 7 years).
© PJM Interconnection 2011. All rights reserved
Page 34 of 136
The report discussion summary indicates that there was general interest in acquiring the MAREL program
although it was noted that MAREL would not replace GEBGE but, rather, be used in conjunction with it. This
is the same philosophy in place today – using PRISM (an improved GEBGE) with MARS (an improved
MAREL).
Although there have been many and significant changes in the bulk electric Industry since this evaluation,
the drafting team is not aware of any fundamental technical changes to the calculations and algorithms used
so that the rationale of complementary tools is still valid (i.e. the solution techniques developed in the 1970s
and 1980s are still used in today’s Industry methods). Please see the section Complementary Tools for
further details.
© PJM Interconnection 2011. All rights reserved
Page 35 of 136
PJM’s Recent Comparison Efforts
During 2002-2004, PJM Staff conducted rather detailed comparisons between PRISM and MARS. These
efforts were the first steps in implementing and integrating MARS into the database schemas and analysis
that are an integral part of the overarching Applications for Reliability Calculations (ARC) environment. PJM
Staff hired a GE consultant to assist and help develop a method to:
1.
2.
3.
Compare the models
Compare the calculations performed
Compare the available standard output
The goal of this work was to determine the best ways to consistently assess the two applications.
A major difference between PRISM and MARS is how load is modeled and assessed. This comparison
further revealed that:
32

Modeling comparisons should begin with the Capacity Model, holding all the various load model
options to a given, flat unchanging shape.

Capacity model items of interest include PRISM’s two state variance and two-state EEFORd. Both
models need to represent all outages as captured in the GADS data. MARS can translate the
EEFORd values into its needed transition states using the UNT-FORS table. One identified
improvement for PRISM involves modeling transition table-like values.

The generation model values in the UNT-MXCP, UNT-CAPS, UNT-DATA, MNT-UNOP can be
matched identically to those used in the PRISM model (see appendix C3, slide 9, as an example of
this kind of comparison). Development of an automated process for this is an identified future effort.

PJM uses the MARS UNT-DERT table to model the same units used for the 2,500 MW ambient
derate in PRISM. This modeling can be identical.

When modeling generation resources for PJM’s Mid-Atlantic region, PRISM and MARS gave
identical results (difference in LOLE ~.001). This was shown by the Project Team, however the
documentation of this effort was not a Project deliverable and this intermediate step was not saved.

When this work was first performed, MARS had a rounding limitation that could report LOLE to the
nearest 0.001. However, MARS has now been upgraded to show the LOLE to 10 -12 decimals (or
more).

MARS’ deterministic hourly load shape (the in02 file); having 8760 values received much attention.
A conclusion was that these hourly loads could be a “design year”, not necessarily from a single
given historic year. From a comparison standpoint, developing a design year had several
advantages. Adjustments to this hourly shape needs appropriate consideration with the interaction
with the other load models characteristics. However the design year concept has not gained
acceptance by other regional groups. Another valid method to address this is to develop the LODUNCY values based on the 8760 hourly load values, as describe in Appendix C3, slide 15.

In 2004 a comprehensive assessment was made concerning what values are appropriate to use in
the in02 hourly load shape file, so that consistent LOLE values can be assessed. The 2002 historic
hourly loads, for the PJMRTO, were judged to be appropriate for use in LOLE studies. However the
selection of the given calendar year may introduce a modeling consistency issue throughout the
entire set of MARS load tables (See Appendix C3).32

The MARS LOD-DATA and LOD-MTAR tables can be identically matched to the corresponding
PRISM parameters, annual peak load (Load forecast Table B1 & B2) and monthly peaks (Load
Forecast Table B6).
ISO-NE load model processing for selection of LOLE load models.
© PJM Interconnection 2011. All rights reserved
Page 36 of 136

When comparing the load model, we can use one MARS table and the appropriate PRISM
parameters to make comparisons. However, all the various table values need to be evaluated and
assessed as a package to determine the best, most consistent modeling approach. Moreover, the
need to evaluate four tables makes for a complex and potentially time-consuming effort. The MARS
tables involved in the load model translation include:
o
o
o
o
o
LOD-MTAR
LOD-UNCY
LOD-DATA
In02 hourly load shape
Convergence Index used in the solution (CNV-CRIT) LOLE.

The LOD-UNCY table values in MARS are consistent with the PRISM PLOTS load models by
translating the WKPKFQ load model parameters, on an Expected Weekly Maximum (EWM) basis,
to the 7 probabilities typically used by the NPCC CP-8 WG. A spreadsheet calculation is used for
this and shown in Appendix B.

PRISM reports results after the invocation of Load Management (Demand Resources and Energy
Efficiency), based on the PJM Load Forecast Report Table B-8, but before a Voltage Reduction.
This corresponds to the MARS EOP level 3, which occurs just before a Voltage Reduction.

The EOP levels used are:
1)
2)
3)
4)
5)
6)
Operating Reserves
Curtailable Loads
30-Minute Reserves
Voltage Reduction
10-Minute Reserves
Public Appeals.
All levels must be determined, by separate evaluation, for each area modeled.

MARS can be used in conjunction with PRISM to perform LOLE assessments that need:
o
o
o
o
o

Less than annual factor/output considerations such as the Winter weekly Reserve
Assessments.
Hourly assessments such as items addressed by CURTAIL and Load management
duration (How many hours?) and Adequacy value of interruption.
Assessment of Demand Management impacts as it increases to be a significant
portion of the PJMRTO peak load and/or a given LDA’s peak load.
Assessment of Intermittent resources, due to hourly nature of Wind facilities.
Coordination of modeling/assessments with other regions.
Refer to the NPCC Long Range Adequacy Overview discussion of this paper, in the Interregional
assessment section. This is an example of how MARS reporting can be consistent with other
PRISM and RTEP related planning assessments.
In the short term, identified investigation items include:
33

The NYISO’s use and documented Unified Methodology using the Tangent 45 inflection point
calculation to determine the IRM in the MARS assessment. The New York Control Area Installed
Capacity Requirements dated December 5, 2008, pages 27-28 and dated December 4, 2009,
Table A-4 page 28 submitted by the New York State Reliability Council, LLC installed capacity
Subcommittee.33

Investigation into the method and techniques used by the ISONE staff to compare their
Westinghouse model with MARS. Refer to the April 28, 2010 presentation on this topic, at the
NERC LOLEWG meeting (http://www.nerc.com/filez/lolewg.html ).
New York State Reliability Council LLC NYSRC Policy, No 5-3, procedure for establishing New York Area Installed
Capacity Requirements, November 16, 2009)
© PJM Interconnection 2011. All rights reserved
Page 37 of 136

Investigation into the method and techniques used by the MISO staff to determine the IRM for their
region.
Access to models: ARC screen display and GUI layout
After logging into the secure, intra-net web based application, the following is the main selection screen for
the Applications for Reliability Calculations (ARC) environment. This environment conforms to the PJM ITS
standards and process that require a change management process for both the database and the
application. There is a robust system to move changes from the development platform to the Test platform
and finally to the Production platform. The response time and reliability of these systems is paramount for
the PJM staff to perform all activities related to the Adequacy studies.
The Week Peak Frequency system is how one specifies and creates the forecast, probabilistic load models
(PLOTS) used by PRISM. The ARC system Administrative system has five categories of activities which
include defining user access and twenty five database administrative functions. The “View Production
Information Warehouse (PIW) Status” option allows a quick summary of all database elements to identify
when the latest update was performed.
© PJM Interconnection 2011. All rights reserved
Page 38 of 136
The main MARS GUI screen is shown below. All options are indicated on the left, with all input options
perform by the selection task drop down button. The GEMARS Help link at the top of the page allows for
specific and detailed on-line user instructions and examples. The user can do searchable inquires and
default specific information is available in a point-and-click environment. The underlying database schemas
and tables are the “engine” that allow the GUI selections to create the Master Input File (MIF) file necessary
to make assessment runs. There is an understanding between the user and support staff that some text
editing is appropriate, as further development work is necessary to fully automate all GUI selections with the
database schema.
© PJM Interconnection 2011. All rights reserved
Page 39 of 136
The main PRISM GUI screen is shown below. All options are indicated at the top selection categories, with
all input details selected in the underlying GUI screens. The Help selection allows for specific and detailed
on-line user instructions and examples. The user can do searchable inquires and default specific information
is available in a point- and-click environment. The underlying database schemas and tables are the “engine”
that allow the GUI selections to create the PRISM and GEBGE text files necessary to make assessment
runs.
© PJM Interconnection 2011. All rights reserved
Page 40 of 136
#1 Comparing the Models
Table 1 provides a detailed comparison between PRISM and MARS and their relative strengths
and weaknesses. Of the many parameters considered, there are four prominent items to mention
from this table: a) Technology, b) Underlying Data Requirements, c) Planning Horizons and d)
Transmission System Model.
a)
Technology – PRISM uses state-of–the-art technology to improve data integrity,
quality results and user efficiencies.
b)
Underlying Data Requirements – PRISM uses a relatively limited set of statistical
parameters while MARS uses numerous event-driven, data-intensive tables. See
Appendix F for further details.
c)
Planning Horizons – Typically PRISM has long-range planning capabilities while
MARS has a comparatively shorter-term planning horizon. This is due to the
intrinsic nature of the operational parameters used in the MARS modeling. Note
that the MAPPCOR and CAISO do publish planning study results using MARS
for a ten year period.
MARS’ use of Operational related inputs present a long-range assessment study
issue as Operation staff is focused on shorter duration periods. These
operational parameters need feedback and assistance by the Operation’s staff.
Values in the long-range efforts might necessitate more sensitivities and
variations to capture uncertainty.
d)
Transmission System Model – PRISM can evaluate up to two areas at one time
whereas MARS has the capability of evaluating 75 (or more) areas.
#2 Comparing the Calculations Performed
PRISM uses a numerical method to simulate a general convolution method by an array look-up into
a cumulative probability table of unit outages. MARS uses a Monte-Carlo method that iterates on a
deterministic load shape simulating different generating unit outages. All things being equal, the
solution time for PRISM is markedly faster than that for MARS.
PRISM enjoys an efficiency advantage as it is able to automatically solve to a specified LOLE
criteria, saving the user from repetitive, tedious efforts. Investigation into other ISO techniques
would seek solutions to address this efficiency issue in the MARS processing.
#3 Comparing the Available Standard Output
Table 2 shows the comparison of the PRISM and MARS output files. One of the obvious
differences is that the PRISM output is in SAS tables and structures while MARS is a flat text file.
The “Text Pad” tool is used to cut and paste data from the text files while SAS can export data
directly into an Excel spreadsheet. Both methods are easy to extract data for reporting and creating
summaries. Both tools have many summaries and reported values that are useful in LOLE
assessment work.
The PRISM output does have the ability to be used in an OLAP, and by ARC’s use of a GUI and
underlying database schema assures consistency among the two models being reported.
© PJM Interconnection 2011. All rights reserved
Page 41 of 136
Comparison of Attributes Documentation
This section includes comparison of strengths and weaknesses and commentary on specific modeling attributes between PRISM and MARS.
Table 1 addresses attributes and application for each tool. Table 2 discusses eleven specific output comparisons with commentary. Table 3
discusses thirteen specific comparisons of database modeling relationships with commentary.
Table 1 – Model Attributes / Application Comparison Matrix
PRISM
Comparison
Strength
MARS
Weakness
PJM
Developer /
Owner
Developed in-house by PJM Staff;
relatively easy to make revisions and
modifications.
PJM is not staffed nor has the time and
resources to provide in-house subject
matter experts to provide PRISM
training and support to external users.
PRISM based on SAS and Database Schemas
Technology
Newer technologies allow for
improved data assessment and
analysis of results
Requires robust underlying information
systems – as a “stand-alone system”,
could be cost-prohibitive.
PJM – however, external requests have been made for the
predecessor application GEBGE
Current Users
Several LSE staffs use and
contribute to the methods deployed
in PRISM. Full and complete
documentation is available for
participant review.
© PJM Interconnection 2011. All rights reserved
Only PJM and ISO-NE staff use a
similar calculation approach. This
allows the opportunity to clarify and
produce concise answers to detailed
technical questions.
Page 42 of 136
Strength
Weakness
General Electric International, Inc. (GEII)
GE supplies a “turn-key” modeling
product and provides quality training and
excellent support to licensed MARS
users.
MARS based on FORTRAN and “flat text” files
Well-known programming structure;
enables easy, quick, manual changes.
Allows for efficient delivery of
executable code for updates that secure
proprietary intellectual property rights.
Older technologies are less flexible;
difficult to upgrade without wholesale
system changes
Multiple licensees
MARS enjoys a wide user network; this
is potentially valuable for interregional
reliability determinations and Broader
Regional Market applications. MARS
enjoys the distinction of an LOLE
product with widespread use among
almost all of North America. Full and
complete documentation is available for
participant review.
PRISM
Comparison
Strength
MARS
Weakness
GADS, historic loads, operational experience, transfer limits
Underlying data
requirements
PRISM has a comprehensive but
more limited amount of data inputs.
The nature of the data inputs do not
rely on other resources to assist in
compiling the data.
The numerical methods used for the
cum Probability table allow a quick
and direct solution (~ 15 minutes).
There is no confidence interval as
one calculates the exact LOLE.
No ability to produce confidence
interval value. No ability yet to
produce output for all 21 numerical
method states.
Yes
Determine
Capacity
Benefit Margin
(CBM)
The advantage is that PRISM only
needs a single value. Once
determined, this is not a
maintenance or otherwise
burdensome task.
34
35
Reporting LOLE is a single value,
before invocation of Voltage
Reductions. Clarity of basis of LOLE.
GADS, historic loads, operational experience, transfer limits
MARS needs more specific Operations
related data. This might cause a
modeling concern if other groups need
to perform an assessment for this data
and there is a resource (manpower)
limitation.
Monte Carlo solution / Standard Error basis usually a 0.05 value.
Can report out many aspects of the
solution process, tracking each solution
with a good summary output of solution
details.
Possible slow solution (some examples
35
were up to 18 hours) but depends
upon processing speed, modeling
assumptions and number of
simultaneous areas evaluated.
Not typically determined for a region with many areas
Only a single value to represent paths
involved in simultaneous import limit.
After invoking load management. (Not disconnecting firm load)
Reporting of
LOLE results
Weakness
Useful in detailing specific event data
(GADS 97 card) and events seen in
operations.
Direct convolution method / Solution tolerance is 0.001
Solution Basis34
Strength
No ability to automatically assess
LOLE at other EOP levels. Need
assessment for each desired EOP
level.
A very detailed transmission pipe
representation.
The transmission pipe sizes are
required for the model. If significant
Load flow work is needed, this could
cause a delay or additional effort to
make appropriate assumptions.
at EOP margin states
Calculation and reporting LOLE at up to
10 EOP levels.
Need Engineering Judgment
concerning the best reported value,
when considering consistent
comparison. Allows defining levels
differently which can muddy the water
when comparing results among
regions.
Solution Basis is determined using a 2-CPU server running Windows 2003 Operating System with 18 gigabytes of RAM and CPU speed of 3.0 GHZ – as applied to large models,
such as those modeling the bulk of the Eastern Interconnection.
Eighteen-hour (18) run times were seen in detailed models using dynamic interface transfer limits, as this is part of the core calculations that requires the largest percentage of
solution time. Refer to Figure 5C for the Monte Carlo simulation process.
© PJM Interconnection 2011. All rights reserved
Page 43 of 136
PRISM
Comparison
Strength
MARS
Weakness
Two: 1) PJM RTO, 2) World
Simultaneous
Study Areas
Up to two areas. This can be an
advantage if proper consideration of
assumptions and its use can be
established to simplify modeling.
For some assessments, PRISM is
limited in assessment capabilities such
as evaluation of contributions from
more than one surrounding neighbor.
One transmission interface (pipe) between areas
Interface
Transfer Ties
One transmission tie, which allows all
pertinent assessment work to
happen outside of the LOLE model.
No probabilistic model for the
transmission tie. Only one transfer pipe
can limit modeling and coordination
with known operations practices.
Planning
Modeling Style
Due to its probabilistic load and
generation model, this is only a
planning division tool.
Increased development to coordinate
with Operation practices.
1 year to 11 years ahead
Planning
Horizon
Data Input
Requirements
36
PRISM offers wider time range.
PRISM’s use of statistical
parameters are useful in the time
frame
Primary use is not l close in seasonal
assessments or less than annual
evaluation efforts.
Three primary input categories-Load, Capacity, Tie Size
PRISM less complicated, less time
intensive
Does not have breadth and depth of
data inputs for reporting out many
different LOLE aspects.
Strength
Weakness
Multiple (up to 75 used)
MARS exhibits much greater flexibility;
virtually unlimited
Increased data requirements.
Multiple (up to 125 used)
MARS exhibits much greater flexibility
Increased data requirements.
Planning, Operational
MARS allows for Operational data to be
specifically modeled
Need for operational data, increased
effort and potential delays or use of
best estimates to ensure best quality
data.
6 months to 5 years ahead
MARS is very useful for close in
assessment work, such as seasonal
assessment that is 3 to 9 months in the
36
future.
MARS can be used to perform long
range assessments, but this
necessitates increased efforts for a
quality forecast data model
(Operational data such as transmission
limits, megawatt values for Emergency
Operation levels).
Many separate tables for modeling values (~50)
Many, many Tables, allowing a plethora
of detailed options that allow Operations
related assessment. Very good for close
in assessments (1-5 years).
Much more complicated, with model
needing all tables of interest before a
single run can be made. Might force
use of best approximations.
California ISO Planning Reserve Margin – 2010 – 2020, May 21, 2010 – 2020. http://www.caiso.com/279d/279ded0337f20.pdf and MAPPCOR December 30, 2009 LOLE
Study 2010-2019, look out for a 10 year period.
© PJM Interconnection 2011. All rights reserved
Page 44 of 136
PRISM
Comparison
Modeling
specifics of
interest
Strength
MARS
Weakness
PRISM models include: Weekly Standard. Normal distribution and Expected
Weekly Maximum (EWM); Forecast monthly peaks; Peak Load Ordered Time
System (PLOTS) file-Mean and STD; Sensitivities to assess EOPs and transfer
ability; Capacity model based on PJM Manual M-22 statistics; EEFORd; TwoState Variance; Unit maintenance; DR & EE; Solved load and 50/50 forecast;
Unit commercial probability; BTM and uncommitted resources
Many details can be modeled and
assessed.
PRISM allows all states (possible
loads) to be evaluated in a given
base case. This is a forecasted
model, with a basis in recent, historic
daily peak demand.
PRISM does not allow an easy, direct
correlation of load shape to a given
Operations period.
Describing and explaining the load
shapes is not straight forward and
requires considerable time, energy,
and a given level of expertise for a
complete understanding.
Based on 10-year historical data
Load Shape
Forecast Probabilistic shape.
Need to assess shapes and
association with other regions for best
selection of time period.
Magnitude ordered – using PLOTS Load Model Shape (Weekly
mean, STD)
Load Model
Shape
Forecast Probabilistic shape.
© PJM Interconnection 2011. All rights reserved
Weakness
MARS Tables includes: LOD-UNCY, LOD-MTAR, in02 file, EOP-DATA, INFTRLM, ELU-DIST, UNT-DERT, UNT-FORS, FCT-DATA, INF-DLYM, INF-DATA,
INT-ONLY, NUM-TRNS, MNT-OPTN, MNT-UNOP, UNT-MXCP, UNT-CAPS, UNTDATA, LOD-DATA, CNV-CRIT, GEN-AREA, GEN-POOL, GEN-CASE
Many details can be modeled and
assessed.
Probabilistic
Load Model
Type
Strength
Need to assess shapes and
association with other regions for best
selection of time period.
Page 45 of 136
Deterministic
MARS allows for specific operation
condition modeling. Explanation of
deterministic values used is rather
straightforward. Use of load forecast
uncertainty table places a weighting of
the probability a given load value is
expected to occur.
MARS Load Model is not a forecast
shape; it is based on historic events.
Needs periodic and independent
assessment work to determine best
historic shape to use. MARS also
needs separate assessment to
determine appropriate load forecast
uncertainty table values, for each area
in mode.
Committee selection of “Base Year” Load Shape
MARS can apply a forecast monthly
distribution and a load forecast
uncertainty to make the deterministic
shape in line with probabilistic resource
model in Monte Carlo simulations. See
Appendix C3.
Since MARS uses a deterministic load
shape, it is constrained by a subjective
and engineering judgment selection
process.
Chronologically ordered – Hourly Load Shape
Straight forward to implement, once a
year has been chosen.
Deterministic based on historic year. If
the shape is many years ago, one can
lose the relationship with current trends
and forecast shapes used in 50/50
forecast.
PRISM
Comparison
Strength
MARS
Weakness
Daily
Peak Load
Assessment
Load Diversity
Forecast Error
Factor (FEF)
Expected
Weekly
Maximum
(EWM)
PRISM modeling starts with the
same hourly loads; it moves to the
more critical daily peaks thus
streamlining any maintenance or
periodic review issues.
1) Simulation based on hourly loads
which capture historic diversity.
2) Consistent with Load forecast.
3) Lower resource requirements to
perform assessment.
A stakeholder and PJM staff
documented values and rational.
PRISM can only address daily peak
assessments. Cannot perform Hourly
assessment using LOLH or EUE.
Yes
1) Resources and expertise to perform
assessment
2) Limited to two area model.
3) No ability to capture dynamics
between sub regions that make up
world region.
Yes
Yes
Lack of increase for future delivery
years.
Documented using the 1st order
statistic in proper aggregating the
daily values into each week modeled.
Yes
Load Forecast
Uncertainty
(LFU)
These are determined from the
appropriate WKPKFQ model and
EWM values.
This directly corresponds to the
monthly load factor used in PRISM or
the 18 Control Card loads.
© PJM Interconnection 2011. All rights reserved
Weakness
Hourly Load Shape, MARS table: in02 file
Granularity can cause issue as hourly
shapes need periodic detailed
assessments.
MARS can use more granular approach
1) Able to simultaneously assess multi
area diversity. Inherent in the 8760
hourly loads (in02 file) data
modeling specification.
2) Can be made consistent with the
monthly peak load forecast.
Yes
1) Resource and expertise to perform
assessment work.
2) Must decide on consistent model
characteristics for a system wide
study.
Not typically done, but could be included.
Can be Incorporated in the LOD-UNCY
table
No way to determine specifics
concerning this modeling attribute.
Not needed for the daily peak
calculations. This uses a different load
model than PRISM, based on 8760
hours. See Appendix C3 for further
discussion (slides 10-15).
Needs coordination with probabilistic
load distribution if consistency and
result reporting is desired to be readily
available between the two tools.
Yes but needs further efforts
MARS table: LOD-UNCY
Use allows the load model shape to be
evaluated at various level of
probabilities.
Yes
Monthly
Forecast Loads
Strength
Page 46 of 136
MARS table: LOD-MTAR
Use allows the forecast monthly load
model shape to be adjusted to match the
50/50 forecast.
PRISM
Comparison
Strength
MARS
Weakness
No
Load
Management
Assessment
PRISM uses an application called
CURTAIL. This evaluates the
reliability benefit for each invocation
of a load management event.
Uses the EEFORd and a two state
model to represent either a full on or
full off scenario.
CURTAIL has limited documentation
and would need time to fully document
process and rational.
Requires GADS event data for 5 year
period.
Represented by a statistic
Forced Outages
Weakness
MARS can model and calculate hourly indices
Probabilistic
Capacity
Modeling
Strength
The PRISM modeling initiated the
Industry to consider including the
EFORd statistic into the IEEE
Standard regarding this matter, IEEE
STD 762.
MARS can assess things related to Load
Management duration, wind unit
performance, other intermittent
resources, etc.
Probabilistic
Uses transition states or Forced outage
rates to approximate transition states.
Yes
Basis for each unit’s forced outage rate
to determine unit availability.
Done outside of PRISM w/ supplementary GE program using
NUM-TRANS table
© PJM Interconnection 2011. All rights reserved
MARS allows non-EFORd values to be
input. Care needs to be made to make
sure all values are on the same basis
and consistence with IEEE STD 762.
MARS offers robust solution
Basis for each unit’s forced outage
rate to determine unit availability.
The primary unit two state
performance statistic, EEFORd, is
based on the EFORd. See PJM
Manual 22.
Transition Rate
Requires GADS event data for 5 year
period.
Represented by a table
Yes
Equivalent
Demand Forced
Outage Rate
EFORd
Typically this is performed by specific
sensitivity work.
Not available in PRISM yet. An
identified future enhancement.
Page 47 of 136
MARS table: UNT-FORS
Application included in MARS
PRISM
Comparison
Strength
MARS
Weakness
Strength
Weakness
Yes
Wind Unit
Performance
Other
Intermittent
Capacity
Resources
Modeled per the existing guidelines,
outlined in PJM Manual 21, Appendix
B
Yes
Need GADS event data for proper
modeling. Proxy is data turned into
GADS administrators over previous 3
year peak period.
Yes
Modeled per the existing guidelines,
outlined in PJM Manual 21, Appendix
B.
One transmission tie only. No issue
with data collection.
Need GADS event data for proper
modeling. Proxy is data turned into
GADS administrators over previous 3
year peak period.
Identified improvement could be to
model a probabilistic set of values.
Relatively Low
Modeling
Resources
(Staff)37
MARS offers robust solution. Modeled
per the existing guidelines, outlined in
PJM Manual 21, Appendix B.
Shorter runtime is a major
advantage for mitigating PJM staff
resource limitations.
37
Need GADS event data for proper
modeling. Proxy is data turned into
GADS administrators over previous 3
year peak period.
Deterministic
Variation of results, offers robust solution
and reporting.
Could cause delays or extra effort due
to resource limitations to determine
modeling values.
Higher
Data requirements are captured by
PJM’s RAP department. No known
resource limitations.
Typically ~ 5-15 minutes
Solution
Runtime38
Need GADS event data for proper
modeling. Proxy is data turned into
GADS administrators over previous 3
year peak period.
Yes
Deterministic
Transmission
System
Modeling
MARS offers robust solution for
intermittent resources. Modeled per the
existing guidelines, outlined in PJM
Manual 21, Appendix B
Potential delays and limitations of
quality data due to resource limitation
of non PJM RAP staff.
Long ( 30 min - 18-hour) runtime due to multiple iterations (~2500
runs)
Can be addressed by using a large
server network or running on a super
computer cluster.
Possible times of up to 18 hours, but
access to a large network environment
can serve to significantly reduce this to
about 30 minutes.
Reference Resource and cost Assessment Section for details. Specific assessment efforts vary. Depending on the assessment, PRISM or MARS might show lower resource
estimates.
38
Solution Basis is determined using a 2-CPU server running Windows 2003 Operating System with 18 gigabytes of RAM and CPU speed of 3.0 GHZ – as applied to large models,
such as those modeling the bulk of the Eastern Interconnection.
© PJM Interconnection 2011. All rights reserved
Page 48 of 136
PRISM
Comparison
Strength
MARS
Weakness
Solution based on size of Cumulative Probability Table
Runtime
Characteristics
Strength
Network complexity, binding issues can impact run time
If appropriate run time options are
selected, Results stored have plethora of
details for the specific calculations
performed in the Monte Carlo
simulations. See figures 5A-5C.
PRISM employs look-up table
approach
Yes
Hourly Data
Assessment
PRISM starts with the hourly loads.
Hourly loads subject to several
automated checks and OLAP review
to insure data quality. Possible
assessment work for neighboring
regions as the hourly loads for
select time period are mandatory.
© PJM Interconnection 2011. All rights reserved
Weakness
Combining all sub-regions that
comprise the World region assumes
close coordination of many control
areas. These control areas are
geographically diverse which might not
represent operational practices.
Page 49 of 136
Yes
Hourly data is used per separate
determination of best values for study.
Consider “Design Shape” instead of
selecting historic hourly loads.
Table 2 – OUTPUT Comparison Matrix
Output Comparisons
Table 2 shows eleven separate output summaries that PRISM produces, and the corresponding MARS
output files. These summaries indicate the results of the calculations, giving various aggregations and
details. These summaries are the “standard” output, or otherwise stated, this is the output for most PRISM
and MARS runs, using the most commonly used input parameters.
PRISM can output many more calculations details, by setting the “DEBUG” option before submitting a
PRISM case. The DEBUG option causes many more output files to be created for a given case. The
DEBUG option can cause use of more computer resources and require more solution time. Once the
DEBUG option is set, it will apply to all submissions in a given environment. This allows someone working
in the development environment to work on developing calculation changes, needing scrutiny of all the
details, while those in the Production environment can perform normal work and not be affected by the
additional DEBUG output. See Appendix E for further details.
The analogous items in MARS to the DEBUG option in PRISM, is the many and various input parameters in
the CNV-CRIT, GEN-TIME, and GEN-CASE tables.
© PJM Interconnection 2011. All rights reserved
Page 50 of 136
Table 2 – OUTPUT Comparison Matrix
Output File
PRISM Output
MARS Output
Comments & Comparisons
1. ALMTABLE (Demand Resources)
 Active Load Management (ALM) refers to the
former term used for Demand Resources
(DR). The output is derived from four specific
runs:
1. Base
2. Insufficiency
3. Necessary
4. Initiation


Not available in MARS.
This is a group of specific and automated analysis
which would need to be developed using MARS.




 All these are inputs into the CURTAIL
application.
 This output reflects the reliability values of DR,
the maximum amount of DR that saturates its
reliability value at 10 interruptions, and the
number of times a given amount of DR needs
to be interrupted for its reliability value to
saturate

The Load Management calculation details are part
of the desire to treat the contractual obligations,
Operational practices and planning assessment on
a consistent basis.
The calculations measure the reliability benefit of
the number of LM interruptions.
Contractually the limit is 10 for the summer peak
season.
Historic calculations verify that the full year benefit
is similar (identical) to the summer only benefit.
There are several issues surrounding DR with
assessments and evaluations being performed to
discern any different and new approach due to
using Load Management in a resource construct
(Bids into Capacity market).
2. CUMPROBTABLE (Cumulative Probability of Outage)
 The cumulative probability table is a twocolumn array, indexed by the amount of MWs
to be evaluated for an outage, typically the
columns represent: 1) daily peak load, and 2)
the associated cumulative probability. The
cumulative probability is based on the unit
characteristics of the model.
 For a two-area run there are two separate
tables: Area 1(PJMRTO) and Area 2
(Neighboring Region).
© PJM Interconnection 2011. All rights reserved



Page 51 of 136
Output Table (OT) 10 (OT10) can be used to
determine the area margins, but this will only yield
the LOLE for each load level evaluated and the
megawatt amount of the capacity.
However, the distribution characteristics of the
aggregated unit total are unknown for each
replication.
These negative margin values, for each load level
and each EOP level can be outputted by specifying
the input parameters in the GEN-TIME table, "Print
hourly output" with caution given that this can
result in a very large output.


The PRISM margin values for any of the values
looked up in the cumulative probability table are
not saved or known outside the calculation
process as the determination is only if the load
value looked up can be served by available
generation, it is an "on" or "off" switch.
If running this with the PRISM DEBUG options, the
Cumulative probability table can be saved, but the
intermediate LOLE equation values for all points in
the distribution are not saved (yet).
PRISM Output
Output File
MARS Output
Comments & Comparisons
3. OUTAGE_SCHEDULE

Parts of the OT7 file output. Interestingly there are
53 weeks for this table showing the individual unit
planned pattern. This shows weekly categories:
o
Fixed daily maintenance
o
Partial fixed maintenance
o
Scheduled maintenance
o
Partial Scheduled Maintenance
o
Unit not yet installed
o
Unit Retired

Two MARS summaries, for each calendar year in
the study are shown.
 One of the primary summaries of calculated
results showing the weekly:
o LOLE
o Mean of Load distribution after
Monthly adjustment
o Load distribution Standard Deviation,
o Expected Weekly Maximum
o Total Standard Deviation (after
applying 1st order statistic and
Forecast Error Factor(FEF)),
o Capacity amount on planned outage
o Reserves in percent and megawatts.

OT8 File Output gives detailed results, on a weekly
or Monthly basis for each calendar year
incorporating several summaries of these values:
o Weekly Indices, LOLE (Days), LOLE
(Hours), LOEE (MWh)
o Expected Number of Days at Specified
Margin States – Weekly
o Monthly Indices, LOLE (Days), LOLE
(Hours), LOEE (MWh)
o Expected Number of Days at Specified
Margin States – Monthly
o Interface Flows – Monthly

In a general sense the outputted values show the
necessary details.
A process is in place for both applications to
request and change the outputs as justification
and benefits are appropriate.
A given summary, due to the nature of the
difference between the analytic solution
(convolution) and the Monte Carlo solution might
be cost prohibitive so that a direct comparative
output can be made.
There are many output summaries in the OT7,
OT8, and OT11 files.
 Shows annual LOLE and RI, separately for
two area model

OT7 file output gives detailed results on a weekly
or monthly basis, for several categories with the
"Weekly Reserves Summary" category of interest.

OT11 provides further detailed area results, on
weekly, monthly load level (LOD-UNCY), and EOP
level basis for each calendar year; summary
categories include:
o
For each LOD-UNCY level, Calculated
Indices (LOLE (days/yr), LOLE (hrs/yr)
LOEE, Frequency, Duration) for both
isolated and interconnect system.
o
For each LOD-UNCY level the expected
number of days at specified EOP margin
state
o
For each LOD-UNCY level, Import and
Export Interface flows
 Listing of generation units that are unavailable
due to a planned outage for the 52 weeks in
each delivery year assessed in a given run.
4. PERIOD_PROBABILITY
© PJM Interconnection 2011. All rights reserved
Page 52 of 136



PRISM Output
Output File
MARS Output
o
o
o
o
o
o
Comments & Comparisons
Annual EOP usage for given LOD-UNCY
level
Weekly Indices (LOLE (days/yr), LOLE
(hrs/yr) LOEE, Frequency, Duration) for
given LOD-UNCY level load
Weekly Expected number of days at
specified EOP margin state for given load
level
Monthly Indices ( LOLE (days/yr), LOLE
(hrs/yr) , LOEE) for given load level
Monthly Expected number of days at
specified EOP margin state for given load
level
Monthly Import and Export Interface flows
for given load level.
5. PRISM_PARAMETERS
 Identifies user name, runtime, input file name,
location of output file, various solution
parameters.
 Useful for trouble shooting and debug efforts.


The various files used in the study are stored in the
same directory.
These can be found quickly by going to the
directory_Help of the ARC system and selecting
the GEMARS editable directories.

For MARS runs, there is no record of users as the
information is stored in the directories per the
Windows OS.
OT6 shows
o Number of replications in study
o Convergence Tolerance
o Standard Error

MARS Solution time can be determined, assuming
one saves the MIF input file just before
submission, by subtracting the time of the OT files
from the input MIF file.
6. PRISM_TIME
 Indicates amount of time for each set of
analysis

7. SUMMARY TABLE
 Primary output summary indicating:
o Solution load
o Installed Capacity
o Week of peak
o Installed Capacity (ICAP) Reserves
o Available reserves(after planned
outages),
o CBM value
o RI
© PJM Interconnection 2011. All rights reserved
 OT 9 provides general results for each calendar
year, using several summary categories:
o Program Option Summary
o Interface Rating Data
o Area Capacities and Loads
o Calculated Indices - Both Isolated and
Interconnected LOLE as days per year,
hours per year,
o Frequency Duration (expected number of
days per year at specified margin states
o Interface Flows
o Annual EOP Usage - (expected number of
times per year)
Page 53 of 136
Output File
PRISM Output
MARS Output
Comments & Comparisons
8. SYSPAR (System Parameters)
 SYSPAR indicates:
o
Installed Capacity (ICAP)
o
Load Distribution mean
o
Load Distribution Standard Deviation
o
ICAP Distribution Mean
o
ICAP Distribution Variance
o
Average Unit Size
o
Average Forced Outage Rate
o
Average Planned Outage (MW)


OT 7, OT 8 and OT 9 includes the underlying data
(EX individual unit) that are used in these
summaries.
However, the OT7 file does show aggregated load
model statistics on a weekly basis.
9. UNITDESCRIPTIONS (Generation Unit Descriptions)
 List of all units in the case: name, unit number
assigned case number, geographical number

OT7 includes individual unit data statistical data in
three summaries:
1. Weekly Individual generation unit
maintenance and outage schedule
2. Available Energy by non-Thermal unit
type(Monthly)
3. Thermal Unit Summary (Summer & Winter
Ratings, FOR, planned weeks of outage)
10. UNITVALUESBYWEEK (Weekly Generation Unit Values)
 List of all units in the case, by delivery year:
name, unit number, Variance, EEFORd, Adder
or change unit, rated value for each of 52
weeks, on or off status for each of 52 weeks.
© PJM Interconnection 2011. All rights reserved

Page 54 of 136
OT7 shows the individual unit data statistical data
in three summaries:
1. Weekly Individual generation unit
maintenance and outage schedule
2. Available Energy by Non-Thermal Unit
Type(Monthly)
3. Thermal Unit Summary (Summer & Winter
Ratings, FOR, planned weeks of outage)
 These summaries are for the distributions for the
primary model input categories: Load, Generation,
and Transmission.
 Some summaries might use a distribution,
perhaps a confidence interval applied, for
application in the MARS Monte Carlo solution.
Output File
PRISM Output
MARS Output
Comments & Comparisons
11. OLAP (On-Line Application Program)
 A cube using an On-Line Application Program
(OLAP) uses ~ 80 specific datasets built from
the primary SAS input database tables and
resulting output files.
 This allows many different and unique
combinations of the underlying dataset details
and allows efficient investigation and
verification of data quality of both the input
data and assessment results.
 The key in this is the appropriate joining of the
load data, defined by sub zones, with the
generation resources, defined by individual
units, using Holistic techniques’ (HOLAP) to
correctly define complex relationships used to
build the database schemas.
 This allows the planning engineer user to
display many facets of the data, in a point and
click GUI, otherwise not available without ITS
staff expertise and resources.
 OLAP metadata summaries include:
o
Forecasted Loads and Reserves, by
subzone
o
Historical Loads, by study area
o
PRISM Reports
o
Generation Owner Web Comparison
Report
o
Peak Week Frequency reports
o
Generator class average comparison.
© PJM Interconnection 2011. All rights reserved

Page 55 of 136
MARS does not have this capability as it uses text
output files.








Online analytical processing, or OLAP
(pronounced /ˈo-lap/), is an approach to quickly
answer multi-dimensional analytical queries.
OLAP is part of the broader category of business
intelligence, which also encompasses relational
reporting and data mining.
The typical applications of OLAP are in business
reporting, and business process management
(BPM)
The term OLAP was created as a slight
modification of the traditional database term
OLTP (Online Transaction Processing).
Databases configured for OLAP use a
multidimensional data model, allowing for
complex analytical and ad-hoc queries with a
rapid execution time.
OLAP methods borrow aspects of navigational
databases and hierarchical databases that are
faster than relational databases.
The output of an OLAP query is typically
displayed in a matrix (or pivot) format.
The dimensions form the rows and columns of
the matrix; the measures form the values.
Table 3 – DATABASE Modeling Relationship Matrix
Table 3 shows thirteen categories related to the data base schemas in use. These schemas use SAS and
Oracle tables. Table 3 discusses the relationship between the two tool’s established data base items. It has
been said that “It is all about the data!”, and that still continues to be the case in conducting the various
Adequacy assessment work. An important and significant amount of effort is needed to ensure high data
quality. High data quality is a mantra for the PJM staff involved in the Adequacy study efforts. One of the
processing methods in place is comparing our SAS database with known detailed Industry data which might
be available by participation in Industry groups, with Industry consulting firms, or with public data sources.
© PJM Interconnection 2011. All rights reserved
Page 56 of 136
Table 3 – DATABASE Comparison Matrix
Database
File
PRISM Database
MARS Database
Comments & Comparisons
1. ALL Monthly Load Forecast
 This database of the monthly peak load
forecasts is for each identified zone of the
PJMRTO or neighboring World regions.
 The PJMRTO data is from the PJM Load
forecast report typically dated in January. The
neighboring World data is based on the NERC
Electric Supply and Demand (ES&D) report.
 The PJMRTO data is posted on the PJM web
site as well, but the database has more
granularities and incorporates many
complexities not available in the public data
posting.





This database of the monthly peak load forecasts
is for each identified zone of the PJMRTO or
neighboring World regions.
The PJMRTO data is from the PJM Load forecast
report typically dated in January.
The neighboring World data is based on the NERC
ES&D report or publically available data posted on
other neighboring ISO web sites.
The PJMRTO data is posted on the PJM web site
as well, but the database has more granularity and
incorporates many complexities not available in the
public data posting.
The MARS data does require different summaries
in addition to those used in the PRISM model, due
to using different area cut-sets (the cut-set define
an area's boundary).

The LOD-UNCY-01 table can currently
accommodate monthly shape distribution values.
The PJM staff has asked for this input's granularity
to be improved so that either weekly or daily peak
distribution parameters can be specified.
The maximum number of points that can be used
in this distribution is 10.
The process to determine these points is
discussed in Appendix C and an example shown in
Appendix B.
The LOD-UNCY data is stored in the SAS tables,
on an area basis, but is currently manually inputted
into the MIF.


This data is very similar in structure between the
two models with both based on the same data
sources, either the PJM load forecast report or
NERC ES&D.
Some of the modeling needs for MARS require the
most granular sub zone data due to drawing cutset lines around a given control area. PRISM
aggregates things into one of two categories,
either a PJM RTO or World region.
2. Non-Normal weekly load shapes
 The database allows for weekly load
distributions to be input.
 The distribution can be from external
assessments to give the correct forecast
shape for each week in the model.
 Currently this shape is used for both area 1
and area 2, with work in progress to specify
different weekly shapes for the two areas
(PJMRTO and neighboring World region).
 These are the Characteristic (load) Data
Curve (CDC) files in the database which are
selected in the PRISM GUI parameters.
 There is no limit for number of points used in
these distributions.
© PJM Interconnection 2011. All rights reserved





Page 57 of 136

PJM staff is working to produce the weekly shapes
used in PRISM and to improve the SAS code so
that different shapes can be used for Area1 and
Area 2.
PJM is also discussing with the GE staff the
benefits of improving the LOD-UNCY table's
granularity.
Database
File
PRISM Database
MARS Database
Comments & Comparisons
3. CLASS_AVERAGES
 Values listed by categories based on type,
size, and primary fuel of a unit.
 These values primarily come from the existing
PJM Fleet of units, about 1400 units. NERC
pc-GAR data is used for any categories that
do not have a sufficient number of PJM units.
 These values are used when a unit does not
have 60 full months of GADS event data.
 The statistical values involved in this are the
POF, EFORd, Two-State variance, EEFORd,
and EMOF.



The data source is the same as for PRISM, which
is used in the UNT-FORS, UNT-MXCP, MNTUNOP tables.
However MARS is based on the transition rate
matrix so the UNT-TRNS and UNT-CAPS tables
must be accommodated using the known data.
Directly calculating the UNT-TRNS and UNTCAPS from the GADS event data would improve
the MARS model and allow for measuring the
approximations used in determining the tables'
values based on the two-state values used in
PRISM.

MARS uses the same SAS data sources as
PRISM but any manipulation for sensitivity analysis
is done manually.
Automating this to avoid any manual intervention,
using Text Pad, would be additional effort.

This is a small number of units but can be tricky to
achieve high data quality due to the external data
(not the "normal" SAS planning data) and its
changing nature.
MARS uses the same SAS data sources as
PRISM. This data is used to automatically populate
the UNT-DATA, UNT-FORS, UNT-MXCP, and
MNT-UNOP tables from the SAS data as
described for PRISM.
It would be an improvement to directly populate the
UNT-TRNS and UNT-CAPS tables from the GADS
event data (97 Card).
Currently the UNT-CAPS data uses appropriate
estimated values.
The UNT-TRNS values are generated using the
INT-ONLY Table, documented in version 3.00.

If MARS is to use the Transition state data (UNTTRNS & UNT-CAPS), additional development
work is needed to automate this data process.
Currently the data source is the same for both
tools.

MARS as the ability to represent the full unit
characteristics based on all recorded GADS
events, using the transition state matrix, instead of
the two-state approximations.
To determine the transition state values from the
GADS event data would require additional
development, which is not trivial.
4. EXTERNAL UNITS
 PRISM tracks all units that hold an agreement
of non-recallability to serve PJM load 1st.
 These units are included into the PJM RTO for
study purposes.
 The location of these units is an important
aspect of the modeling data so they are
placed into the correct zone when performing
sensitivity analysis.
 This is one of several operation database links
that coordinates our model with agreements /
practices used in operations.


5. FUTURE_GENS_PJM
 This is the forecast future units for both the
PJMRTO and the external world areas.
 Information is for individual units capabilities
and when they are in-service, retired, or have
modifications (up-ratings).
 The primary source of this modeling is the
PJM generation Interconnection Queues.
 The values entered in this data base reflect
the commercial probability for a given study
model.
 This data takes considerable time to develop
and verify.
© PJM Interconnection 2011. All rights reserved




Page 58 of 136

Database
File
PRISM Database
MARS Database
Comments & Comparisons
 The units are given the class Average unit
performance values based on its category
(i.e., Fossil steam, Nuclear, Combined Cycle,
Wind).
 The number of units for each neighboring
world area is based on the appropriate
reserve level per the study assumptions and
NERC Electric Supply & Demand (ES&D)
assumptions.
6. NORMAL
 This data base representation uses 21 points
to appropriately model the Normal distribution.
 This SAS data set is automatically referenced
as the default distribution to use in all PRISM
assessments.



The analogous item in MARS is the LOD-UNCY
table values.
However consideration of these values should be
in the context of consistent application with the
8760 deterministic hourly loads entered in the IN02
file. See discussion in Appendix C3.
The LOD-UNCY values are manually entered via a
text editor. However a SAS table is available
(LODUNCYDATA) for future automation efforts.


For a week's worth of daily peak values, PRISM
uses 105 values (5 week days X 21 points each
day), and MARS can use 70 values (7 days X 10
values each day).
o While the 70 MARS values can be unique
as each day peak is different, PRISM’s 105
values use 21 unique values –same daily
peak for each day.
The automation of this data for MARS analysis
has been identified as a valued effort in LOLE
assessments by the PJM staff.
7. OPERATING GENERATORS
 This SAS table has more than 30 columns of
detailed information that allows automated
input into all possible PRISM assessment
scenarios.
 This is a data base for all existing (iron in the
ground) units.
 There are over 1200 PJMRTO existing units
and 4000 World units in this data base.
 The existing units modeled in MARS have the
same SAS data source as PRISM.
 The tables populated automatically include UNTDATA, UNT-FORS, UNT-MXCP, and MNT-UNOP
tables.
 The two state unit values are typically selected for
use as the transition state matrix cannot yet be
automatically created.

The existing units modeled in MARS have the
same data source as PRISM.
However automation would need to be developed
to use the complete capabilities of MARS.



MARS Transition state matrix values for World
units would be a new development work item.
This table is a result of processing the operating
generator details table.
Most of the World unit model could come from
exchange of MARS data with neighboring ISOs.
Non Disclosure Agreements (NDA) are necessary
as data is confidential. See Operating agreement,
page 196(Section 18.17.1 paragraph B)
8. OPERATING GENERATORS DETAIL
 This SAS table is the basis for the Operating
Generator SAS table.
 Several small units, especially in the World
area, need combining so that the rounding of
units' rating to the nearest tens does not
significantly reduce the amount of available
capacity because of many small units (less
© PJM Interconnection 2011. All rights reserved


Page 59 of 136

This data source is typically only used for a few
sensitivities and assessment work.
The primary use of this data base source is to
create the operating generators table.
Database
File
PRISM Database
MARS Database
Comments & Comparisons
than 5 MW).
 This has over 30 columns of detailed
information that allows automated input into all
possible PRISM assessment scenarios.
 This is a data base for all existing (iron in the
ground) units.
 There are more units in this table than the
operating generators table.
9. ZONE MAP
 There are several assessment categories that
a given sub zone can be involved with.
 The PJMRTO and the neighboring World
region is separated into various "cut sets" so
that a given region can be assessed and
modeled correctly.
 A cut-set is a boundary line used to define a
given area (like a property line for a home).
 One of the significant items of interest is
correctly representing the diversified peak
load (the sum of individual sub zones peaks
usually do not equal a regional peak).
 This data base indicates child parent
relationships. These relationships include; a
location key, a zone number, level associated
with each location key and zone number, zone
name, and a character- based zone name
code.




The SAS table of the MARS Zone List is
comparable to the PRISM zone map.
The purpose is the same as PRISM’s which is to
map the correct load and capacity values together
based on a geographic "cut set" boundary.
The cut sets are predetermined by many in the
planning process.
MARS modeling requires that the correct load
parameters and capacity parameters go into
several more areas than the PRISM assessment
work.





The mapping of the combination of the load and
capacity data into various zones can be a tedious
process, especially for newly defined zones.
Zone map changes can have significant impacts
as this is the foundation for all modeling.
As experience grows in assessing a given zone,
the previous efforts assist in assessing the data
quality of the given model being developed.
OLAP tools can significantly help assessing,
analyzing and reporting modeling and providing
observations that are hidden in conventional two
dimensional efforts (spreadsheets).
The basis for all assessment work is the definition
of the cut-set boundaries and the resulting load
modeled in each defined area.
10. Emergency Operating Procedure (EOP) Data
 Modeled by manually adjusting the PRISM
case and external assessment work.
 There is no direct input or requirement for this
data.






© PJM Interconnection 2011. All rights reserved
Page 60 of 136
The Emergency Operating Procedures are
required for most Industry models.
This is the EOP-DATA table.
There are typically six levels modeled for the RRS
studies and interregional work with the CP-8 WG.
These values are automatically inputted into the
text file via selections in the MARS GUI.
The two SAS tables involved are the EOPDATA
and EOPNAMES tables.
The six EOP levels are: 1) Operating Reserves, 2)
Load Management, 3) 30-minute reserves, 4)
Voltage Reduction, 5) 10-minute reserves and 7)
Appeals for public curtailment.



More data and maintenance efforts are needed for
the MARS modeling.
Coordination with the Transmission Planning and
Operation departments is needed for maintaining
this data base.
Increased coordination with neighboring region
staffs will improve the data quality for external
regions in the Study model.
Database
File
PRISM Database
MARS Database
Comments & Comparisons
11. Interface Transfer Limits
 Defined in Schedule 4 (sheet 29) of the
Reliability Assurance Agreement (RAA).
 A single transfer pipe (interface) size value of
3,500 MW that reserves this amount, over all
PJM tie lines with our neighbors, for
requesting emergency assistance from the
neighboring regions at the discretion of the
Operation staff.
A set of values that define the transfer limits
between two areas defined in the study model.
These SAS tables are used to automatically
populate the INF-DATA, INF-TRLM tables.
There are two SAS data base tables: INFDATA
and INFTRLM.


The MARS_SEASON SAS data base table
indicates what months are included in which of the
four seasons, used in the reporting of results.

The granularity for the PRISM table is greater
(weeks) than that used for MARS (Months).

These default parameters are different than those
used in PRISM.
The intent is to ease the typical use and
assessment options specified in the GUI to
effectively be the same effort as built in the PRISM
GUI.
However MARS has a smaller set of defaults that
have warranted definition.
These defaults include: 1) For the CNV-CRIT table:
tolerance, margin state to solve at, minimum
number of replications, maximum number of
replications, random number seed value, and 2) for
the GEN-CASE tolerance.

Because PRISM has been used repeatedly for
many years, more efforts have been made to
reduce the tedious, repetitive tasks especially
those that can cause errors in performing
assessments done by PJM staff members.
Consideration of these efficiencies when moving
to perform comparable MARS assessments
should be evaluated as many years of efforts has
increased the assessment efficiency using PRISM.





More data and maintenance efforts are needed
for the MARS modeling.
Coordination with the Transmission Planning and
Operation departments is needed for maintaining
this data base.
Need to assess and verify MARS assessments
can demonstrate use of 3500 MW CBM value,
per sheet 29 of RAA.
12. SEASONS
 The seasons SAS data base indicates which
weeks are included in which of the four
seasons, used in the calculating and reporting
of results.
13. PARMS - Default Parameters
 There are several and various default
parameters defined to ease typical use
specified in the GUI.
 This includes the zone definitions, run
controls, sub zone ratios, zone rollups, and a
geography reference file.
 SAS data base tables involved in the default
parameters include: DM_Geography_XREF,
DM_LAS_ZONES,DM_PJM_AREA_MAP,
DM_ZONE_RATIOS, DM_ZONE_ROLLUPS
© PJM Interconnection 2011. All rights reserved



Page 61 of 136

Summary of Tables 1, 2, & 3 Comparisons
Table 1 Observations
Table 1 shows a comparison of PRISM and MARS for various parameters, modeling characteristics, data
requirements and other considerations. The most prominent comparisons are summarized as follows:
PRISM

An advantage of PRISM is its use of statistical input data requirements. PRISM has relatively
few statistical parameters that incorporate the full model. PRISM also uses a forecast
probabilistic load model.

PRISM’s solution time is faster – and thereby allows time for additional sensitivity cases to
more fully assess system impacts.

With regard to resources, PRISM does not require any additional PJM staff, beyond the
Resource Adequacy Planning (RAP) department, to collect and perform a full resource
adequacy assessment.

The major weakness of PRISM is that it is limited to modeling the “World” as a single outside
entity. This limitation forces assumptions that are not fully supported by known operational
practices. Assessment by the 1993 PJM Load and Capacity Working Group (LCWG) verified
that modeling the world as a single outside entity was appropriate for LOLE studies. Noting
that review of transmission constraints and load diversity, having reserves below 10% might
prove a benefit of modeling a multi-area world region.39

PRISM cannot perform any hourly LOLE assessments, and cannot determine LOLH or EUE.
MARS

The significant advantage of MARS is its ability to model many areas, perform hourly
calculations and include parameters more directly related to Operations.

MARS is well-supported and in wide use for resource Adequacy planning throughout North
America.

MARS is comparatively time-intensive with regard to data collection and maintenance of the
required inputs.

Due to its inherent structure using Monte Carlo replications to meet the Standard Error solution
criteria, MARS does require longer solution times.

MARS output offers robust user-defined results for various assessment requirements.

In some MARS parameters, certain assumptions and approximations are used to populate and
maintain tables to represent specific operational scenarios.

MARS can be used to comply with any new energy metrics and forth coming requirements, per
the NERC Planning Committee directives. MARS can report the identified new metrics, LOLH
and EUE. Historically these two metrics have not been reported within the electric Industry for
bulk power systems.

Load diversity can be explicitly modeled for each area in a multi-area system wide
assessment. Each area’s load characteristics are not influenced by another area.
39
Evaluation of PTI’s Multi-Area Reliability Program MAREL - November 1993 - developed by the PJM
Load & Capacity Working Group (L&CWG)
© PJM Interconnection 2011. All rights reserved
Page 62 of 136
Table 2 Observations
MARS has several load level, Emergency Operating Procedure level, and Interface flow summaries that are
not available in the PRISM output. Additional PRISM load related summaries are currently being developed
by the PJM staff to report out the intermediate values at the 21 points of the daily peak load distribution.
PRISM has capacity distribution summaries that are not available in the MARS output. The MARS output, by
processing the OT10 and OT7 files, can be used to develop similar capacity summaries which include the
use of the post processing tool Histogram. Care should be taken so that any new summaries are
appropriate for the solution algorithm used (Analytical-PRISM, or Monte Carlo-MARS).
PRISM does use a database schema both for its input and output results. This allows a mapping of the
relationships between these data and is defined in an Online Analytical Processing (OLAP) metadata
process that allows assessment and reporting of several complex summaries.
Both tools offer many detailed outputs to perform LOLE assessments.
Table 3 Observations
An often overlooked aspect of LOLE assessment work is the importance and significance of the mantra “It is
all about the data!” In these data measuring efforts all aspects of the data, typically limited by identifying
significant characteristics and resource limits, ensure high data quality. Having high data quality allows for
confidence, correct interpretation of reported results, appropriate decision making and high accuracy of final
reported values. The final values are used in markets where even small changes can cause stakeholders
concern due to cost fluctuations to adhere to Adequacy requirements. Having an established production
grade environment database process mitigates fluctuations, and allows for appropriate changes that are
verifiable, tested and approved by the proper staff authorities. It conforms to a culture of compliance with
Standards and process that demonstrate by auditing by a 3rd party that the best practices are used in the
underlying data systems used to perform LOLE assessments.
The MARS work is performed using the same main GUI and underlying database as PRISM. This GUI is in
the ARC application (see the access to models: ARC screen display and GUI layout section). Due to the
general philosophy to use a database schema for all aspects of the Adequacy modeling and results, the
MARS text file is consistent with the modeling used for PRISM. Any adjustments to the modeling, to ensure
100% consistency, are done using a text editor (Text Pad), adjusting the input files used for PRISM (.dat file)
and MARS (MIF file – in05).
PRISM and MARS are both parts of the main ARC process used by the PJM staff. In this process there are
many data relationships established to assist automating a consistent model between the two tools. High
data quality is paramount throughout the established Adequacy assessment process- especially in light of
providing the most benefit at the lowest cost for all PJM members.
There are more categories of input data for MARS than for PRISM. This allows MARS more flexibility and
the potential to perform assessments that PRISM cannot perform. It also requires increased coordination
among PJM department staff and requires more time for the PJM ITS support staff to maintain and update
the data.
One aspect to consider is that while the PJM staff attempts to perform most of the Adequacy work, MARS is
likely to increase RAAS member technical representatives’ responsibilities and efforts. This is seen in MARS
assessment work by neighboring ISOs and potentially several of the increased input categories, used in
MARS, might be best determined by an on-going dialog with member technical representatives.
© PJM Interconnection 2011. All rights reserved
Page 63 of 136
Resource & Cost Assessment
The seven “topical” categories listed below were evaluated to compare staff resources and estimated costs
of conducting the PJM Resource Adequacy assessments. The values shown are ”ballpark” estimates only
— and shown for comparison only. For determining staff resource time, the Full-Time Equivalent” (FTE)
was used; this reflects the time a PJM Staff person would take to complete a given task. The labor rate
used for this analysis was assumed to be $75.00 per hour.
1.
Annual Reserve Requirement Study (RRS)
The Annual Reserve Requirement Study (RRS) has historically only been performed using
PRISM, with all the existing facilities geared toward PRISM requirements. It would take
significantly more resources to perform the RRS using MARS; mainly due to start up efforts to
accommodate the different model specifics and the experience level of the PJM staff. An initial
MARS assessment, using engineering estimates and appropriate engineering judgment could
enable a preliminary MARS assessment to gain experience in both the required modeling
impacts and how best to capture and report the final results of the assessment. An initial
MARS assessment (using estimates) might give insights concerning further use of MARS for
the RRS.
2. Capacity Emergency Transfer Objective (CETO) Analysis
Due to several aspects for new modeling and operational specific parameters, it is anticipated
that using MARS for this would require significantly more resources. Both applications will
result in a final CETO result which could be used in the RTEP and RPM stakeholder process.
MARS efforts would need to have verification, several checks and scrutiny which accounts for
the significant increase in resources.
3. Short-Term MARS Investigations
The efforts outlined are from previously identified PJM staff work tasks. This identifies and
increases the priorities for these efforts so that they can be considered for possible completion
in the future. See Appendix G for further details on this topic.
4.
Winter Weekly Reserve Target Analysis
Currently PRISM is lacking an appropriate load model for weekly assessment work as required
in this analysis. Developing a PRISM load model useful in weekly assessments is not justified
by this sole analysis effort. MARS is coordinated with PRISM; see Table 3 concerning the
relationship of the models, so that the typical MARS assessment is consistent with the PRISM
model and other PRISM assessment work.
5. Ambient Derate Analysis
The existing UNT-DERT table in MARS has potential application to quantify LOLE impacts of
any events not already covered in GADS reporting. PRISM must address this issue by external
assessment work; requiring operational data not typically used in the other RRS assessment
efforts. Identifying and using the operational data is anticipated to be a significant effort.
© PJM Interconnection 2011. All rights reserved
Page 64 of 136
6. “World Region” Multiregional Analysis
PRISM cannot address more than two regional assessments. Through the coordination efforts
and established model relationships between PRISM and MARS, it is anticipated that this
MARS assessment would require a medium amount of resources.
7. Intermittent Resources
PRISM does not have the ability to model intermittent resources on an hourly basis.
Intermittent Resources include types such as hydro, wind and solar units. The MARS efforts
outlined represent initial efforts. At some point a thorough technical assessment should be
considered as these types of resources become a greater percentage of the PJM generation
fleet.
Table - Resource & Cost Summary
PRISM Resources
Task Description
FTE (hours)
DEV
PROD
MARS Resources
COST
DEV
FTE (hours)
COSTS
PROD
DEV
PROD
DEV
PROD
1
Annual Reserve
Requirement Study
1795
$$$
3547
2100
$$$$$
$$$$
2
Capacity Emergency
Transfer Objective (CETO)
Analysis
728
$$
3020
968
$$$$$
$$$
3
Short-Term MARS
Investigations
4
Winter Weekly Reserve
Target Analysis
5
Ambient Derate Analysis
6
7
700
$$
$
72
$
$$$
544
$$
“World Region” Multiregional
Analysis
624
$$
Intermittent Resources
948
$$$
FTE:
2680
112
$$$$
1160
Full-time Equivalent in hours … Labor rate used for cost estimation =$75 per hour)
st
DEV:
Development Environment – environment where all initial efforts are performed, 1 environment to
demonstrate initial successful results before moving to the Test environment and finally to the Production environment.
PROD:
Production Environment – final environment used for approved assessment work. Moving
applications into this environment requires adherence to a comprehensive and structured change management process.
Cost Key:
Very Small
Small
Medium
Large
Very Large
$
$$
$$$
$$$$
$$$$$
= < $10,000
= $10,000 - $55,000
= $55,000 - $150,000
= $150,000 - $250,000
= > $250,000
The above Task 1 comparison estimates did receive other ISO feedback, from SME(s). Because other’s
MARS estimates were in the general ball park, the Resource and Cost summary table values were deemed
reasonable. The other task estimates (2-7) were not reviewed further assuming they are reasonable too.
© PJM Interconnection 2011. All rights reserved
Page 65 of 136
1. Annual Reserve Requirement Study
PRISM
Task Description
MARS
FTE
Production / Maintenance
FTE
Development
FTE
Production / Maintenance
Modeling and Database efforts
595
800
640
Assessment analysis
Reporting
Training and development
520
320
120
750
560
447
480
320
167
Interaction with SME/Vendor/ Support staff
240
400
333
Duration
Identified Outside Resources
6.7 months
None
8.7 months
6.7 months
GE staff, Operations Staff,
Operations Staff, Transmission
Transmission Planning staff, ITS
Planning staff
staff
Outside Resource effort
0
240
80
New Data efforts
0
350
Successful completion Measure
Similar RRS report as 2009 in
time for stakeholder review and
use by PJM Board of Managers
Not defined (yet) ... consider
NYISO, ISONE, NPCC, MISO
reports as starting template.
Successful Completion Due date
Mid-Sept
Mid-Sept
Similar as those shown for
PRISM but needs clarification
and specifics as it applies to
additional MARS modeling and
assessments for input items.
80
Report as defined in initial effort
in time for stakeholder review
and use by PJM Board of
Managers
Mid-Sept
Similar as those shown for
PRISM but needs clarification
and specifics as it applies to
additional MARS modeling and
assessments for input items.
Milestones
Collection of Outside region data,
Generation Owner Review,
Approval of assumptions, draft
report to RRAWG for review
Total Identified FTE
1795
3547
2100
Cost (Assumed average rate)
$$$
$$$$$
$$$$
Comments
Depending on
Model
considerations
Depending on
Model
considerations
Notes:
FTE = Full Time Equivalent in hours. Labor rate used for cost estimation = $75 per hour .
Model considerations include assumptions for GADS event data, estimating regions EOP-DATA table values, estimating INF-TRLM table values, LOD-UNCY table
values, and design year values for in02 file.
Primarily data gathering efforts.
© PJM Interconnection 2011. All rights reserved
Page 66 of 136
2. Capacity Emergency Transfer Objective (CETO)
PRISM
MARS
FTE
Production / Maintenance
160
FTE
Development
200
FTE
Production / Maintenance
200
Assessment analysis
320
500
400
Verification checks
Reporting
Training and development
160
40
16
160
200
800
160
40
16
Task Description
Modeling and Database efforts
Interaction with SME/Vendor/ Support staff
32
500
32
1.5 months
None
9 months
Operations Staff, ITS staff
1.5 months
Operations Staff, ITS staff
Outside Resource effort
0
160
40
New data coordination efforts
0
500
80
Successful completion Measure
Full 24 LDA completed by TP
staff deadline for RTEP(TEAC
mtg) and RPM posting
Not defined(Yet). Consider
NYISO, ISONE, NPCC, MISO
reports as starting template.
Full 25 LDA completed by TP
staff deadline for RTEP(TEAC
mtg) and RPM posting
Successful Completion Due date
On-Going
On-Going
On-Going
Duration
Identified Outside Resources
Milestones
Coordination with PSSE
machine lists, incorporation of
latest load forecast data,
Updates of CIR values,
generation Queues,
uncommitted resources.
Redevelop automation for quick
consistency and repeatability: Coordination with PSSE machine
Coordination with PSSE machine lists, incorporation of latest load
lists, incorporation of latest load forecast data, Updates of CIR
forecast data, Updates of CIR
values, generation Queues,
values, generation Queues,
uncommitted resources.
uncommitted resources.
Total Identified FTE
728
3020
968
Cost (Assumed average rate)
$$
$$$$
$$$
Notes:
FTE = Full Time Equivalent in hours
One delivery year assessment, related to RPM or RTEP
Primarily data gathering efforts.
© PJM Interconnection 2011. All rights reserved
Page 67 of 136
Comments
This is for 25 LDAsadditional effort needed
to produce consistent
load model
Need evaluaion and
updates for EOP table
values
3. Short-term MARS Investigations
MARS
FTE
Production Environment
Task Description
Investigate NYISO Unified Methodology
80
Investigate ISONE comparison of Westinghouse model with MARS
60
Interpret and develop reporting uses for data in OT6,OT7,OT8,OT9, and OT11
files
Investigate MISO use of reporting IRM with MARS
Compare and analysis in02 file shapes reporting differences seen in 2002
calendar year, 2005 calendar year, 2006 calendar yr, and design year values.
Duration
Consider items related to neighboring region model and assess impacts
60
80
240
3 months
Cooperation with: NYISO staff,
ISONE staff, MISO staff, SERC
staff
LOD-UNCY values - Load uncertainty
80
EOP-DATA values - Emergency Operating Procedures levels
100
INF-TRLM values - Transmission pipe sizes
60
Specific Path related contracts
60
Severe case assumptions or sensitivity assumptions
60
Evaluation to aggregate supplied details
200
Successful Completion Due date
Mid-August
Total Identified FTE
700
Cost (Assumed average rate)
$$
Notes:
FTE = Full Time Equivalent in hours
© PJM Interconnection 2011. All rights reserved
Page 68 of 136
Comments
Known documenation and
SME
NERC LOLE WG agenda
item
Started as part of this paper
Design year needs
development efforts by PJM
staff
Assumed cooperation and
availablity from neighboring
region staffs
Some efforts performed on
this topic
Some efforts performed on
this topic
Some efforts performed on
this topic
Some information available
for northern paths
Development work needed as
this would be a new effort
4. Winter Weekly Reserve Target Analysis
MARS
PRISM
FTE
Development
FTE
Production / Maintenance
Modeling and Database efforts
800
40
Assessment analysis
Reporting
Training and development
320
80
240
32
16
0
Task Description
Interaction with SME/Vendor/ Support staff
Duration
Identified Outside Resources
Milestones
24
32
16
0
500
16
0
1.5 Weeks
1.5 weeks
None
None
240
8
Need to develop a new weekly
load model to replace TRUE.
0
0
Need to develop a new weekly
load model. Previously TRUE
was used, investigated use of
model used by ITROM
application used by load
forcasters.
0
New Data efforts
Successful Completion Due date
Need to develop a new weekly
load model to replace TRUE.
FTE
Production / Maintenance
12 months
Operations Staff, Load
Forecasting Planning staff, ITS
staff
Outside Resource effort
Successful completion Measure
Comments
500
Similar RRS reported section Similar RRS reported section
as in 2009. In time for for PC
as in 2009. In time for for PC
stakeholder review and
stakeholder review and
recommendation to OC before recommendation to OC before
Winter period.
Winter period.
Mid-Oct
Mid-Oct
Completion of RRS base case Completion of RRS base case
and sensitivity analysis,
and sensitivity analysis,
producing MARS model using producing MARS model using
ARC
ARC
Similar RRS reported section as
in 2009. In time for for PC
stakeholder review and
recommendation to OC before
Winter period.
Mid-Oct
Completion of RRS base case
and sensitivity analysis,
producing MARS model using
ARC
Total Identified FTE
2680
112
72
Cost (Assumed average rate)
$$$$
$
$
Notes:
FTE = Full Time Equivalent in hours
Model considerations include assumptions for GADS event data, estimating regions EOP-DATA table values, estimating INF-TRLM table values, LOD-UNCY table values, and
design year values for in02 file.
Primarily data gathering efforts.
TRUE = Temperature Related Universal Effect - load model
© PJM Interconnection 2011. All rights reserved
Page 69 of 136
5. Ambient Derate Analysis
PRISM
FTE
Production / Maintenance
Task Description
MARS
Comments
Need operational data
and compilation of
statistical measures
FTE
Production / Maintenance
Statistical Modeling and Database efforts
1000
Use deration matrix to assess Adequacy impacts
N/A
160
80
40
4 Months
Operations Staff
40
0
Reporting similar to previous
efforts
Mid-Jul
80
40
2 Months
Operations Staff
24
0
Establish method, evaluation
scenarios
Mid-Jul
Total Identified FTE
1160
544
Cost (Assumed average rate)
$$$
$$
Reporting
Training, development, documentation
Duration
Identified Outside Resources
Outside Resource effort
New Data efforts
Successful completion Measure
Successful Completion Due date
Milestones
Notes:
FTE = Full Time Equivalent in hours
N/A = Not Applicable
© PJM Interconnection 2011. All rights reserved
Page 70 of 136
240
Comments
Need operational data
and compilation of
statistical measures
MARS uses the UNTDERT
6. "World Region" Multiregional Analysis
MARS
FTE
Production Environment
Task Description
Create initial single area that is consistent with model used in PRISM for the
neighboring region.
Divide single area model into its components, based on Operational
considerations, and Interregional group's modeling. Verify consistentcy for
infinate Transmision limits and sharing of resources
Develop all needed modeling table values, UNT-DATA, INF-TRLM, EOP-DATA,
LOD-UNCY, RES-SHAR
Training, development, documentation
Reporting
Duration
Identified Outside Resources
Outside Resource effort
Coordinate data model with neighboring regions
Successful completion Measure
Successful Completion Due date
Milestones
120
160
160
40
40
2 Months
MISO Staff, SERC Staff,
NPCC Staff, NYISO Staff,
ISONE Staff
24
120
Establish method, evaluation
scenarios
Mid-Aug
Total Identified FTE
624
Cost (Assumed average rate)
$$
Notes:
FTE = Full Time Equivalent in hours
N/A = Not Applicable
PRISM is not capable to perform this multi-Area assessment
© PJM Interconnection 2011. All rights reserved
Page 71 of 136
Comments
7. Intermittent Resources
MARS
FTE
Production Environment
Task Description
Develop modeling method for wind units, using hourly profile(s).
Develop modeling method for Hydro units, based on GADS data.
Investigate and appropriately use previous assessment work documented in
technical assessments (EWITS) and other ISO methods.
Investigate Manual 21, Appendix B method and calculations for existing wind
units. Evaluate Adequacy impact of using a perfect performing unit at a reduced
rating in comparison of availability shown in hourly profiles for full nameplate
rating.
Training, development, documentation
Reporting
Duration
Identified Outside Resources
Outside Resource effort
Coordinate data model with neighboring regions
Successful completion Measure
Successful Completion Due date
Milestones
120
160
160
320
60
60
2 Months
MISO Staff, GE Staff, NYISO
Staff, ISONE Staff
8
120
Establish method, evaluation
scenarios
Mid-Dec
Total Identified FTE
948
Cost (Assumed average rate)
$$$
Notes:
FTE = Full Time Equivalent in hours
N/A = Not Applicable
PRISM is not capable to perform this hourly model assessment
© PJM Interconnection 2011. All rights reserved
Page 72 of 136
Comments
Frequently Asked Questions (FAQs)
In the conduct of this study, there were many questions by RAAS participants and Stakeholders. These
questions were placed into five categories as follows:
Category 1: MARS as the primary LOLE tool
Q1. What factors most concern PJM for replacing PRISM with MARS? What impacts would be seen in
time and economic resources?
PRISM is a well-defined, valid and thoroughly known product for PJM that offers its staff great
efficiencies in performing assessment work due to its ongoing development over many years. Based on
an initial assessment, PJM Staff has determined that performing MARS assessments will ultimately
yield the same (or very similar) results as PRISM. Study outputs are defined in the PJM agreements
and marketplace requirements (see the RAA, RPM Manual 18 and RTEP Manual 14B).
A detailed MARS assessment is considered essential, due to several technical issues involved in
translation of PRISM load modeling to MARS modeling. There are many details to consider for full
MARS implementation at PJM; all such related study efforts have not yet been uncovered or fully vetted
through the technical stakeholder process.
For further information, refer to the following sections: Comparative Strengths, Complementary Aspects
of PRISM and MARS, and Ongoing LOLE Assessment Work.
Q2. Why is PJM reluctant to replace PRISM with MARS?
PJM is not reluctant to use MARS as a tool, as it started to implement MARS in 2004. PJM actively
uses MARS in performing LOLE assessment work. PJM staff has been involved with Industry models
and assessments for many years; it is PJM’s position that the proper tool should fit the task at hand. In
these efforts, PJM Staff invites opportunities to review the other known Industry tools such as Gridview,
SERVM, and the Westinghouse model.
In PJM’s view, no single tool emerges with a “one-size fits all” advantage for all the LOLE assessment
desired by the PJM staff and its Stakeholders. In evaluating and recognizing the complementary aspect
of these modeling tools, PJM Staff strives to employ a consistent and yet flexible assessment strategy.
PJM Staff actively implemented MARS on the PJM production grade systems which started back in
2002 (with previous efforts to evaluate a multi-area tools launched in 1987). For the past decade, PJM
has recognized the benefit and value that a multi-area tool such as MARS brings to LOLE assessment
work.
See sections Comparative Strengths, Complementary Aspects of PRISM and MARS, and Ongoing
LOLE Assessment Work for further details.
Q3. Please recap the discussion at the December 3, 2010 RRAWG (predecessor to RAAS) meeting
concerning the MARS solution process.
Mr. Glenn Haringa, of General Electric International, Inc. (GEII), the developing company of MARS,
joined the December 3, 2010 RRAWG meeting to clarify and explain aspects of the MARS solution
process.
A.
How does MARS’ “pipe” values come into play for transmission limits? Please relate
this to FCITC values determined in load flow assessments.
© PJM Interconnection 2011. All rights reserved
Page 73 of 136
There are load flow based assessment
techniques that can help determine the pipe
size (summation of Individual Interface tie
sizes) transfer limits used by MARS. In
addition to observing the limits on the
individual ties, MARS also allows you to
define interface groups (input table INFGRPS) which allow for the modeling of
simultaneous transfer limits. The interface and
interface group transfer limits can be varied
hourly based on the availability of specified units and the value of area, pool or system loads
(input table INF-DYLM per MARS User Manual). The degree to which the interface and
interface group limits are modeled can increase total solution time.
The extracted image above shows an example of interface groupings and dynamic limits for
the Long Island Area in the New York Control Area (CA).
B.
How does MARS address or model diversity in its assessment?
The load diversity within a given regional model is determined by the timing of the loads in the
various subareas that comprise the regional model. Due to different geographical parameters
for a given hour, not all subareas typically peak at the same time, MARS inherently addresses
this by its use of an hourly chronological load shape (in02 file), specified for each modeled
subarea. The issue of diversity is addressed early in the MARS modeling process with a
separate and up-front assessment to determine the appropriate 8760 hourly loads (load
shape). The area’s month-to annual ratios (LOD-MTAR table), as determined by load
forecasting reports, can also influence these deterministic hourly load shapes.
C.
Are there any tools or techniques for choosing the 8760 hourly load values?
Hourly loads determine MARS’ load shape characteristics. The selection of the hourly loads to
use involves performing assessment work related to load forecasting combined with historical
peak demand periods. This effort includes weather and economic uncertainty (I’m not sure
what these would include). This assessment is uniquely focused on the LOLE modeling of
peak load demand, but associated with load forecasting. In this assessment are considerations
of the demand periods being both severe (high) and also high for a long duration. Different
operational concerns exist for single peak demand periods as opposed to many peak demands
over extended time periods (i.e. high daily peaks for two weeks).
The Industry groups cited in the Interregional Assessments section of this report have
practices and established methods to choose this input modeling parameter. This topic has
been discussed at length at the NPCC CP-8 WG, and due to different regional operational
issues, the techniques vary.
D.
How does MARS uses its linear program algorithm to reach a solution? Explain how it
performs reserve sharing but not load loss sharing.
The presentation slides from the December 3, 2010 RRAWG meeting included three slides (9 12) that illustrated the MARS solutions and the Standard Error (These slides are Figures 6A6C in this report). An orange line shows how the LOLE solution relates to the Standard Error.
As Mr. Haringa explained, the solution (number of replications) depends on, among other
factors, the reliability of the system being modeled. A more reliable system therefore has a
slower convergence.
The MARS solution results are unique because of the algorithm used, and repeatable due to
the way in which the random numbers that are used to simulate random events are generated.
In this system- wide solution for large regional models, the default is to share assistance
between areas on an equitable basis. This approach mitigates individual areas’ LOLE states,
without the need for a pre-determined priority. This approach prevents an area from providing
assistance to a deficient area to such an extent that it negatively impacts the LOLE state of the
assisting area. This avoids making a decision as to which areas have priority in reserve
© PJM Interconnection 2011. All rights reserved
Page 74 of 136
sharing over others in receiving assistance. However, MARS has capability of specifying a
priority list for assigning known assistance priorities among the areas that comprise the large
regional model (Table RES- PRIO).
The MARS solution does not model or simulate the ability to share loss of load events among
areas. In a consistent manner it is important to note that the PJM RTO or other balancing
authorities do not invoke load shedding to provide emergency assistance to neighboring
control areas.
E.
From an academic viewpoint, theoretically, can MARS have more than one unique
solution for a large regional model?
A given MARS model, as specified by the input data, will have one unique solution. A large
regional model has many moving parts, involving generation, load, and transmission. These
work closely together in a coordinated, consistent and robust basis. In a bulk grid system there
are several modeling techniques for the successful grid operation (i.e. increasing a unit’s
output can alleviate a transmission limit).
As several coordinated, interrelated modeling aspects are addressed, there can be different
but appropriate results due to the different handling of the modeling in the solution process. In
general, these different scenarios and approaches yield similar trends and insights’ regarding
Bulk system needs and attributes. By choosing different subareas and regional cut-sets, the
total regional model does not change. Defining different subareas or redefining attributes of the
same subareas of this total region can however give different results. The solution depends on
the input data which is dependent on the study assumptions.
The total regional LOLE is associated with the lowest common denominator or the subarea
having the highest LOLE. If a subarea with a high LOLE is combined with a subarea with a
very low LOLE, the overall regional LOLE can improve significantly, but you have effectively
eliminated the transfer limits that had previously existed between the two areas. For example
in a system wide model, there are 4 areas each with 20% IRM and a fifth area with 5% IRM. If
these areas are re-combined into 3 areas, the resulting LOLE could be much lower than the
five-area model (the fifth area drives the resulting LOLE). However, the choice for planning
models is dependent on actual operational limits and methods used by the Dispatchers in
controlling the Bulk grid. (I.e. so combining the 5 areas into 3 might be inconsistent and
inappropriate to match operation practices).
Many scenarios are performed in Industry studies that address various modeling assumptions.
Upon review by experienced Engineering staffs and technical oversight groups (such as the
RAAS ) a comprehensive interpretation of all assessments can be given for use in Planning,
Markets, and Operations.
Q4. What steps are needed to encourage adoption of MARS as the LOLE tool of choice, going forward?
PJM staffs have currently implemented and use MARS. Over the years historic PJM groups have
endorsed a multi-area modeling tool. Due to resource and cost considerations, PJM has maintained its
internal PRISM tool as the primary tool. Over time some stakeholders have come forth to promote
MARS as the primary tool. However other stakeholders express caution indicating that PRISM
addresses the current Adequacy needs. The stakeholder process should be deliberate and
comprehensive, on technical grounds –with consideration of resources. Many stakeholders believe that
PRISM is a proven tool which meets PJM Adequacy needs in a cost effective and timely manner.
It is acknowledged, on technical merits, PRISM has limitations –and that MARS is needed to address
those limitations. While all stakeholders do see merits and strengths in using MARS, the debate is if
MARS should replace PRISM fully or if the current hybrid system (using both tools) be maintained.
Appendix G outlines several items for possible future consideration. The Resource and Cost
assessment section indicates the level of commitment needed to fully implement MARS to replace
PRISM. If some of the identified efforts can be prioritized and sub divided into phases, the transition
toward a full MARS implementation could be realized with a gradual cost. This subject was not the
focus of this technical comparison: providing any conclusions or recommendations for next steps was
avoided, as directed by the RAAS.
© PJM Interconnection 2011. All rights reserved
Page 75 of 136
At the first MARS user group meeting (October 2010), GEII staff announced the development of a stateof-the art GUI, database, and output reporting for MARS. Currently MARS has a text based input and
output format that comes from its 1980’s development. Improvements of the efficiency and
effectiveness of any new MARS product; using a state of the art GUI, database schema, and output
reporting which significantly enhances the PJM staff use; might prove a good point in time for further
incorporation into PJM’s ARC application.
The interregional planning efforts involving PJM staff might also prove to be the catalyst for MARS to
evolve as the primary tool at PJM.
Category 2: MARS resource needs and computing speed
Q5. Does running MARS take more “horsepower”, including computing power and manpower, to perform
studies?
Well, yes and no. It depends on what organization is running the MARS assessment work, the
experience and skill of the staff involved in the study, and the specific assessment at hand. MARS does
need many specific discrete table entries that might require additional assessment work compared to
PRISM.
Experienced staff performing MARS LOLE assessment work can considerably mitigate any burden due
to needed extra effort. For legitimate concerns in the output tables, deeper investigation into the issue
can be time-consuming and yet, cannot be avoided.
PRISM does have an advantage in that its input parameters are statistically-derived parameters from
the base information that both applications need and use. These statistics are defined in the Industry,
like IEEE Standard 76240, and result in fewer input sets required for the analysis to begin. Strict process
adherences to Standards ensure a high level of accuracy in application of the PRISM calculations.
However these statistical quantities do require an experienced staff, versed in statistics, and their
application in order to properly conduct a Study.
Q6. Could MARS runs be processed faster (with fewer resources) than what GE or PJM has already
indicated? – (As shown in the Resource and Cost Assessment Section)
As part of the “Next Steps” process, test runs can be made to perform some basic run-time
comparisons. Work of this nature can be rather complex and labor-intensive – and it would have to be
scheduled and approved by PJM management. Resources required to do such studies can vary widely
and there are many factors that can govern completion. The numbers referenced in this report have
been reviewed by PJM Staff involved in LOLE assessment work, and by other SME experts using
MARS – and are PJM’s best estimates per that review. Other industry SMEs experienced with MARS
were also solicited for their feedback about these estimates based on their MARS experience – and that
feedback is shown in the discussion about these best estimate values.
40
IEEE Standard 762 refers to IEEE Standard definition for use in reporting electric generating unit reliability, availability and
productivity. IEEE Std-2006 (revision of IEEE STD 762-1987) - approved 9/16/2006 by IEEE-SA Standards Board.
© PJM Interconnection 2011. All rights reserved
Page 76 of 136
Q7. What important details cause MARS to have long run times?
The MARS solution is iterative, as it requires a Monte-Carlo solution, while the PRISM solution is a
direct table look up technique. It would be expected that PRISM would have a faster solution time. This
is similar to how a load flow (i.e. Newton-Raphson) technique compares to a short-circuit study (i.e.
Thevenin equivalent) technique. The question is then: Are the two tools comparable in solution times?
Many modeling and specifics of a given assessment come into play for computer resources needed to
reach a solution. The issue discussed concerning dynamic interface transfer limits indicates a MARS
increased solution time. Not every study requires this modeling. If it is desired to compare single area
solution times, it would be expected that the MARS solution would have comparable solution times as
PRISM, with both solving in a reasonable time period (i.e. in a few minutes).
Figure 3C shows the overall MARS solution steps with step 7 iterating, due to dynamic transfer limits,
causing a slow solution for a large region, many area model (Process flow step: Calculate Emergency
assistance and resulting area margins which involves the dynamic interface transfer limits.) .
Q8. What are the details of why MARS requires much more staff time than PRISM?
PRISM has several staff efficiency advantages; 1) an established database schema; 2) many
automated administrative functions for maintenance and 3) user GUI options. MARS has been
incorporated into this environment, but in a supporting role. MARS’ additional data needs (more
granular modeling) require more work for the PJM Staff. This is primarily due to the Planning,
Operations and Market division staffs having an individual focus for detailed assessments. Coordination
of these individual PJM divisional assessments is currently done on a higher level (don’t get lost in the
weeds) to ensure consistent modeling.
As stated in this report, PRISM models are not necessarily suited for operation specific interpretations.
This mitigates detailed discussions and efforts about specific modeling parameters between the various
PJM divisions. However MARS models can be suited for operations specific interpretations which would
increase coordinating efforts with potential changes to accommodate underlying data base schema
changes.
Q9. What activities are done for the hours listed in the MARS Resource and Cost Assessment section?
They seem way too large?
In looking at the first task in the Resource and Cost Assessment section, the Annual Reserve
Requirement study, there are several staff members involved who comprise various skill sets and
departments. These efforts require close coordination so that the desired Engineering technical
assessment capability be performed effectively and efficiently. Further details concerning the rows
shown: Modeling and database efforts; Outside Resource effort; Interaction with SME/Vendor/ Support
staff; are as followings. The value in parenthesis indicates how many individuals are involved.
Task Description(# of staff)
Administrative Staff (2)

Document editorial review

Posting, organization, and administration of PowerPoint, spreadsheet, word
document
Engineering Assistant(2)

Straightforward editorial review

Create Spreadsheet checks for data and assumptions

Create spreadsheet summaries of detailed results.

Create documentation items(screen Captures, How-To)

Measure and validate correct data
Database Administrator(2)

Run Extract-Transfer-Load (ETL) processes.

Develop and maintain SSIS routines.

Coordinate with RAP staff to maintain data source changes.
SAS Admin and ITS support staff(2)

Move SAS data
© PJM Interconnection 2011. All rights reserved
FTE Hours for all
staff / Annual
period
40
400
160
320
Page 77 of 136

Trouble shoot GUI/data issues

Create Change Management tests

Oversee Change Management process

Maintain SAS libraries/schemas/Tables

Coordinate ITS response to Users

Coordinate with RAP staff to maintain data source changes.
GUI / OLAP developer(1)

Develop new data processing

Maintain Data processing

Assist in moving solutions to PRD

Document and coordinate proposed changes with SME (RAP,CA,ITS)
Documentation(4)

Adhere to ITS Standards documents and format

Adhere to ITS Standards for coding

Create “How-To” documentation

Maintain existing data documentation
Testing and Change Management(5)

Development environment

Test Environment

Production Environment
Internal PJM data gathering and cleaning(4)

Load forecast

Generation Forecast

Transmission Forecast

EOP

Contracts
Data gathering for external sources: NERC, FERC, SERC, NPCC, MISO (3)

Load forecast

Generation Forecast

Transmission Forecast

EOP

Contracts
Data cleaning of external sources(2)

FERC 714 process Hourly Loads

NERC ES&D data

NYISO MARS Data

ISO-NE MARS data

NERC LTRA report

Transition to Old NERC boundaries(Pg 34-36 of 2010 RRS)
Coordinating with Operations staff(2)

Maintain data and review with Operations staff

Develop documentation of new data and review with operations staff to
establish process.
Additional Modeling of PJM control Areas(4)

Transmission planning assessments
a. Significant development effort requiring limited Engineering staff
resources.
b. Develop a MARS multi-area model that breaks the PJMRTO into is
discrete control areas which are driven by known operational
constraints.
Resource Adequacy Planning Senior Engineering Staff oversight (3)

One day of review per week over critical 3 month modeling period.

Provide examples
Formal meetings to coordinate modeling among Resource Adequacy Planning Staff,
Transmission planning department, Operation Staff, database administrators, SAS
developers, Java GUI developers, and staff involved in Change Management.(5)
Working meetings to informally gather in same room to share and draw on other’s
viewpoints while each individual task is be performed – aides in producing a consistent
product and keeping all involved on the same page.(5)
© PJM Interconnection 2011. All rights reserved
Page 78 of 136
160
80
80
30
60
60
24
320
250
120
416 – This is not
additional time, but
time allocated for
performing some of
the above tasks.
While this is not meant to be comprehensive, the above attempts to provide verification of the number of
hours shown in task 1 of the Resource and Cost assessment section. Similar details were provided by
the other ISO staffs verifying that the PJM estimates are in line with other neighboring staff efforts.
Q10. What incremental efforts are meant to be done by PJM related to items listed in the MARS Resource
and Cost Assessment section?
Although the PJM staffs are currently using MARS for several assessment topics, the full capability of
MARS within PJM’s system is not yet fully functional (i.e. it cannot replace PRISM). Items needing
improvement include:
Automated GUI parameters and options,


Addressing missing data in the model,

Maintenance of data and assessment methods,

Incorporation of other ISO’s MARS methods into an appropriate adaptation for PJM,

Measuring and calculation techniques for all the various MARS parameters,

Ensuring the input parameter package is complete and balanced.
A specific item would be to break the PJMRTO model into its internal Control Areas (six or more), from
the current four-area model. This involves Operations and Transmission Planning Staff efforts. There
are many significant non trivial efforts that would need to be done. An initial list is shown in Appendix G.
These are possible examples for what incremental efforts could be performed, with these items part of
the estimates in the Resource and Cost section.
Category 3: MARS Advantages
Q11. Does MARS represent a big advantage due to its obvious multi-area capability?
MARS not only has an advantage in multi-regional study work but can also perform hourly
assessments. MARS can model many specific operational events for close-in assessments. PRISM
should NOT be used for operational planning of bulk system requirements as its statistical methods may
or may not comply with specific actual operational experience. Typically the operational planning groups
use different techniques better suited for near term operational requirements. PRISM is a probabilistic
model well-suited for planning of future bulk grid requirements. Along with multi-area capability, the
ability to perform hourly chronological assessments and LOLH and EUE calculations will be important to
address reporting, considering the NERC-PC is requiring these items to be reported at a future date.
MARS Monte Carlo solution technique and its discrete inputs allow for evaluation of specific operational
conditions, with reported results usually indicating the number of invocations of a given Emergency
Operation Procedure (EOP) level or a confidence level of the observed assessment conditions.
Q12. Would the full implementation of MARS enhance PJM’s contribution to the interregional planning
initiatives such as the Interregional Planning Stakeholders Advisory Committee (IPSAC)?
Since 2005 PJM staff has used MARS to complement PRISM (see Q19 for more details).
PJM Staff is actively involved in performing these assessments and functions. At times, PJM staff takes
the lead in running all appropriate analysis and reviews such joint assessments. For some of these
efforts, PJM gathers all the modeling data and implements the MARS solutions – while at other times,
the other ISO staffs take project leads. The lead on any such efforts depends on available resources,
modeling and assessment expertise, given priority for a given ISO, availability of application tool that
best fits study objectives. MARS is useful for determining adequacy metrics such as LOLE, LOLH, and
EUE.
The IPSAC website includes agendas and details about these efforts. Other interregional efforts include
the NPCC CP-8 WG, and SERC’s RAWG, in which all interregional groups are currently using MARS.
http://www.pjm.com/committees-and-groups/stakeholder-meetings/stakeholder-groups/ipsac.aspx
© PJM Interconnection 2011. All rights reserved
Page 79 of 136
Q13. What is MARS’s biggest advantage?
MARS’ biggest advantage is that it can model an unlimited number of defined areas. This feature is
useful in assessing transmission system conditions and impacts on resource adequacy metrics such as
LOLE, LOLH, and EUE.
MARS does have other advantages, if: 1) the underlying hourly load data, 2) GADS events and 3)
operational data is available.
o
It can explicitly model operational events. This allows for adequacy assessments for
measuring various changes and scenarios to these events.
o
The numerous entries allow for reporting out on specific Operational define levels, in
accordance with defined Emergency Operating Procedures (EOPs).
o
MARS can capture, through its detailed input tables (See Appendix F), energy related
modeling specifics. This includes intermittent resources (Wind), demand side management
programs, and load shedding.
The typical MARS modeling practice of specifying many areas for the surrounding neighboring systems
is typically done by Interregional assessments. If transmission limits between the defined areas are
unknown, there could be significant cost and effort involved.
Category 4: Clarify MARS limitations
Q14. Why is the MARS’ output in flat text files (which have to be cut and pasted into Excel) for further
assessment efforts?
MARS has both Text and Binary final outputs. However almost all of the results, processed for reporting
LOLE results and calculations, are in the text format files. The binary file format is very useful in using
the histogram supporting application, to process the replication and capacity distributions. The PJM
Staff does not have or know of a better method to process the many MARS results in the various MARS
output tables. Using a manual text editor tool is an efficient method for single, one-off type of reporting.
However for more complex or iterative processing, this is manually intensive. There would have to be
process to define and store a database schema, for the various MARS output results. Similar to what
has been done for PRISM; this would mitigate the PJM manual efforts in processing and coordinating
results.
Potential changes to the underlying data base schema require costly and significant resources.
Q15. Why is MARS is limited to a 5 year look-ahead?
MARS does not have any time period limitations, from a modeling, calculation or processing
perspective. Results can be reported for any time period deem appropriate, by the user.
Other Control Areas report MARS results for a ten year forward period (i.e. CAISO, MAPPCOR).
However, as seen in the recent NERC PC decisions about approved metrics (by GTRPM Task Force),
operations based metrics (energy) are appropriate for either a 2 or 5 year look-ahead. The uncertainty
in forecasting values beyond that time period was deemed inappropriate.
MARS requires operational specifics to forecast the reported metrics (i.e. LOLE, LOLH, and EUE).
PRISM does not need these operational specifics, as its model uses forecast probabilistic parameters
for the individual components of load, capacity, and transmission. Hence the convolution and resulting
LOLE states are not solely contingent on any underlying operational condition or event. Note that
whatever the tool used the uncertainty of the reported forecast values increases as a function of time.
At some point the uncertainty of the reported metrics is deemed to be either not useful or causes a
misunderstanding of the reported information. MARS has been seen to be highly useful in the Industry
to report out clear, valid metrics 5 years into the future, based on the PJM Staff’s interregional study
experience to date.
© PJM Interconnection 2011. All rights reserved
Page 80 of 136
Q16. Why does MARS has many more input tables compared to PRISM?
Although both MARS and PRISM have complex, and specific input data requirements, MARS does
have 50 input table categories while PRISM has 23 input control card categories. Of the MARS 50 input
table categories approximately 22 tables, comprising 121 parameters, are typically in a large regional,
multi-area model. See Appendix F for further details.
While these MARS input details allows many variations and sensitivities to be assessed it also
necessities a detailed database schema to track and maintain any data changes, especially any
operational related parameters.
Category 5: PRISM Advantages
Q17. What is PRISM’s biggest advantage?
PRISM’s biggest advantage is that it is a quick and accurate solution, using a limited amount of
statistical input parameters (EEFORd, EFORd, Variance, and POF).
PRISM does have other advantages, if: 1) the underlying hourly load data and 2) GADS event data is
available.
o
PRISM’s use of statistical parameters mitigates any re-evaluation of the discrete
events/observations. These statistical quantities are Industry proven and have been
developed over many decades in the context of application of correct mathematical and
numerical techniques.41 42
o
The numerous entries required in MARS can lead to significant investigation and re-occurring
sensitivity analysis to quantify impacts in each study. For example, PRISM’s use of a single tie
size is an advantage in that extensive analysis was conducted in determining this value, and
this robust analysis continues per FERC requirements to report a Simultaneous Import Limit
(SIL) with the resulting Capacity Benefit Margin (CBM) value being one value agreed to in the
stakeholder process and stipulated in the PJM Agreements. It is clear and unambiguous what
CBM value should be used for the transmission pipe size. This removes any need to
investigate “what if” analysis, saving the PJM members time & money in each annual study.
PRISM’s modeling practice of placing the entire surrounding neighboring systems into one region has
been evaluated and discussed by the RAAS in-depth. A recent White paper, titled “World Modeling
Region – Technical Issues” on this subject is on the PJM web site at:
http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/res-adeq/20090515-worldmodeling-region-tech-issues.ashx .
It is recognized that several of the assumptions of this modeling can be investigated for improvement by
evaluation and assessment work using MARS’s multi-region capability.43
Q18. What specifics concerning PJM’s blended approach is used by both PRISM and MARS? Please
specify when each tool is used and the reason behind using this specific tool for the identified task
at hand.
PRISM (including the PLOTS load model) is used as the primary tool for: 1) annual RRS reporting,
CETO reporting, 2) assessment of the load model time period, and 3) most resource adequacy
sensitivity work. PRISM calculates the LOLE metric for demonstrating adherence to any defined criteria
or standard. PRISM and PLOTS are based on the existing database schema, which includes the
established data gathering efforts. Both of these tools are integrated into a GUI that increases the
efficiency and quality of PJM staff assessment efforts. There is currently no non-annual load model
(Seasonal or monthly) available as previous efforts were not carried over into the new SAS, ORACLE,
41
Power System Reliability Evaluation, - Gordon and Beach, Science Publishers, –1970 – by Roy Billinton
Probability Calculation of Generation Reserves - March 1969 - by C.J. Baldwin and published by The Westinghouse Engineer
Evaluation of PTI’s Multi-Area Reliability Program MAREL - November 1993 - developed by the PJM Load & Capacity Working
Group (L&CWG
42
43
© PJM Interconnection 2011. All rights reserved
Page 81 of 136
JAVA systems. The previous non-annual modeling was deemed to be not practical for implementation
going forward. Other modeling is seen as more appropriate (i.e.: Fuzzy logic techniques). This makes
PRISM useful for annual assessment efforts, as no seasonal or monthly focused modeling is done with
PRISM. The underlying database schema is foundational in nature. Changes to the database are
therefore resource intensive, and costly.
MARS is used for: 1) the winter weekly reserve target, 2) Wind and DR type assessment that need
hourly patterns that are energy based, and 3) Interregional coordination efforts with neighboring ISOs.
Another potential application for MARS could be to analyze the ambient impact on units and the
relationship to the LOLE. MARS is used in the winter target assessment as PRISM does not have a
non-annual probabilistic load model. MARS is used in the interregional coordination efforts as a
consistent tool avoids translations issues between different tools. MARS will be used to satisfy the
upcoming NERC reporting requirement of LOLH and EUE metrics.
For further discussion about why each tool is used, please see the section Complementary aspects of
PRISM and MARS. For further information about what is performed please see the sections Ongoing
LOLE assessment work and Database modeling relationship matrix.
© PJM Interconnection 2011. All rights reserved
Page 82 of 136
Interregional Assessments

NPCC Regional Assessment
The Northeast Power Coordinating Council (NPCC) conducted an Interregional Long Range Adequacy
44
Overview that used PJM’s 2005 RRS model as the basis for the modeling used for PJM in this MARS
assessment. This study is an example of a regional assessment to demonstrate a comparison of MARS results
with PJM Planning Division results. The graphs below reflect the results of this analysis.
The chart below is from Page 34 of this report and illustrates how LOLE increases through the Study time period.
During 2004-2006, PJM Planning Division assessments indicated reliability needs for system enhancements and
facility upgrades. Significant changes were planned as recorded in the RTEP process and the implementation of
the Reliability Pricing Model (RPM) marketplace. This chart, from MARS assessment work was consistent with
other PJM Planning Division work and gave appropriate valuations of impacts on an LOLE basis.
44
Northeast Power Coordinating Council “Interregional Long Range Adequacy Overview” (November 28, 2006), conducted by the
NPCC CP-8 Working Group. The full report can be obtained on the NPCC website
© PJM Interconnection 2011. All rights reserved
Page 83 of 136

IPSAC Regional Assessment
PJM participates in interregional planning activities with all its interconnected Planning Coordinator
neighbors and all Planning Coordinators throughout the Eastern Interconnection pursuant to various
interregional agreements. Two of these agreements specify the formation of an Inter-regional Planning
Stakeholder Advisory Committee (IPSAC) for the purpose of allowing for review of and input to
coordinated system planning activities by all stakeholder groups.
o
The Joint Operating Agreement between PJM and the MISO
o
The Northeast ISO/RTO Planning Coordination Protocol among PJM, the NYISO and the ISONE.
PJM activities related to these IPSAC groups and the related interregional agreements can be followed
at the web pages for the NYISO, MISO, PJM and others. Related meetings and information that are
organized and run by PJM will be shown as PJM meetings in the space below. Notice and information
of IPSAC meetings organized and run by others will be disseminated via email to the PJM IPSAC, PC
and TEAC “exploder” lists. Registration information for meetings run by others will be included with
those emails.
IPSAC past meeting materials are posted at: http://www.pjm.com/committees-and-groups/stakeholdermeetings/stakeholder-groups/ipsac-ny-ne.aspx ( see 12-18-2009 Agenda item “Interregional
Planning”, and 06-30-2009 Agenda item “NYISO / PJM focused Study” ).
IPSAC meeting materials related to the Midwest ISO are posted at: http://www.pjm.com/committeesand-groups/stakeholder-meetings/stakeholder-groups/ipsac-midwest.aspx .
PJM staff has performed MARS assessments which involved up to a 28-area model of the Northeast
(PJM, NYISO and ISO-NE) in these efforts. Some figures and summaries concerning MARS methods
and solutions in this Report are extracted from these detailed Industry modeling efforts.

SERC Reliability Assessment Working Group
The PJM staff participates in the SERC Reliability Corporation’s45 , Reliability Assessment Working
Group (RAWG). This does include review of MARS modeling and assessments performed for the
SERC region and its sub-regions. This modeling includes the neighboring- external systems. The
PJMRTO model represents portions of the SERC sub regions and a portion of the SERC neighboring
external region.
Participation in the SERC RAWG requires a signed NDA and adherence to confidentially rules as
defined in the SERC Reliability Corporation’s stakeholder process. PJM Staff has access to the secure
RAWG portion of the SERC web site due to compliance with these rules.
The public SERC web site is at: http://www.serc1.org/Application/HomePageView.aspx
45
SERC is one of nine regional electric reliability councils under North American Electric Reliability Corporation (NERC)
authority. SERC was formed in 2005, as the successor to the Southeast Electric Reliability Council (also known as SERC).
© PJM Interconnection 2011. All rights reserved
Page 84 of 136
Glossary
Also see PJM Manual 35, Definitions and Acronyms at:
http://www.pjm.com/~/media/documents/manuals/m35.ashx
Adequacy – the ability of a bulk electric system to supply the aggregate electric demand and energy
requirements of the consumers at all times, taking into account scheduled and unscheduled outages of
system components. One part of the Reliability term.
Applications for Reliability Calculations (ARC) – ARC is the principle computer environment that allows for all
LOLE Adequacy assessment work to be performed. There are five parts of this application: 1) Week Peak
Frequency (WKPKFQ); 2) PRISM; 3) MARS; 4) Administration (Security and database); 5) PIW view (PJM
Information Warehouse).
Bulk Power System (BPS) – The Bulk Power System (BPS) refers to all generating facilities, bulk power
reactive facilities, and high voltage transmission, substation and switching facilities. The BPS also
includes the underlying lower voltage facilities that affect the capability and reliability of the generating and
high voltage facilities in the PJM Control Area. As defined by the Regional Reliability Organization, the
BPS is the electrical generation resources, transmission lines, interconnections with neighboring systems,
and associated equipment, generally operated at voltages of 100 kV or higher. Radial transmission
facilities serving only load with one transmission source are generally not included in this definition.
Capacity – The amount of electric power (in megawatts) that can be delivered to both: 1) firm energy to load
located electrically within the PJM Interconnection and 2) firm energy to the border of the PJM Control
Area for receipt by others. Installed capacity (ICAP) and unforced capacity (UCAP) are related measures
of this quantity.
Capacity Benefit Margin (CBM) – the CBM (in megawatts) represents the amount of import capability that is
reserved for the emergency import of power to help meet LSE load demands during peak conditions and
is excluded from all other firm uses. This is defined on page 104 of the Reliability Assurance Agreement
(RAA) , posted October 14, 2010.
Capacity Emergency Transfer Limit (CETL) – CETL represents the sub-area’s actual import capability as
determined from power flow studies. The sub-area criterion is satisfied if its CETL is equal to or exceeds
its CETO (see below). PJM’s CETO/CETL analysis is typically part of the PJM’s deliverability
demonstration; see PJM Manual 20, and PJM Manual 14B, attachment C for details.
Capacity Emergency Transfer Objective (CETO) – CETO represents the sub-area’s import capability objective
required to satisfy the RFC’s resource adequacy requirement LOLE. CETO is an independent evaluation
that is done in a coordinated and consistent manner with the annual RRS.
Central Limit Theorem (CLT) – In probability theory, the CLT states conditions under which the mean of a
sufficiently large number of independent random variables, each with finite mean and variance, will be
approximately normally distributed. In more general probability theory, a CLT is any of a set of weakconvergence theories. The formal mathematical theorem states:
1) If you have a simple random sample of “N” observations from a population with mean µ and
a standard deviation σ. 2) And if “N” is large, then 3) the sample average is approximately
normally distributed with a mean µ and standard deviation σ / N ½.
Two important practical implications of the Central Limit Theorem are: a) even if the sample values are not
normally distributed, the sample average is approximately normally distributed, and b) because the
sample average is normally distributed, you can use the Empirical Rule to summarize the distribution of
sample averages.46
Complementary – Something that perfects, completes, or adds to something47. Both PRISM and MARS add its
unique features and capabilities to each other to enhance (add to) a given individual application of each
tool.
46
SAS System for Elementary Statistical Analysis, Second Edition – 1997 – by Sandra D. Schlotzhauer and Ramon C. Littell,
PHD, published by the SAS Institute Inc, Cary NC, USA, pages 170-171.
47
Webster’s Dictionary and Thesaurus, Deluxe Edition, Nichols Publishing Group, Copyright 2001.
© PJM Interconnection 2011. All rights reserved
Page 85 of 136
Control Area (CA) – Electric power system or combination of electric power systems bounded by
interconnection metering and telemetry. A common generation control scheme is applied in order to:
o
o
o
o
o
Match the power output of the generators within the electric power system(s) plus the energy
purchased from entities outside the electric power system(s), with the load within the electric
power system(s);
Maintain scheduled interchange with other Control Areas, within the limits of Good Utility
Practice;
Maintain the frequency of the electric power system(s) within reasonable limits in accordance
with Good Utility Practice and the criteria of the applicable regional reliability council of NERC;
Maintain power flows on Transmission Facilities within appropriate limits to preserve reliability;
and
Provide sufficient generating Capacity to maintain Operating Reserves in accordance with
Good Utility Practice.
Convolution – the term “convolution” is used to depict the relationship between two input distributions of load
and capacity and the resulting loss of load distribution. Engineering problems approach this term from
considering the system as a set of weighting coefficients. The convolution method is called the recursive
method. The load model curve is thereby “convolved” (joined) with the generation system model for
computing the LOLE.
CP-8 Working Group – An NPCC planning group that is focused on LOLE studies for NPCC regional
assessments, review and compliance to NPCC planning protocols and Standards (such as demonstrating
an LOLE of 0.1 days/Year) . This group consists of representatives from the five control areas within
NPCC with some participation from neighboring regions’ staff. Participation requires a Non-Disclosure
Agreement as confidential information is required in the group efforts.
CURTAIL – An analysis tool based on legacy techniques that evaluate the load management’s reliability benefit
in terms of LOLE. The number of load management interruptions and the amount of load management are
key inputs in determining the benefit for each load management interruption. Typical assessment using
this tool focus on the reliability value’s saturation (measured in terms of LOLE) of load management and
the maximum amount of load management that cannot degrade the reliability value when 10 interruptions
are invoked.
Cut-Set – A line or boundary that defines a geographical demarcation to distinguish, typically, a Locational
Deliverability Area (LDA). All electrical facilities within a cut-set boundary are modeled within this zone.
Delivery Year (DY) – The Delivery Year (DY) is the twelve-month period beginning on June 1 and extending
through May 31 of the following year. As changing conditions may warrant, the PJM PC may recommend
other Delivery Year periods to the PJM Board of Managers. In prior studies, the DY was formerly referred
to as the “Planning Period”.
Deliverability – Deliverability is a test of the physical capability of the transmission network for transfer capability
to deliver generation capacity from generation facilities to wherever it is needed to ensure, only, that the
transmission system is adequate for delivery of energy to load under prescribed conditions. The testing
procedure includes two components: (1) Generation Deliverability; and (2) Load Deliverability. (See PJM
Manual 14B)
Demand Resource (DR) – A resource with the capability to provide a reduction in demand. DRs are a
component of PJM’s Load Management (LM) program. The DR is bid into the RPM Base Residual
Auction (BRA). See Load Management (LM) and ILR.
Demand Resource (DR) Factor – Associated with the Forecast Pool Requirement (FPR) factor and specified in
the Reliability Assurance Agreement (RAA). This factor is annually approved by the PJM Board of
Managers. The factor is multiplied by the amount of demand resources to yield the resulting value used in
PJM Markets.
Demand – The rate at which electrical energy is delivered to or by a system or part of a system, generally
expressed in kilowatts or megawatts, at a given instant or averaged over any designated interval of time.
Demand is equal to load when integrated over a given period of time. See Load.
Deterministic System – A deterministic system has a single result or set of results for a given set of input
parameters. A deterministic algorithm is an algorithm which behaves predictably. Deterministic
mathematical functions always produce the same output given a certain input. (Compare to Probabilistic
System.)
© PJM Interconnection 2011. All rights reserved
Page 86 of 136
Diversity – Diversity is the difference of the sum of the individual maximum demands of the various subdivisions
of a system, or part of a system, to the total connected load on the system, or part of the system, under
consideration. The two regions modeled in the RRS are the PJM RTO and the surrounding “World”
region. If the model has peak demand periods occurring at the same time, for both regions (PJM RTO
and World), there is little or no diversity. The peak demand period values are determined as the Expected
Weekly Maximum (EWM). A measure of diversity can be the amount of MWs that account for the
difference between a Transmission Owner zone’s forecasted peak load at the time of its own peak and the
coincident peak load of PJM at the time of PJM peak.
Effective Equivalent Demand Forced Outage Rate (EEFORd) – is used for reliability and reserve margin
calculations. For each generating unit, this outage rate is the sum of the EFORd plus ¼ of the equivalent
maintenance outage factor. See PJM Manual 22.
Electricity Reliability Organization (ERO) – The generic name used in the U.S. Energy Policy Act of 2005 to
refer to the independent entity that would be given the authority to develop and enforce mandatory
reliability standards for the North American bulk power system. NERC was designated as this “electricity
reliability organization” by FERC on July 20, 2006. “ERO” refers to NERC’s role, but “ERO” is not an
official name.
Empirical Rule – If data is from a normal distribution, the Empirical rule gives a quick and easy way to
summarize the data. This rule states: 1) 68% of the values are within one standard deviation of the mean.
2) 95% of the values are within two standard deviations of the mean 3) more than 99% of the values are
within three standard deviations of the mean. For a normal distribution about 68% of the values occur
48
between µ-σ and µ+σ, corresponding to the Empirical Rule.
Equivalent Demand Forced Outage Rate (EFORd) – the portion of time that a generating unit is in demand,
but is unavailable due to a forced outage. See PJM Manual 22.
Equivalent Maintenance Outage Factor (EMOF) – for each generating unit modeled, the portion of time a unit
is unavailable due to maintenance outages. See PJM Manual 22.
Expected Unserved Energy (EUE) – is the expected number of MWh of load that will not be served in a given
year. EUE indicates the capability of the Transmission system to continuously serve all loads at all
delivery points while satisfying all planning criteria. Information required for computing EUE includes: a)
frequency of each contingency (outage/year), b) duration of each contingency (hr/outage) and c) unserved
MW load for each contingency.
Expected Weekly Maximum (EWM) – the weekly peak load corresponding to the 50/50 load forecast, typically
based on a sample of 5 weekday peaks. The EWM parameter is used in the PJM PRISM program. Also
see PJM Manual 20 pages 19-23.
Forecast Error Factor (FEF) – FEF is a value that can be entered in the PRISM program per Delivery Year to
indicate the percent increase of uncertainty within the forecasted peak loads. As the planning horizon is
lengthened, the FEF generally increases 0.5% per year.
Forced Outage Rate (FOR) – FOR is a statistical measurement as a percentage of unavailability for generating
units and recorded in the GADS. FOR indicates the likelihood a unit is unavailable due to forced outage
events over the total time considered. It is important to note that there is no attempt to separate out
forced outage events when there is no demand for the unit to operate.
Forecast Peak Load – the expected peak demand (Load) representing an hourly integrated total in megawatts,
measured over a given time interval (typically a day, month, season, or delivery year). This expected
demand is a median demand value indicating there is a 50% probability actual demand will be above or
below the expected peak.
Forecast Pool Requirement (FPR) – the FPR represents the total UCAP requirement (percent) for the PJM
Control Area — determined by taking 100% plus the percentage of IRM for the PJM Control Area. The
percentage FPR is required pursuant to the Reliability Assurance Agreement (RAA) and approved by the
Reliability Committee pursuant to Schedule 4 of the RAA.
GEBGE – GEBGE is a resource adequacy calculation program, used to calculate daily LOLE that was jointly
developed in the 1960s/1970s by staff at General Electric International, Inc. (GEII) and Baltimore Gas and
Electric (BGE). The GEBGE program has since been largely superseded and replaced by PJM’s PRISM
in the conduct and evaluation of IRM studies at PJM. (See PRISM.)
48
SAS System for Elementary Statistical Analysis, Second Edition – 1997 – by Sandra D. Schlotzhauer and Ramon C. Littell,
PHD, published by the SAS Institute Inc, Cary NC, USA, page 136
© PJM Interconnection 2011. All rights reserved
Page 87 of 136
General Electric International, Inc. (GEII) – GEII is a global infrastructure, finance and media company taking
on the world’s toughest challenges. From everyday light bulbs to fuel cell technology, to cleaner, more
efficient jet engines, GE has continually shaped our world with groundbreaking innovations for over 130
years. See the GEII web site for further details.
Generator Availability Data System (GADS) – As defined in the NERC Generating Availability Data System
(GADS) (http://www.nerc.com/page.php?cid=4|43 ), PJM fully supports GADS using the eGADS web
based application. All generating facilities taking part in the PJM markets are required to submit unit
statistical performance and reliability data to determine the value of the facility as an unforced capacity
(UCAP) resource, by using PJM’s eGADS application found as one of the eSUITE applications available
on the PJM web site (https://esuite.pjm.com/mui/ ).
Generator Forced / Unplanned Outage – An immediate reduction in output, capacity, or complete removal from
service of a generating unit by reason of an emergency or threatened emergency, unanticipated failure, or
other cause beyond the control of the owner or operator of the facility. A reduction in output or removal
from service of a generating unit in response to changes in or to affect market conditions does not
constitute a Generator Forced Outage.
Generator Maintenance Outage – The scheduled removal from service, in whole or in part, of a generating unit
in order to perform necessary repairs on specific components of the facility approved by the PJM Office of
Interconnection .
Generator Planned Outage – A generator planned outage is the scheduled removal from service, in whole or in
part, of a generating unit for inspection, maintenance or repair – with the approval of the PJM.
Graphical User Interface (GUI) – a GUI is a Java-based web environment that allows the user to point and click
selections to create a desire model, assessment options, and analysis parameters. Supported and
maintained by the PJM Information Technology Services (ITS) staff.
Institute of Electrical and Electronic Engineers Inc. (lEEE) – the IEEE is a non-profit professional association
dedicated to advancing technological innovation related to electricity. IEEE is chartered to advance the
theory and practice of Electrical, Electronics, Communications and Computer engineering, as well as
computer science, the allied branches of engineering and the related arts and sciences. IEEE serves as a
major publisher of scientific journals and a conference organizer. It is also a leading developer of industrial
standards (having developed more than 900 active industry standards) throughout a broad range of
disciplines. IEEE has more than 395,000 members in more than 160 countries, 45% outside the United
States.
Installed Capacity (ICAP) – commonly refers to “iron in the ground” – or rated capacity of a generation unit (in
megawatts) prior to derating or other performance adjustments.
In02 file – One of the 3 necessary MARS input files that have the 8760 hourly loads used in the calendar year
assessment. This is a discrete distribution of the load model shape used in the Monte Carlo simulations.
A separate set of 8760 values is needed for each area in the model. This is a text file, using the historic
EEI format for loads (80 columns, 12 hours (am) on 1st line, 12 hours (pm) on 2nd line.
In05 file – One of the 3 necessary MARS input files typically referred to as the Master Input File (MIF). This is a
text file that has all the MARS tables specified for the assessment.
ln17 file – One of the 3 necessary MARS input files that can have hourly shape information, such as for wind
units and Demand Side Management(DSM) profiles. This file is a text file but can be blank if no hourly
shapes are part of analysis.
Import Capability –Import Capability (in megawatts) is a single value that represents the simultaneous imports
into PJM that can occur during peak PJM system conditions. The capabilities of all transmission facilities
that interconnect the PJM Control Area to its neighboring regions are evaluated to determine this single
value.
Installed Reserve Margin (IRM) – IRM represents the amount of aggregate generating unit capability above the
forecasted peak load that is required to meet a given resource adequacy level. IRM is expressed as a
percentage, and is calculated by dividing the total installed capacity (ICAP) by the net peak load (after
accounting for qualified forecast Load Management). The IRM is used to determine the Forecast Pool
Requirement (FPR). The PJM IRM is the level of installed reserves needed to meet the ReliabilityFirst
Corporation criteria for an LOLE of one day, on average, every 10 years
© PJM Interconnection 2011. All rights reserved
Page 88 of 136
Independent Electricity System Operator (IESO) – is the nonprofit Regional Transmission Organization (RTO)
responsible for managing Ontario's bulk electricity power system and operating the wholesale market. The
IESO is one of the five main control areas within the NPCC. (http://www.ieso.ca/ ).
Independent System Operator (ISO) – An Independent System Operator (ISO) is an organization formed at the
direction or recommendation of the Federal Energy Regulatory Commission (FERC). In the areas where
an ISO is established, it coordinates, controls, and monitors the operation of the electrical power system,
usually within a single state or province.
Independent System Operator of New England (ISO-NE) – is a regional transmission organization (RTO) and
not-for-profit corporation responsible for reliably operating the bulk electric power generation, transmission
system and wholesale electricity markets in New England. Created in 1997 and with headquarters in
Holyoke, MA, the ISO-NE control extends throughout New England including Maine, New Hampshire,
Vermont, Rhode Island, Massachusetts and Connecticut. (www.iso-ne.com)
Locational Deliverability Areas (LDAs) – are zones that comprise the PJM RTO as defined in the RAA
schedule 10.1 and can be an individual zone, a combination of two or more zones, or a portion of a zone.
There are currently 25 LDAs within the PJM footprint.
Load - Integrated hourly electrical demand, measured as generation net of interchange. Loads generally can be
reported and verified to the tenth of a megawatt (0.1 MW) for this report.
Load Management (LM) - Load Management, previously referred to as Active Load Management (ALM),
applies to interruptible customers whose load can be interrupted at the request of PJM. Such a request is
considered an emergency action and is implemented prior to a voltage reduction. This includes both
Demand Resources (DR) and Interruptible Load for Reliability (ILR).
Load-Serving Entity (LSE) – an LSE is a utility company, distribution company or power marketer that provides
the distribution, customer, and energy services for natural gas and electricity. An LSE may also be
referred to as a Utility Distribution Company (UDC).
Loss of Load Expectation (LOLE) - Generation system Adequacy is determined as LOLE and is expressed as
days (occurrences) per year. This is a measure of how often, on average, the available capacity is
expected to fall short of the restricted demand. LOLE is a statistical measure of the frequency of firm
load loss and does not quantify the magnitude or duration of firm load loss. The use of LOLE to assess
Generation Adequacy is an internationally accepted practice. The term expectation is not in the ordinary
sense but adheres to a mathematical definition (See Reference 4).
Loss of Load Hours (LOLH) – represents the hours of unserved energy for a given delivery year.
Multi-Area Reliability Simulation (MARS) – developed by General Electric International, Inc., the MARS (or
GE-MARS) model is a probabilistic analysis program that uses sequential Monte Carlo simulation to
analyze the resource adequacy for multiple areas. MARS is used by ISOs, RTOs, and other
organizations to conduct multi-area reliability simulations.
Metadata – metadata is essentially a data specification – the purpose being to define the many relationships
among various and different database sources that are related yet distinct and separate. This term and
specification technique is used in several state-of-the-art database analysis tools to quickly assess and
report complexities and quantify relationships that otherwise were not practical due to resource limitations
and knowledge transfer among the many database schemas. The Microsoft product SSIS (SQL Server
Integration Services) and SSRS (SQL Server Reporting Services) are examples of available OLAP tools
within the Information Technology Services (ITS) industry. Metadata is used to join together the various
load parameters with the generation parameters, on a time and geographical basis.
Midwest Independent System Operator (MISO) – is a not-for-profit corporation responsible for reliably
operating the bulk electric power generation, transmission system and wholesale electricity markets in
the Midwest. Created in 1996 and with headquarters in both Carmel, IN and St. Paul MN, MISO’s control
extends throughout through 13 states and one Canadian province, including all or parts of: Illinois,
Indiana, Iowa, Michigan, Minnesota, Missouri, Montana, Nebraska, North Dakota, Ohio, Pennsylvania,
South Dakota, Wisconsin and Manitoba (Canada). (www.midwestiso.org)
Monte Carlo Simulation Methods – as mathematical methods for assessing risk probabilities, Monte Carlo
methods involve computational algorithms that rely on repeated random sampling to compute their results.
Monte Carlo methods are often used in simulating physical and mathematical systems where it is
unfeasible or impossible to compute an exact result with a deterministic algorithm. Monte Carlo methods
are especially useful in studying systems with a large number of variables, coupled degrees of freedom
and significant uncertainty in inputs, such as risk assessment.
© PJM Interconnection 2011. All rights reserved
Page 89 of 136
Developed at the time during development of the World War II-based Manhattan Project, Monte Carlo
simulations have been successfully applied in space exploration and oil exploration, actual observations of
failures, cost overruns and schedule overruns are routinely better predicted by the simulations than by
human intuition or alternative "soft" methods. The alternative to Monte Carlo methods is to solve the
problems using the mathematics of probability. This can be performed using numerical methods using
established techniques involving cumulative probability arrays.
New York Independent System Operator (NYISO) – is a not-for-profit corporation responsible for reliably
operating the bulk electric power generation transmission system and wholesale energy markets in New
York State. The NYISO was created from the former New York Power Pool (NYPP) in 1999 and has its
headquarters in Albany, NY. (www.nyiso.com)
New York State Reliability Council (NYSRC) – is a nonprofit, sub-regional electric reliability organization
(ERO) within the NPCC. Working in conjunction with the NYISO, the NYSRC’s mission is to promote and
preserve the reliability of electric service on the New York Control Area (NYCA) by developing,
maintaining and updating reliability rules which shall be complied with by the New York Independent
System Operator (NYISO). (www.nysrc.org)
North American Electric Reliability Corporation (NERC) – is a nonprofit corporation based in Princeton, NJ
(Plans to move to Atlanta GA). NERC was formed in 2006 as the successor to the North American
Electric Reliability Council (also known as NERC). NERC supports the electric utility industry in
promoting reliability and adequacy of the bulk power transmission in the electric utility systems of North
America. NERC's mission states that it is to "ensure that the bulk power system in North America is
reliable." NERC oversees eight regional reliability entities and encompasses all of the interconnected
power systems of the contiguous United States, Canada and a portion of Baja California in Mexico.
(www.nerc.com)
Northeast Power Coordinating Council (NPCC) – is a regional electric reliability organization within NERC that
is responsible for ensuring the adequacy, reliability, and security of the bulk electric supply systems of the
Northeast region comprising parts or all of: New York, Maine, Vermont, New Hampshire, Connecticut,
Rhode Island, Massachusetts, and the Canadian provinces of Ontario, Quebec, Nova Scotia, New
Brunswick, and Prince Edward Island. The five control areas that comprise the NPCC region include:
NYISO, ISO-NE, Ontario (IESO), Hydro Quebec, and the Maritimes. (http://www.npcc.org/ )
Online Analytical Processing (OLAP) – OLAP (/o-lap/), is an approach to quickly answer multi-dimensional
analytical queries. OLAP is part of the broader category of business intelligence, which also
encompasses relational reporting and data mining. The typical applications of OLAP are in business
reporting and business process management (BPM). OLAP was derived from the traditional database
term OLTP (Online Transaction Processing). Databases configured for OLAP use a multidimensional
data model, allowing for complex analytical and ad-hoc queries with a rapid execution time – borrowing
aspects of navigational databases and hierarchical databases that are faster than relational databases.
The output of an OLAP query is typically displayed in a matrix (or pivot) format. The dimensions form the
rows and columns of the matrix; the measures form the values.
Peak Load – The Peak Load is the maximum hourly load over a given time interval, typically a day, month,
season, or delivery year. (Refer to Forecast Peak Load.)
Peak Load Ordered Time Series (PLOTS) – The PLOTS load model is the result of the Week Peak Frequency
(WKPKFQ) application. This is one of the load model’s input parameters. This is discussed in the 2010
Reserve Requirement Study Report, load forecasting and WKPKFQ parameters section of Part II –
Modeling and analysis.
Peak Season – Peak Season is defined to be those weeks containing the 24th through 36th Wednesdays of the
calendar year. Each such week begins on a Monday and ends on the following Sunday, except for the
week containing the 36th Wednesday, which ends on the following Friday. Please note that the load
forecast report(s) used for studies define peak season as June, July and August.
Pipe Size – Determined by load flow based assessment techniques of individual transmission facilities to assess
a single value representative of the composite set of transmission facilities. The pipe size is a summation
of Individual Interface tie sizes to represent transfer limits between areas used by LOLE modeling
methods.
PJM Interconnection Inc. – PJM is a regional transmission organization (RTO) responsible for reliably
operating the bulk electric power generation, transmission system and wholesale energy markets
throughout all or parts 13 states and the District of Columbia, including: Delaware, Illinois, Indiana,
© PJM Interconnection 2011. All rights reserved
Page 90 of 136
Kentucky, Maryland, Michigan, New Jersey, North Carolina, Ohio, Pennsylvania, Tennessee, Virginia,
West Virginia and the District of Columbia.
Planned Outage Factor – A factor measured in weeks per year that indicates the forecast amount, for each unit
to capture the following outages as recorded in PJM’s eGADS application: A unit’s outage of
predetermined length that is scheduled well in advance of its occurrence. The outage must be included in
a regularly issued maintenance schedule at least one month prior to the starting date of the outage; i.e.,
the outage must appear in two consecutive issues of the PJM Unit Maintenance Schedule prior to starting
the outage. See PJM Manual 22 for further details.
Planning Committee – The PJM Planning Committee (PC) is established under the Operating Agreement
(PDF) and has the responsibility to review and recommend system planning strategies and
policies as well as planning and engineering designs for the PJM bulk power supply system to assure
the continued ability of the member companies to operate reliably and economically in a competitive
market environment. Additionally, the PC makes recommendations regarding generating capacity reserve
requirement and demand-side valuation factors. The RAAS reports to this Committee.
Probabilistic Reliability Index Study Model (PRISM) – PRISM is PJM’s primary planning reliability program.
Developed by PJM Staff, PRISM replaced GEBGE which was a FORTRAN language program. The
models are based on statistical measures for both the load model and the generating unit model. This is
a computer application developed by PJM that is a practical application of probability theory and is used
in the planning process to evaluate the generation adequacy of the bulk electric power system.
Probabilistic System – A probabilistic system will have results that vary, due to observable certainty described
by its system distribution parameters. Probabilistic systems (also called stochastic model, process, or
system) are often solved with Monte-Carlo methods – where a computer program uses a pseudo random
number generator to provide values of the attributes in the system that can vary. The program provides an
assessment of the uncertainty of results. Typically, a large number of runs (trials or iterations) are made.
Summary statistics may include the value that occur most frequently (mode), the mean value, and low and
high range, for instance the 10% and 90% percentile. The standard deviation and histogram of results
may also be part of the summary information. There is no single standard presentation as this will depend
on the application. (Compare to Deterministic System.)
Probability Density Function (PDF) – is P (x) of a continuous distribution is defined as the derivative of the
(cumulative) distribution function D (x). The probability density function (PDF) of a continuous distribution
is defined as the derivative of the (cumulative) distribution function, with a full mathematical definition of
the PDF shown at the following link (Wolfram MathWorld – the web’s most extensive mathematics
resource): http://mathworld.wolfram.com/ProbabilityDensityFunction.html.
Probable Load – A given amount of load (typically in megawatts) and the associated probability for that load
value is the probable load. Probable Load is the relationship and association between a given load value
and the likelihood that this load amount will occur.
Resource Adequacy Planning (RAP) – the PJM RAP Department is responsible for the conduct of the annual
Forecast Pool Requirement (FPR) development process and associated Unforced Capacity (UCAP)
Obligation process. The development of the FPR begins with the Reserve Requirement Development
Process utilizing Load Forecasts as well as other resource reports. Encompassing both Planning and
Actual data, RAP considers the impact of load, generation, demand resources and transmission.
Reliability – in a bulk power electric system, is the degree to which the performance of the elements of that
system results in power being delivered to consumers within accepted standards and in the amount
desired. The degree of reliability may be measured by the frequency, duration, and magnitude of adverse
effects on consumer service. Bulk Power electric reliability cab be addressed be considering two basic
and functional aspects of the bulk power system – adequacy and security.
ReliabilityFirst Corporation (RFC) – is one of nine regional electric reliability councils under NERC authority.
RFC began operations in 2006 as the successor to three other reliability organizations: The Mid-Atlantic
Area Council (MAAC), the East Central Area Coordination Agreement (ECAR) and the Mid-American
Interconnected Network (MAIN). Based in Akron, OH, RFC lies within the Eastern Interconnection and
covers territory stretching from the Eastern United States to the lower Great Lakes, covering all of the
states of Pennsylvania, New Jersey, Delaware, Maryland, West Virginia, Ohio, Indiana, and portions of
the states of Wisconsin, Michigan, Illinois, Kentucky and Virginia. (www.rfirst.org)
Reliability Index (RI) – is a value that is used to assess the bulk electric power system’s future occurrence for a
loss-of-load event. A RI value of 10 indicates that there will be, on average, a loss of load event every ten
years. A given value of reliability index is the reciprocal of the LOLE.
© PJM Interconnection 2011. All rights reserved
Page 91 of 136
Reliability Pricing Model (RPM) – RPM is PJM’s capacity-market model. Implemented in 2007, RPM is based
on making capacity commitments three years ahead, is designed to create long-term price signals to
attract needed investments in reliability in the PJM region. The long-term RPM approach, in contrast to
PJM’s previous short-term capacity market, includes incentives that are designed to stimulate investment
both in maintaining existing generation and in encouraging the development of new sources of capacity –
resources that include not just generating plants, but demand response and transmission facilities.
The RPM model works in conjunction with PJM’s Regional Transmission Expansion Planning (RTEP)
process to ensure the reliability of the PJM region for future years. The RPM includes the continued use
of self-supply and bilateral contracts by load-serving entities (LSEs) to meet their capacity obligations. The
capacity auctions under the RPM obtain the remaining capacity that is needed after market participants
have committed the resources they will supply themselves or provide through contracts.
Regional Transmission Expansion Plan (RTEP) – PJM’s RTEP identifies transmission system additions and
improvements needed throughout the PJM footprint. Studies are conducted to test the transmission
system against mandatory NERC national standards as well as PJM regional standards. RTEP looks
ahead 15 years into the future to identify transmission overloads, voltage limitations and other reliability
standards violations.
RTEP results are used to assist PJM’ Planning Dept. in developing transmission plans to resolve
violations that could otherwise lead to overloads and black-outs. These plans are examined for their
feasibility, impact and costs and are discussed throughout the development process with PJM
stakeholders. When reliability criteria violations are detected, PJM then develops solutions to mitigate
those violations. RTEP recommendations are submitted periodically to PJM’s independent Board of
Managers to resolve identified reliability criteria violations. Once mitigation plans are approved, they
become part of PJM’s overall RTEP.
Sub-Regional RTEP Committees (SR RTEPs) have been formed for the Mid-Atlantic, Southern and
Western sub-areas to provide review and input of sub-regional RTEP projects and provide
recommendations to the TEAC.
Regional Transmission Organization (RTO) – Each entity (a) that owns, leases or otherwise has a possessory
interest in facilities used for the transmission of electric energy in interstate commerce, (b) that provides
Transmission that is a party to the PJM Transmission Owners Agreement and PJM Operating Agreement.
Reserve Requirement Study (RRS) – Performed annually, the objective of PJM’s RRS is to derive a single
calculated percentage of IRM that represents the amount above peak load that must be maintained to
meet the RFC adequacy criteria. The RFC adequacy criteria are based on a probabilistic requirement of
experiencing a loss-of-load event, on average, once every ten years. (The RRS is also referred to as the
R-Study.)
Reserve Requirement Assumptions Working Group (RRAWG) – this former group previously reported to the
PJM PC. Beginning in January 2011 this RRAWG group was renamed the Resource Assumptions
Analysis Subcommittee (RAAS).
Resource Adequacy Analysis Subcommittee (RAAS) - reporting to the PC, the RAAS assists PJM staff in
performing the annual Reserve Requirement Study (RRS), setting LOLE study assumptions, performing
CETO assessments and maintains the reliability analysis documentation. This subcommittee name
replaced the Reserve Requirement assumptions Subcommittee (formed in January 2011 replacing the
RRAWG) name.
Restricted Peak Load - For the given forecast period, the restricted peak load equals the unrestricted peak load
minus anticipated load management.
SAS – SAS Institute offers several software products and analytical solutions driven by a proprietary 4th
generation computer language. It is owned by SAS Institute Inc. , providing software solutions since
1976. SAS Inc. is the largest independent vendor in the business intelligence market, based in Cary,
North Carolina.
Security – refers to the ability of the bulk electric system to withstand sudden disturbances such as electric short
circuits or unanticipated loss of system components or switching operations. One part of the Reliability
term.
© PJM Interconnection 2011. All rights reserved
Page 92 of 136
SERC Reliability Corporation (SERC) - SERC is one of nine regional electric reliability councils under North
American Electric Reliability Corporation (NERC) authority. SERC was formed in 2005, as the successor
to the Southeast Electric Reliability Council (also known as SERC) and is based in Charlotte, NC. The
SERC region serves all or parts of Alabama, Arkansas, Florida, Georgia, Illinois, Iowa, Kentucky,
Louisiana, Missouri, North Carolina, Oklahoma, South Carolina, Tennessee, Texas and Virginia.
(www.serc1.org )
Simultaneous Import Limit (SIL) – The SIL is the amount (megawatts) that can be imported into the PJM RTO
over all transmission tie-line facilities during capacity emergency conditions. The SIL determined for FERC
compliance filings, such as Order No. 697, can use different assumptions than used for forecasting future
planning requirements, and has specific, detailed study assumptions specifications. However the SIL is
one of several considerations used in review of the Capacity Benefit Margin (CBM).
Summer Net Dependable (SND) – is the rating for a given generation unit is used in the summer period. All
processes use the SND rating as the basis for evaluating a unit. See PJM Manual 21 for further details.
Transmission Expansion Advisory Committee (TEAC) – the TEAC provides advice and recommendations to
aid in the development of the RTEP and SR RTEPs.
Unforced Capacity (UCAP) – Installed capacity rated at summer conditions that are not on average
experiencing a forced outage or forced derating, calculated for each Capacity Resource on the 12-month
period from October to September without regard to the ownership of or the contractual rights to the
capacity of the unit.
Unrestricted Peak Load – this represents the metered load plus estimated impacts of Load Management.
Variance - A measure of the variability of a unit's partial forced outages which is used in reserve margin
calculations. This term is also called the two-state variance, a result of the PRISM model only having two
states either full on or full off. See PJM manual 22 for further details.
Weak Peak Frequency (WKPKFQ) – A computer application that creates the Peak Load Ordered Time Series
(PLOTS) load model; its input data is hourly load data. This application is part of the ARC application.
Weather Normalized Loads - The weather-normalized loads are estimated seasonal peak assuming median
peak day weather conditions. The weather-normalized loads are also referred to as 50 / 50 loads.
Winter Weekly Reserve Target – part of the annual RRS process. PJM Staff develops the Winter Weekly
Reserve Target, covering the 13-week winter season (December through February.). This parameter is
used for operation’s planned outage scheduling of the PJM generation fleet. PJM uses the MARS
program to evaluate this parameter.
Zone / Control Zone - An area within the PJM Control Area, as set forth in PJM’s Open Access Transmission
Tariff (OATT) and the Reliability Assurance Agreement (RAA). Schedule 10 and 15 of the RAA provide
information concerning the distinct zones that comprise the PJM Control Area.
© PJM Interconnection 2011. All rights reserved
Page 93 of 136
References
1.
Multi-Area Generation Reliability – Phase I Testing Report - October 22, 1991 - developed by the New
York Power Pool (NYPP) Reliability Evaluation Task Force for the ESEERCO Project EP87-28
2.
Generating Reserve Capacity Determined by the Probability Method - March 25, 1947 - by Giuseppe
Calabrese, presented at the AIEE Midwest meeting of November 3, 1947
3.
Probability Methods Applied to Generating Capacity Problems of a Combined Hydro and Steam
System - August 14, 1947 - by E.S. Loane and C.W. Watchorn; presented at the AIEE Midwest meeting of
November 3, 1947
4.
Reliability Evaluation of Power Systems – 1984 – by Roy Billinton and Ronald Allan, published by the
Plenum Press, New York
5.
NPCC Regional Reliability Reference Directory #1 “Design and Operation of the Bulk Power System”
(December 1, 2009) - section 5.2, page 9.
6.
MAPP Loss of Load Expectation (LOLE) Study, MAPPCOR and MAPP Composite System Reliability
Working Group (December 30, 2009);
http://www.mapp.org/ReturnBinary.aspx?Params=584e5b5f405c567900000002cb
7.
Mid-Atlantic Area Council (MAAC) Document A-1, MAAC Reliability Principles and Standards (March 30,
1990).
8.
MAIN Guide #6, Generation Reliability Study 20045-2014 (September 27, 2005).
9.
Standard BAL-502-RFC-02 (December 4, 2008) - ReliabilityFirst Corporation (RFC) region planning
resource Adequacy Analysis, Assessment and Documentation,: http://www.nerc.com/files/BAL-502-RFC02.pdf .
10. Midwest Reliability Organization (MRO) Standard RES-501-MRO-01, planned Resource Adequacy
Assessment (December 29, 2007);
http://www.midwestreliability.org/04_standards/approved_standards/mro_standards/RES-501MRO-01_Final_20071229_Clean.pdf .
11. PJM Generation Adequacy Analysis: Technical Methods - October 2003 – developed by PJM Staff;
available at: http://www.pjm.com/planning/resource-adequacy-planning/~/media/planning/resadeq/20040621-white-paper-sections12.ashx.
12. Probability Calculation of Generation Reserves - March 1969 - by C.J. Baldwin and published by The
Westinghouse Engineer, This paper is copyright protected but can be purchased online at Infotrieve; article
information accession number 00434361 (800-422-4633; www.infotrieve.com).
13. Evaluation of PTI’s Multi-Area Reliability Program MAREL - November 1993 - developed by the PJM
Load & Capacity Working Group (L&CWG)
14. Power System Reliability Evaluation, Mathematical Expectation - page 12, Gordon and Beach, Science
Publishers, –1970 – by Roy Billinton and Reinvent Legacy Software with SAS, the Web, and OLAP Reporting, Appendix
B –2007 – available at: http://www.qlx.com/whitepapers/25-2008.pdf .
15. Re-invent a Legacy System with SAS, the Web and OLAP reporting, November 13, 2007, 2007
Northeast SAS user group conference – Applications Big and Small, Paper Number 10.
http://www.lexjansen.com/nesug/ , Appendix : Download the code (2369 KB)
16. Schedule 15 of Reliability Assurance Agreement (RAA) (link) and PJM Manual 14B, pages 49-50 (Link),
December 02, 2010.
17. Power System Reliability Evaluation, Mathematical Expectation - page 12, Gordon and Beach, Science
Publishers, –1970 – by Roy Billinton and Reinvent Legacy Software with SAS, the Web, and OLAP
Reporting, Appendix B –2007 – available at : http://www.qlx.com/whitepapers/25-2008.pdf .
18. The Scientist and Engineer's Guide to Digital Signal Processing , By Steven W. Smith, Ph.D, copyright
© 1997-2006 by California Technical Publishing – http://www.dspguide.com/pdfbook.htm
© PJM Interconnection 2011. All rights reserved
Page 94 of 136
19. Course notes, module PE.PASU19.5 on Generation adequacy evaluation, Convolution techniques,
item U19.7.3 - page 43, by Dr. James McCalley’s. http://www.ee.iastate.edu/~jdm/ee653/ee653schedule.htm
20. Course notes: Electrical Power System Reliability, part3 Discrete Convolution Method - page 30,
copyright 1995, by Dr. Chanan Singh, http://www.ece.tamu.edu/People/bios/singh/coursenotes/part3.pdf
21. Northeast Power Coordinating Council Tie Benefits Methodology, by Glenn Haringa and Philip Fedora,
November 5-6, 2008, Best LOLE Practices meeting held at California ISO offices, Agenda item 8. See slide
12, last bullet.
22. New York Control Area Installed Capacity Requirements for the Period May 2010 through April 2011 Technical Study Report - December 4, 2009 – developed by the New York State Reliability Council, LLC
(NYSRC) Installed Capacity Subcommittee (ICS)
23. New York State Reliability Council LLC NYSRC Policy, No 5-3, procedure for establishing New York
Area Installed Capacity Requirements, November 16, 2009)
24. Westinghouse Capacity Model and GE MARS, by Fei Zeng ISO New England, April 27-28, 2010, North
America Electric Reliability Corporation (NERC) Loss-Of-Load-Expectation (LOLE) Working Group, Portland
OR. Agenda item 7d – http://www.nerc.com/filez/lolewg.html .
25. California ISO Planning Reserve Margin – 2010 – 2020, May 21, 2010 Final Report to California
Independent System Operator for Planning Reserve Margin (PRM) Study - 2010 – 2020.
http://www.caiso.com/279d/279ded0337f20.pdf
26. IEEE Standard 762-2006: IEEE Standard definition for use in reporting electric generating unit
reliability, availability and productivity. IEEE Std-2006 (revision of IEEE Std 762-1987) - approved
9/16/2006 by IEEE-SA Standards Board. To obtain a copy: http://standards.ieee.org/findstds/standard/7622006.html )
27. Northeast Power Coordinating Council Interregional Long Range Adequacy Overview - (November 28,
2006), conducted by the NPCC CP-8 Working Group. http://www.npcc.org/docSearch.aspx .
28. SAS System for Elementary Statistical Analysis, Second Edition – 1997 – by Sandra D. Schlotzhauer
and Ramon C. Littell, PHD, published by the SAS Institute Inc, Cary NC, USA
29. 2009 PJM Reserve Requirement Study – 11-year Planning Horizon: June 1st 2009 to May 31st 2020 –
October 13, 2009 - Analysis performed by PJM Staff with review by the PJM Reserve Requirement
Assumptions Working Group (RRAWG)
30. NYISO load model processing for selection of LOLE load models –New York State Reliability Council
(NYSRC), Installed Capacity Subcommittee (ICS) Meeting, Albany, NY, June 1, 2010. Footnoted: Draft –
For discussion only.
31. Reliability Assessment of Electrical Power Systems using Monte Carlo Methods – 1994 – by Roy
Billinton and Wenyuan Li, published by the Plenum Press, New York
32. PJM Manual 22, Generator Resource Performance indices, revision 15, June 1, 2007 - :
http://www.pjm.com/~/media/documents/manuals/m22.ashx
33. PJM Manual 21, Rules and Procedures for Determination of Generating, revision 9, May 1, 2010 :
http://www.pjm.com/~/media/documents/manuals/m21.ashx
34. PJM Manual 20, PJM Resource Adequacy Analysis, revision 3, June 1, 2007 :
http://www.pjm.com/~/media/documents/manuals/m20.ashx
35. A Monte Carlo Simulation Model for Multi-Area Generation Reliability Evaluation, by A. A. Chowdhury,
L. Bertling, B. P. Glover, G.E. Haringa, June 11-15, 2006, International conference on Probabilistic Methods
Applied to Power Systems, Stockholm Sweden.
36. Reliability Evaluation of Engineering Systems: Concepts and Techniques – 1984 – by Roy Billinton
and Ronald Allan, published by the Plenum Press, New York
37. Reliability Pricing Model (RPM) Auction User Information, 2013/2014 Scenario Analysis Results.
Simulation # 4 of 22. These simulations are available on the PJM Web Site.
© PJM Interconnection 2011. All rights reserved
Page 95 of 136
38. Measurement Practices for Reliability and Power Quality, a toolkit of reliability measurement
practices, June 2004, by J.D. Kueck and B.J. Kirby, Oak Ridge National Laboratory, P.N. Overholt, US DOE, L.C.
Markel, Sentech, Inc. Report link. DOE Research report abstract.
39. PJM Load forecast report, January 2010. Available on PJM web site at this location.
40. 2010 PJM Reserve Requirement Study – 11-year Planning Horizon: June 1st 2010 to May 31st 2021 –
September 30, 2010 - Analysis performed by PJM Staff with review by the PJM Reserve Requirement
Assumptions Working Group (RRAWG). Report posted here.
41. Federal Energy Regulatory Commission, 18 CFR part 40, Docket No. RM10-10-000 – Planning
Resource Adequacy Assessment Reliability Standard, issued October 21, 2010. Notice of Proposed
Rulemaking (NOPR).
42. Reliability Assurance Agreement among load serving entities in the PJM Region , Rate Schedule FERC
No. 44. Agreement posted here.
© PJM Interconnection 2011. All rights reserved
Page 96 of 136
APPENDICES
o
Appendix A – “MARS Model vs. PRISM Model Overview” - PJM Staff presentation for the
April 15, 2008 RRAWG meeting
o
Appendix B - Calculations to translate from PLOTS load models used in PRISM to the LODUNCY table values used in MARS
o
Appendix C -- Load Modeling Comparison Issues

Appendix C1 – Load Distribution

Appendix C2 – Baldwin Paper: Choice of Sigma

Appendix C3 – ISO-NE’s Comparison of Westinghouse Capacity Model and GEMARS
o
Appendix D – MARS Solution Techniques
o
Appendix E – Transparency of process and steps used to perform LOLE calculations
o
Appendix F – Detailed Input Parameter Requirements
o
Appendix G – Possible future investigation activities
© PJM Interconnection 2011. All rights reserved
Page 97 of 136
Appendix A – MARS Model vs. PRISM Model Overview PJM Staff presentation
“MARS Model vs. PRISM Model Overview”
PJM Staff presentation for the April 15, 2008 RRAWG meeting
The following PJM Staff presentation was delivered at the 15 April 2008 RRAWG meeting, per requests to show how MARS and PRISM are being used and how they compare.
This presentation was developed to capture the efforts of several PJM staff, but was not meant to provide full and comprehensive documentation for how to coordinate
consistent modeling between the different tools. The PJM staff recognizes that thorough documentation needs development; however this was identified as a lower priority task.
The important points of this presentation information include: 1) PJM staff have been using MARS and PRISM in concert since 2004; 2) PRISM and MARS are parts of the
overall production environment Applications for Reliability Calculations (ARC) ; 3) While MARS is driven using the ARC GUI, MARS has many more manual text entries ; 4) The
key component for consistent models is the load model entries in four MARS tables. ; 5) There are more MARS categories of input requirements and more data than PRISM. ;
6) MARS provides capabilities that PRISM does not and they are used in a complementary fashion. ; 7) Comprehensive documentation will require additional efforts.
© PJM Interconnection 2011. All rights reserved
Page 98 of 136
© PJM Interconnection 2011. All rights reserved
Page 99 of 136
© PJM Interconnection 2011. All rights reserved
Page 100 of 136
© PJM Interconnection 2011. All rights reserved
Page 101 of 136
© PJM Interconnection 2011. All rights reserved
Page 102 of 136
Appendix B – Translation from PLOTS to LOD-UNCY table
Calculations to translate from PLOTS load models used in
PRISM to the LOD-UNCY table values used in MARS
The Appendix B spreadsheet maps the steps in the calculation process to translate the PRISM load model to the
MARS load model so that modeling can be consistent (See Appendix C3, slides 12-15).
© PJM Interconnection 2011. All rights reserved
Page 103 of 136
The values are processed from left to right – and then from top to bottom. The teal-highlighted values represent a
distribution for the mean while the light blue-highlighted values represent the final EWM distribution values.
1.
Run Week Peak Frequency (WKPKFQ) to create the Peak Load Ordered Time Series (PLOTS) Load
Model
This is a probabilistic load model typically based on 7-10 years of history and the current load
forecast growth rates. This creates the mean and standard deviation for each week in the model (152). This is shown in the Mean seasonal Trend and the Standard Deviation columns (yellow
highlighted area). These are the starting values for the translation process.
2.
Translate the ARC weeks into calendar months.
The ARC weeks correspond to a delivery year time period. A delivery year starts around the 1st of
June and ends the following May. The values are gathered into the appropriate calendar month, as
shown in the in the columns before the dark bold line. I.e. ARC week 1 is calendar week 20, which is
in mid-May.
3.
Calculate Mean Values (teal-highlighted area)
The calculated values of variance, which is the square of the standard deviation, and the variance
times the mean help to determine a monthly weighted variance and standard deviation(standard
deviation is the squared root of the variance). These weighted Standard deviations for each month
are used to determine the 7 LOD-UNCY values at the given number of STD away from the mean,
assuming a Standard normal distribution. These values are shown in the teal-highlighted region.
4.
Final EOP-UNCY Table values using EWM
The monthly weighted Expected Weighted Maximum (EWM) is determined from the monthly
weighted standard deviation values. [ EWM = 1* (1+1.16295* Wtd. STD Dev) ] .
o
o
o
o
o
The 7 EOP-UNCY values for each month are the mean values in the teal-highlighted area
divided by the monthly EWM values (calculated from Wtd. STD Dev values shown on that
row for that month).
These EOP-UNCY table values now take on the distribution using the EWM methods
deployed in the PLOTS and PRISM load model.
A recognized limitation is that one can only subset the LOD-UNCY into monthly aggregates.
The ability to specify weekly or daily LOD-UNCY values is not available yet. A request to GE
staff has been submitted to be considered for future MARS improvements.
The final LOD-UNCY values are shown in the gray-highlighted area.
It is worth noting that the ISO-NE staff presented recent material that indicates a similar process is involved to join
their Westinghouse model with the MARS model (See Appendix C3).
The ISO-NE Westinghouse modeling is analogous to what is used in GEBGE (which was the basis for PRISM).
They cite the use of the MARS LOD-UNCY table is a significant step in ensuring that the load models are
consistent and measure the same load model attributes/distributions.
A primary interest is the tails of the load model distributions between the two applications, which was discussed at
length at several RRAWG meetings during the 2009 RRAWG assessment of load model issues. The tails of these
distributions is where most of the LOLE occurs hence attention to the details of the distributions at these values is
very significant and important in the LOLE assessment efforts. (Example RRAWG presentations:
1. http://www.pjm.com/~/media/committees-groups/working-groups/rrawg/20090910/20090910item-03a-reporting-on-50-50-solution-loads.ashx
2. http://www.pjm.com/~/media/committees-groups/working-groups/rrawg/20090925/20090925item-03a-consistence-between-load-forecast-rrs-peak-load.ashx
3. http://www.pjm.com/~/media/committees-groups/working-groups/rrawg/20090731/20090731item-03a-truncating-normal-distribution.ashx
4. http://www.pjm.com/~/media/committees-groups/working-groups/rrawg/20090701/20090701item-03a-and-03b-load-model-selection.ashx
© PJM Interconnection 2011. All rights reserved
Page 104 of 136
Appendix C – Load Modeling Comparison Issues
Load Modeling parameters to use in the LOD-UNCY table
The LOD-UNCY values are associated with the deterministic inputted 8760 hourly loads that define the shape
used in the MARS daily peak solution method, using a Monte Carlo method.
Topics considered in this Appendix:



C1 - Load distribution granularity and impacts concerning greater granularity.
C2 - Choice of Sigma: Load model truncation discussion in the CJ Baldwin paper.
C3 - Use of LOD-UNCY table for consistent modeling between Westinghouse method and MARS.
This appendix provides further details around a central topic when comparing the PRISM and MARS modeling
and solution results.
Appendix C1 – Load distribution granularity
Load Distribution
Load distribution granularity and impacts concerning greater granularity
Assessment and discussion of the use of 21 points in the PRISM load distribution and the possible need for greater
49
granularity allowed for review of the Westinghouse Engineer paper, March 1969, by CJ Baldwin which is the basis
of the load modeling techniques used in PRISM. The PJM staff also independently performed an evaluation of
impacts around using various points for the daily peak distribution curve.
An issue to consider in this matter is that typically the Industry models that use MARS, apply a 7-point
distribution in the MARS LOD-UNCY table. This table has a maximum value of 10 points. Considering the
weekly distribution, with MARS assessing all seven days of the week, MARS considers 49 points each week (7
points in each of 7 days per week) while the use of the 21-point approach in PRISM considers 105 points per
week (21 points in each of 5 days per week).
There is a direct impact, almost one for one, in the MARS solution time by adding more points in the
distribution.
Assessment of number of points in daily peak distribution
PJM Staff summarized Adequacy analysis and spreadsheet analysis that assessed using up to 321 points in
the load distribution.





Figure C1 shows the change in LOLE, for the peak week, as a function of the number of points in the
load distribution.
The spreadsheet work indicates that the choice of 21 points for the distribution allows for statistically
valid results while addressing the need for solution time.
The results of the 21 point distribution cannot be found to be statistically more or less accurate than
the results seen for up to 321 points.
An issue in this type of assessment work has been typically called “Lumpiness”. This term relates that
the model does well in general aggregate assessment terms, but drilling further into the small values
being calculated may result in unexpected, somewhat quirky results. This might explain the
reason for what is seen when evaluating the 9-point distribution.
In addition, an assessment to evaluate the 41-point distribution was made. The distribution was
truncated at the zero point, and doubled the numbers of points up to 4.2 standard deviations. That
sensitivity case gave an IRM result 0.08% different than the 21 point result, partially explained due to
a different solved RI.
49
C.J. Baldwin, “Probability Calculation of Generation Reserves” (March 1969), published by The Westinghouse
Engineer, This paper is copyright protected but can be purchased online at Infotrieve; article information accession
number 00434361 (800-422-4633; www.infotrieve.com).
© PJM Interconnection 2011. All rights reserved
Page 105 of 136
Figure C1 – LOLE vs. number of points in Load model distribution
The conclusions drawn from this portion of assessment work were:





LOLE computed using different representations is function of the installed reserves
Representations using less than 21 points can result in either lower or higher LOLE’s for a given
reserve- Not suitable for estimating risk
LOLE computed stabilizes with representations of 21 points and beyond.
Increase in number of points beyond 21 has diminishing effects on LOLE.
Choice of representation should also consider computation time
© PJM Interconnection 2011. All rights reserved
Page 106 of 136
Appendix C2 – Baldwin Paper
Baldwin Paper: Choice of Sigma
Choice of Sigma: Load model truncation discussion in the CJ Baldwin paper.
The following PJM Staff presentation was delivered to the 19 August 2009 RRAWG meeting. This
presentation was developed to capture the essence and content of the aforementioned C.J. Baldwin paper
“Probability Calculation of Generation Reserves” (March 1969) paper written by CJ Baldwin, dated March
19699. This paper serves as the basis for the load modeling algorithms used in GEBGE and PRISM.
Per the CJ Baldwin paper the Years/Day Reliability Index (RI) saturates just above 4.0 standard deviations.
The choice of modeling out to 4.2 standard deviations in the PRISM model is justified by this paper’s Table VI,
shown below.
The discussion in the paper states “….A number of engineers (but rarely mathematicians) will argue that utility
loads are in fact not normal because load can never exceed connected load, which is certainly not infinite.
They argue that the load distribution should be truncated at connected load. The argument is valid, but the
important question is the effect of truncation on risk, which is illustrated in Table VI. ……..“
Analysis for the 2008 RRS single area model was performed to see if this type of relationship can be empirically
shown for the PJM model. The results of that analysis, adjusting the 21-point distribution to change the sigma,
shows that 4 standard deviations, or more, is a good modeling parameter choice based on the specifics of the PJM
system model.
© PJM Interconnection 2011. All rights reserved
Page 107 of 136
© PJM Interconnection 2011. All rights reserved
Page 108 of 136
© PJM Interconnection 2011. All rights reserved
Page 109 of 136
Appendix C3 – ISO-NE’s Comparison of Westinghouse Model and MARS
ISO-NE’s Comparison of Westinghouse Model and MARS
LOD-UNCY table for consistent modeling between Westinghouse method and MARS, version 3.00 .
A complete set of the following slides were presented at the 4/27/2010 NERC LOLE WG meeting and are available at: http://www.nerc.com/filez/lolewg.html .
The Westinghouse model is similar to PRISM’s single area model and calculations.
© PJM Interconnection 2011. All rights reserved
Page 110 of 136
© PJM Interconnection 2011. All rights reserved
Page 111 of 136
© PJM Interconnection 2011. All rights reserved
Page 112 of 136
© PJM Interconnection 2011. All rights reserved
Page 113 of 136
Several observations can be drawn from this presentation and subsequent work by PJM staff:

The essence of the comparison between Westinghouse and MARS is the same as the comparison
between PRISM and MARS.

Westinghouse and PRISM use the same analytical (convolution) solution techniques, with PRISM
having the ability to represent two areas while Westinghouse typically is used for a single area.

The PRISM model does not use the 3 moment (“skew-ness”) in its load model.

The capacity distributions between MARS and the analytical models, both Westinghouse and PRISM
can be shown as identical (see slide 9). PJM staff is developing these MARS distribution graphs using
Industry models and the Histogram post processing tool (HST-DATA).

The primary issue for consistency between the analytical models and MARS is the load model
distribution, followed by the transfer pipe limits.

Attention to the high load segment of the load distribution, which has a low probability of occurrence, is
imperative for proper LOLE modeling and assessments.

For single area models, similar LOLE results can be achieved when equivalent input assumptions are
used between the two modeling methods (analytical vs. Monte Carlo).

Equivalent input assumptions necessitates the use of the LOD-UNCY table:
rd
1.
2.
3.
Minimize least-square error between two distributions.
Identify more weight on the high load portion of the load distribution.
LOD-UNCY values will change as the associated hourly load shape values change.
© PJM Interconnection 2011. All rights reserved
Page 114 of 136
Appendix D – MARS public solution techniques
Two public information sources are given that describe the MARS solution methods used in the LOLE calculations
and reporting.
The first public information source, concerning MARS solution calculations and methods, is from the
NPCC CP-8 WG 2004 Summer assessment, Appendix C. This information is available on the NPCC web site
at: http://www.npcc.org/documents/reports/Seasonal.aspx.
NPCC SUMMER 2004
MULTI-AREA PROBABILISTIC RELIABILITY ASSESSMENT
NPCC CP-8 Working Group – May 2004 39 Final Report
APPENDIX C
Multi-Area Reliability Simulation Program Description
General Electric International, Inc.’s Multi-Area Reliability Simulation (MARS) program allows assessment of the
reliability of a generation system comprised of any number of interconnected areas.
Modeling Technique
A sequential Monte Carlo simulation forms the basis for MARS. The Monte Carlo method allows for many different
types of generation and demand-side options.
In the sequential Monte Carlo simulation, chronological system histories are developed by combining randomly
generated operating histories of the generating units with the inter-area transfer limits and the hourly chronological
loads. Consequently, the system can be modeled in great detail with accurate recognition of random events, such
as equipment failures, as well as deterministic rules and policies that govern system operation.
Reliability Indices
The following reliability indices are available on both an isolated (zero ties between areas) and interconnected
(using the input tie ratings between areas) basis:
.

Daily LOLE (days/year)

Hourly LOLE (hours/year)
LOEE (MWh/year)


Frequency of outage (outages/year)
Duration of outage (hours/outage)


Need for initiating Operating Procedures (days/year or days/period)
The CP-8 Working Group used both the daily LOLE and EOP indices for this analysis.
The use of Monte Carlo simulation allows for the calculation of probability distributions, in addition to expected
values, for all of the reliability indices. These values can be calculated both with and without load forecast
uncertainty.
The MARS program probabilistically models uncertainty in forecast load and generator unit availability. The
program calculates expected values of Loss of Load Expectation (LOLE) and can estimate each Area's expected
exposure to their Emergency Operating Procedures. Scenario analysis is used to study the impacts of extreme
weather conditions, variations in expected unit in-service dates, overruns in planned scheduled maintenance, or
transmission limitations.
Resource Allocation among Areas
The first step in calculating the reliability indices is to compute the area margins on an isolated basis, for each hour.
This is done by subtracting from the total available capacity in the area for the hour the load demand for the hour. If
an area has a positive or zero margin, then it has sufficient capacity to meet its load. If the area margin is negative,
the load exceeds the capacity available to serve it, and the area is in a loss-of-load situation.
If there are any areas that have a negative margin after the isolated area margins have been adjusted for
curtailable contracts, the program will attempt to satisfy those deficiencies with capacity from areas that have
© PJM Interconnection 2011. All rights reserved
Page 115 of 136
positive margins. Two methods are available for determining how the reserves from areas with excess capacity are
allocated among the areas that are deficient.

In the first approach, the user specifies the order in which an area with excess resources provides
assistance to areas that are deficient.

The second method shares the available excess reserves among the deficient areas in proportion to the
size of their shortfalls. The user can also specify that areas within a pool will have priority over outside
areas. In this case, an area must assist all deficient areas within the same pool, regardless of the order of
areas in the priority list, before assisting areas outside of the pool. Pool-sharing agreements can also be
modeled in which pools provide assistance to other pools according to a specified order.
Generation
MARS has the capability to model the following different types of resources:





. Thermal
. Energy-limited
. Cogeneration
. Energy-storage
. Demand-side management
An energy-limited unit can be modeled stochastically as a thermal unit with an energy probability distribution (Type
1 energy-limited unit), or deterministically as a load modifier (Type 2 energy-limited unit). Cogeneration units are
modeled as thermal units with an associated hourly load demand. Energy-storage and demand-side management
impacts are modeled as load modifiers.
For each unit modeled, the installation and retirement dates and planned maintenance requirements are specified.
Other data such as maximum rating, available capacity states, state transition rates, and net modification of the
hourly loads are input depending on the unit type.
The planned outages for all types of units in MARS can be specified by the user or automatically scheduled by the
program on a weekly basis. The program schedules planned maintenance to levelize reserves either on an area,
pool, or system basis.
MARS also has the option of reading a maintenance schedule developed by a previous run and modifying it as
specified by the user through any of the maintenance input data. This schedule can then be saved for use by
subsequent runs.

Thermal Units
In addition to the data described previously, thermal units (including Type 1 energy-limited units and
cogeneration) require data describing the available capacity states in which the unit can operate. This is
input by specifying the maximum rating of each unit and the rating of each capacity state as a per unit of
the unit's maximum rating. A maximum of eleven capacity states are allowed for each unit, representing
decreasing amounts of available capacity as governed by the outages of various unit components.
Because MARS is based on a sequential Monte Carlo simulation, it uses state transition rates, rather than
state probabilities, to describe the random forced outages of the thermal units. State probabilities give the
probability of a unit being in a given capacity state at any particular time, and can be used if you assume
that the unit's capacity state for a given hour is independent of its state at any other hour. Sequential
Monte Carlo simulation recognizes the fact that a unit's capacity state in a given hour is dependent on its
state in previous hours and influences its state in future hours. It thus requires the additional information
that is contained in the transition rate data.
For each unit, a transition rate matrix is input that shows the transition rates to go from each capacity state
to each other capacity state. The transition rate from state A to state B is defined as the number of
transitions from A to B per unit of time in state A:
Number of Transitions from A to B
TR (A to B) = _____________________________
Total Time in State A
© PJM Interconnection 2011. All rights reserved
Page 116 of 136
If detailed transition rate data for the units is not available, MARS can approximate the transition rates
from the partial forced outage rates and an assumed number of transitions between pairs of capacity
states. Transition rates calculated in this manner will give accurate results for LOLE and LOEE, but it is
important to remember that the assumed number of transitions between states will have an impact on the
time-correlated indices such as frequency and duration.

Energy-Limited Units
Type 1 energy-limited units are modeled as thermal units whose capacity is limited on a random basis for
reasons other than the forced outages on the unit. This unit type can be used to model a thermal unit
whose operation may be restricted due to the unavailability of fuel, or a hydro unit with limited water
availability. It can also be used to model technologies such as wind or solar; the capacity may be available
but the energy output is limited by weather conditions.
Type 2 energy-limited units are modeled as deterministic load modifiers. They are typically used to model
conventional hydro units for which the available water is assumed to be known with little or no uncertainty.
This type can also be used to model certain types of contracts. A Type 2 energy-limited unit is described
by specifying a maximum rating, a minimum rating, and a monthly available energy. This data can be
changed on a monthly basis. The unit is scheduled on a monthly basis with the unit's minimum rating
dispatched for all of the hours in the month.
The remaining capacity and energy can be scheduled in one of two ways. In the first method, it is
scheduled deterministically so as to reduce the peak loads as much as possible. In the second approach,
the peak-shaving portion of the unit is scheduled only in those hours in which the available thermal
capacity is not sufficient to meet the load; if there is sufficient thermal capacity, the energy of the Type 2
energy-limited units will be saved for use in some future hour when it is needed.

Cogeneration
MARS models cogeneration as a thermal unit with an associated load demand. The difference between
the unit's available capacity and its load requirements represents the amount of capacity that the unit can
contribute to the system. The load demand is input by specifying the hourly loads for a typical week (168
hourly loads for Monday through Sunday). This load profile can be changed on a monthly basis. Two types
of cogeneration are modeled in the program, the difference being whether or not the system provides
back-up generation when the unit is unable to meet its native load demand.

Energy-Storage and DSM
Energy-storage units and demand-side management impacts are both modeled as deterministic load
modifiers. For each such unit, the user specifies a net hourly load modification for a typical week which is
subtracted from the hourly loads for the unit's area.

Transmission System
The transmission system between interconnected areas is modeled through transfer limits on the
interfaces between pairs of areas. The transfer limits are specified for each direction of the interface and
can be changed on a monthly basis. Random forced outages on the interfaces are modeled in the same
manner as the outages on thermal units, through the use of state transition rates.

Contracts
Contracts are used to model scheduled interchanges of capacity between areas in the system. These
interchanges are separate from those that are scheduled by the program as one area with excess capacity
in a given hour provides emergency assistance to a deficient area. Each contract can be identified as
either firm or curtailable. Firm contracts will be scheduled regardless of whether or not the sending area
has sufficient resources on an isolated basis, but they can be curtailed because of interface transfer limits.
Curtailable contracts will be scheduled only to the extent that the sending Area has the necessary
resources on its own or can obtain them as emergency assistance from other areas.
© PJM Interconnection 2011. All rights reserved
Page 117 of 136
The second public information source, concerning MARS solution calculations and methods, is available on the NERC web site at:
http://www.nerc.com/filez/lolewg.html . (Agenda Item 7D)
General Electric International, Inc. (GEII) Presentation Slides
© PJM Interconnection 2011. All rights reserved
Page 118 of 136
© PJM Interconnection 2011. All rights reserved
Page 119 of 136
© PJM Interconnection 2011. All rights reserved
Page 120 of 136
© PJM Interconnection 2011. All rights reserved
Page 121 of 136
Appendix E – Transparency of process to perform LOLE calculations
PRISM
Detailed output of almost all aspects of the input processing, calculation steps, solution parameters and reported results can be observed by using the DEBUG
option for a given PRISM run. All DEBUG information is stored based on the PRISM run number. In the final reporting the PRISM run number is shown as it
allows a quick reference for any follow-up detailed inquires.
The location of the SAS data tables on the PJM Local Area Network is :
\\sas12vwp\sasAdHoc$\sasDepts\CapPlan\\sasAdHoc\sasDepts\CapPlan\PRISM_Debug_Folder
The following summary lists the information available, on an area and delivery year basis. These SAS data table summaries have been used to perform detailed
comparisons with the GEBGE calculations.

List of information in the DEBUG directories, for a given PRISM run number:
o
EWM Calculations - 52 Weekly values for: Original GrowX (historic growth rate), gsdn (mean), EWM, growx (final growth rate using load
forecast).
o
PRISM Parms: Indicates general solution parameters, name of user, solution run time, source location of 1) input file, 2) files containing
capacity base units, adders, changes, maintenance schedules, 3) yearly constants, 4) probability table 5) Capacity and load factors 6)
load models, 7) tie size and 8) solved data results.
o
PRISM Time: Start and end times for various steps in the solution process.
o
Period Probability: Similar to the output shown in Table 2 for the period probability category.
o
Summary Table: Similar to the output shown in Table 2 for the summary table category
o
Syspar (unit model aggregate parameters): Underlying data used to create the output in Table 2 for the Syspar category including
distribution values.
o
Weekly Solution: Values associated with the output shown in Table 2 for the period probability category.
o
Yearly Solution: The automatic solution process loads and resulting RI to iterate to the solution tolerance for the given desired RI (Typically
10 years - per occurrence)
o
Weekly EWM: Information similar to the Expected Weekly Maximum (EWM) calculations.
o
Control Card Data: The term “Control Card” is based on the historic card readers used to input data. This term was carried over to PRISM.
This is a direct list of the individual input data showing what was used for all input values. There are over twenty individual control cards
that are used for all aspects of the model.
© PJM Interconnection 2011. All rights reserved
Page 122 of 136


o
Outage Schedule: Individual unit schedule for planned outages, indicating the starting week and duration of the outage. This schedule is
typically automatically determined by the PRISM algorithms to maintain a constant level of available reserves over the delivery year
period.
o
Probable Peaks: 52 Weekly probable peak (mean) and system parameters (SYSPAR - System Unit Average) variable values.
o
Unit Descriptors: Individual list of Units in the model indicating Area number (1 or 2), Name, Unit number, ID, zone number and zone
descriptor.
There are 3 sub directories for each directory year:
o
Capacity Model: Data sets include added generation, base generation, capacity changes, weekly mlnst (MWs out on PO) variable, scheduled
planned maintenance, Base unit list (out area), All units(Base plus future units)
o
Input Data: added generation, area data, base generation, capacity changes, load model, schedule planned outages, yearly peak loads.
o
Load Model: plots load model variables, weekly load probable peak and capacity, yearly peak load.
The following files are for each Delivery year of a PRISM Run:
o
Area adder units
o
Area capacity changes
o
Area capacity model
o
Area capacity summary
o
Area capacity factors
o
Area cumulative probability table
o
Area load model
o
Area load model factors: weekly load model mean and standard deviation values.
o
Area manual maintenance
o
Area outage schedule
o
Area unit descriptors
o
Area unit list
o
Area unit data set
© PJM Interconnection 2011. All rights reserved
Page 123 of 136
o
Area Unit values by week
o
Area Updated model
o
Area Weekly EWM: Orig_growx, gsdn, EWM (MPP), growx.
o
Area Weekly capacity factors
o
Area data
o
Capacity changes intermediate steps, Temp, Temp2, Temp 3a, and Temp3b.
o
Core calculation intermediate steps: There are 3 places in the calculation process that are available for review. Core1 shows the load model
processing. Core2 shows automatic solution processing. Core5 has a format and data similar to the period probability table discussed in Table
2.
o
Unit availability state in each of 52 periods(Cumprob1)
o
Intermittent steps for capacity processing: Added generation, Adder units, capacity changes, capacity model.
o
Final run flag that indicates best solution iteration in automatic solution: main_finalprob, main_finalprobability, Main_finalrunflag
o
Several capacity model intermediate steps: unitds_1, updated_mode, schedule1, schedule2, unitdescriptors, units_area1, unitvaluesbyweek,
weeklycapadders, weeklycapearlyadders, weeklycapearlychanges, weeklycapmodel, weeklycapsortchanges, weeklycapsortmodel,
weeklycapstartmw.
o
Solution primary load, secondary load, and LOLE during automatic solution process. Weeklyoutptds, yearlyoutputds.
o
Load model intermediate steps involved with load distribution (zprob): weekly sigma value and associated probability (typically 21 points for
each week for each area). Weekly number of points in distribution. (Zprob, zprobtable, zsummary).
It was noted by the RRAWG recently and confirmed by the PJM staff that output of the internal calculated 21-point steps is not currently saved or available after
the final solution is determined. However the PJM staff has identified the issues needed to be addressed for this reporting and have established a plan to
complete the processing to output these intermediate steps. Due to the modular nature of the code and its ability to automatically find a solution, these 21-point
internal calculation values were not originally output due to the large file size.
MARS
OT 10 File
In the “Years to Study and annual Output” Options (pages 3-14 of the February 2009 MARS user manual) the Hourly output is listed in a column format shown in
Table 44. The Clock column starts at Hour 1, for which there are 8760 hours for a single replication. The clock does not get reset, so the value for the 2nd
replication starts at 8761.
© PJM Interconnection 2011. All rights reserved
Page 124 of 136
The load level is associated with the values given in the LOD-UNCY table, typically there are seven values (1 = largest Load level, 4 = 50/50 load level).
The Margin is from the EOP Table; typically there are six EOP levels used: 1 = Operating reserves, 2 = Load Management, 3 = 30-minute reserves, 4 = Voltage
Reduction, 5 =10-Minute reserves, and 6 = Appeals).
The isolated /Interconnected column (ISO/INT) indicates a 1 = Isolated and 2 = Interconnected ISO, and the area name columns show the amount of reserves
that were deficient causing an LOLE event, for that EOP level and LOD-UNCY load amount.
By using these values and the associated input probabilities, each replication calculates the resulting LOLE for the various LOD-UNCY distribution points and
EOP levels defined in the input data. (This output file can be very large (gigabytes).)
To measure and determine the values as shown in the OT9, values could be processed for each replication. The possible outcomes for a study that requires, for
example, 2000 replications, to solve to an acceptable Standard Error, assuming that 500 hours have a non zero risk(a negative value for load) could be 2000
(replications) x 500 (contributing hours) x 7(LFU values) x 6 (EOP levels) = 42 million values to evaluate.
It is important to note that the process would need to place the risk into the correct EOP bucket and that a check needs to be made to make sure that a given
hour is the daily peak, to match the reporting in the MARS Standard output OT files when the solution convergence index is option 3, for LOLE in Days/Year
(reference CNV-CRIT-00 Table). This process needs to be developed and once successful, automated, so it can be part of the GUI used to process any user
chosen model’s area.
Table 44 – MARS Hourly Output
CLOCK
LOAD LEVEL
MARGIN
ISO / INT
Area Margin
4193.0
4505.0
4505.0
4528.0
4529.0
4529.0
5369.0
5369.0
5369.0
5441.0
13265.0
13289.0
13289.0
13289.0
13310.0
13311.0
13312.0
13313.0
1
1
1
1
1
1
1
1
1
1
1
1
1
0
0
0
0
1
1
1
0
0
1
2
1
2
0
1
1
1
2
1
1
1
1
1
2
2
2
2
2
2
2
2
2
2
2
2
2
2
2
2
2
2
-1690.08
-3804.65
-477.65
-652.67
-4675.25
-746.25
-4836.28
-907.28
-1509.28
-2635.19
-613.57
-5173.03
-1244.03
-1846.03
-1449.30
-5661.37
-6784.05
-10463.62
© PJM Interconnection 2011. All rights reserved
Page 125 of 136
Additional details are available in the OT8, OT7, and OT 11 files, as described in Table 2 in the Period Probability category.
Comparison comments.
The determination of the capacity distributions in each of the MARS replications is not a simple and direct effort. The capacity outage distribution summary
might take more time for MARS while the load distribution summary in PRISM takes more time.
There are many, many calculation step values available in both tools. Both tools offer a process to improve the transparency of the processing steps by
requesting improvements to the staff that maintains and owns the intellectual property rights of the tool.
© PJM Interconnection 2011. All rights reserved
Page 126 of 136
Appendix F – Details of Input Parameter Requirements
The following shows the input parameters available for both PRISM and MARS. The specifics given should
allow the reader to discern the level of effort needed to make an analysis run, to observe the differences in
the event driven MARS inputs vs. statistical quantities for PRISM, and to capture the breadth and depth of
what is available to allow the tools to provide assessments for a large number of various and diverse
assessments needs. Both tools need hourly loads and individual unit GADS event data to perform an
assessment.
PRISM
The PRISM file format is automatically generated via a JAVA based GUI. The format uses a Control Card
(CC) category which is a holdover from historic formats used in the 1960s.
CC1 – Report Title
Cols 1-5
Cols 6-61
Control Code (01)
Report Title – One title line per area
CC2 – Report Remarks
Cols 1-5
Cols 6-85
Control Code (02)
Report Remarks
CC3 – Case Identification
Cols 1-5
Cols 6-13
Control Code (03)
Case ID
CC7 – Constant Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-16
Cols 17-26
Cols 27-31
Cols 32-36
Cols 37-41
Cols 42-44
Cols 45-46
Cols 47-56
Cols 57-66
Cols 67-69
Control Code (07)
Area Number
Study Year
Simulated Maintenance (Load?)
Probability Table Limit
Target Solved Reliability Index (RI)
Tolerance
Per Unit Reserve Factor
Days per Week
Calendar Week Equal to Period 1
First Order Statistic
Forecast Error Factor
Step Size
CC8 – Season Data
Cols 1-5
Cols 6-13
Cols 14-21
Cols 22-29
Cols 30-37
Control Code (08)
Spring
Summer
Fall
Winter
CC9 – Case Data
Cols 1-5
Cols 6-10
Cols 11-15
Control Code (09)
Study Start Year
Study End Year
© PJM Interconnection 2011. All rights reserved
Page 127 of 136
Cols 16-17
Cols 18-22
Col 25
Cols 26-27
Cols 28-59
Cols 60-61
Cols 62-63
Cols 64-65
Cols 66-75
Cols 76
Cols 77
Cols 68-107
Cols 108-147
Number of Areas
LAS Year
Modified Two-state (Y/N)
Normal Flag (Y/N)
Non-normal or Normal Curve data set name
CETO/CETL Flag (Y/N)
CETO/CETL Adders Flag (Y/N)
EFORd Flag (Y/N)
Class Average Years
PSSE Flag (Y/N)
OpGen Type (R=Rollup, D=Detail)
PSSE File Name
PSSE Adder File
CC10 – Distribution Data
Cols 1-5
Cols 6-15
Cols 16-25
Cols 26-27
Control Code (10)
Per Unit Sigma
Probability of Occurrence
Week Number
CC12 – Load Levels, Tie Sizes, and Margin Levels
Cols 1-5
Cols 6-10
Col 11
Cols 12-17
Cols 18-23
Cols 24-29
Cols 30-35
Cols 36-41
Cols 42-47
Cols 48-53
Cols 54-59
Cols 60-65
Cols 66-71
Cols 72-77
Control Code (12)
Year
cc12 Type (1-Load Levels, 2-Tie Sizes, 3-Margin Levels)
Value 1
Value 2
Value 3
Value 4
Value 5
Value 6
Value 7
Value 8
Value 9
Value 10
Value 11
CC16 – Capacity Factor Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-19
Cols 20-27
Cols 28-35
Cols 36-43
Cols 44-46
Cols 47-49
Control Code (16)
Area Number
Year
Spring Factor
Summer Factor
Fall Factor
Winter Factor
Start No Maintenance Period
Stop No Maintenance Period
CC17 – Load Model Data
Cols 1-5
Cols 6-7
Cols 8-15
Cols 16-23
Cols 24-25
Cols 26-37
Cols 38-43
Cols 44-93
Control Code (17)
Area Number
Mean or Trend Value
Standard Deviation
Week
Title
Load Model Type (Plots or Trend)
(Text used in recalling saved data)
© PJM Interconnection 2011. All rights reserved
Page 128 of 136
Cols 94-95
Cols 96-104
Season (1, 2, 3, 4)
Study Period (Ex. 1997-2001)
CC18 – Load Factor Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-13
Cols 14-21
Cols 22-29
Cols 30-37
Cols 38-45
Cols 46-53
Cols 54-61
Cols 62-69
Cols 70-77
Cols 78-85
Cols 86-93
Cols 94-101
Cols 102-109
Cols 110-117
Cols 118-125
Cols 126-133
Cols 134-141
Cols 142-149
Cols 150-157
Cols 158-165
Cols 166-173
Cols 174-181
Cols 182-189
Cols 190-197
Cols 198-205
Control Code (18)
Area Number
Year
Number of Factors (4 or 12)
Spring Factor or FCST Load Month 1
Summer Factor or FCST Load Month 2
Fall Factor or FCST Load Month 3
Winter Factor or FCST Load Month 4
FCST Load Month 5
FCST Load Month 6
FCST Load Month 7
FCST Load Month 8
FCST Load Month 9
FCST Load Month 10
FCST Load Month 11
FCST Load Month12
ALM Month1
ALM Month2
ALM Month3
ALM Month4
ALM Month5
ALM Month6
ALM Month7
ALM Month8
ALM Month9
ALM Month10
ALM Month11
ALM Month12
CC20 – Base Unit Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-33
Cols 34-38
Cols 39-43
Cols 44-49
Cols 50-55
Cols 56-59
Cols 60-61
Cols 62-63
Cols 64-65
Cols 66-67
Cols 68-69
Cols 70-71
Cols 72-73
Cols 74-75
Cols 76-77
Cols 78-79
Cols 80-82
Cols 83-122
Cols 123-186
Cols 187-198
Cols 199-203
Cols 204-211
Control Code (20)
Area Number
Unit Number
Unit Name
Summer Capability
Winter Capability
Variance
EEFORd
Maintenance Start Year
1st Maintenance Cycle
2nd Maintenance Cycle
3rd Maintenance Cycle
4th Maintenance Cycle
5th Maintenance Cycle
6th Maintenance Cycle
7th Maintenance Cycle
8th Maintenance Cycle
9th Maintenance Cycle
10th Maintenance Cycle
Zone Number
Zone Name
Owner Index
GEBGE Name
Unique Generator Name
MARS Name
© PJM Interconnection 2011. All rights reserved
Page 129 of 136
Cols 212-217
Cols 218-220
Cols 221-230
Cols 231-240
Cols 241-246
EFORd Value
Gen Class Key
Load Model Code
SAS Loc Key (From DM_TBLS.Geography_Xref)
XEFORd
CC20A – Case Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-51
Cols 52-61
Cols 62-64
Control Code (20A)
Area Number
SAS Loc Key (From DM_TBLS.Geography_Xref)
Zone Name
Load Model Code
Zone Number
CC21 – Adder Units
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-33
Cols 34-37
Cols 38-39
Cols 40-44
Cols 45-50
Cols 51-56
Cols 57-58
Cols 59-60
Cols 61-62
Cols 63-64
Cols 65-66
Cols 67-68
Cols 69-70
Cols 71-72
Cols 73-74
Cols 75-76
Cols 77-81
Cols 82-84
Cols 85-124
Cols 125-188
Cols 189-200
Cols 201-212
Cols 213-220
Cols 221-226
Cols 227-229
Cols 230-239
Cols 240-244
Cols 245-250
Control Code (21)
Area Number
Unit Number
Unit Name
Install Year
Install Period
Summer Capacity
Variance
EEFORd
1st Maintenance Cycle
2nd Maintenance Cycle
3rd Maintenance Cycle
4th Maintenance Cycle
5th Maintenance Cycle
6th Maintenance Cycle
th
7 Maintenance Cycle
8th Maintenance Cycle
9th Maintenance Cycle
10th Maintenance Cycle
Winter Capacity
Zone Number
Zone Name
Owner Index
Load Model Code
GEBGE Name
MARS Name
EFORd
Gen Class Key
SAS Loc Key
Unique Gen Key
XEFORd
CC22 – Capacity Changes (There can be multiple records per unit per year)
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-33
Cols 34-35
Cols 36-40
Cols 41-42
Cols 43-46
Cols 47-48
Cols 49-53
Control Code (22)
Area Number
Unit Number
Unit Name
Period Number
New Capacity
Change Month
Change Year
Unused placeholder
Winter Capability
© PJM Interconnection 2011. All rights reserved
Page 130 of 136
Cols 54-56
Cols 57-96
Cols 97-108
Cols 109-113
Zone Number
Zone Name
GEBGE Name
Unique Gen Key
CC23 – Manual Maintenance
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-33
Cols 34-37
Cols 38-39
Cols 40-42
Cols 43-44
Cols 45-46
Cols 47-48
Cols 49-50
Control Code (23)
Area Number
Unit Number
Unit Name
Change Year
Number of Periods
Zone Number
Period 1
Duration 1
Period 2
Duration 2
The periods and durations continue in pairs for the number of periods specified in columns 38-39.
Zone Descriptor is added to the end of the string starting one column after the periods and durations finish
and is 40 characters long. GEBGE Name is added after Zone Name for 12 characters. Unique Gen Key is
added after GEBGE for 5 characters.
CC24 – Tie Size
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-21
Control Code (24)
Area Number
Year
Tie Size
CC25 – Curtail Data
Cols 1-5
Cols 6-9
Cols 10-19
Cols 20-29
Cols 30-45
Cols 46-50
Col 51
Control Code (25)
Year
ALM
Reserve Value
Generator Name
Large Generator Threshold
ALM Run Type
CC26 – Solve Data
Cols 1-5
Cols 6-7
Cols 8-11
Cols 12-19
Cols 20-21
Cols 22-23
Cols 24-25
Cols 26-27
Cols 28-29
Control Code (26)
Area Number
Year
Load
Automatic (1 is automatic)
Print Seasonal LOLE
Print Planned Outages
Print Loads
Print Stored
CC31 – Generators table for PSSE
Cols 1-5
Cols 6-12
Cols 13-15
Cols 16-24
Control Code (31)
PSSE Bus Number
PSSE Machine Id
PSSE Bus Name
© PJM Interconnection 2011. All rights reserved
Page 131 of 136
Cols 25-54
Cols 55-59
Cols 60-64
Cols 65-66
Cols 67-68
Cols 69-90
Cols 91-100
Cols 101
Cols 102-111
Cols 112-115
Cols 116-119
Cols 120-149
Cols 150-153
Cols 154-155
PSSE Generator Name
PSSE Max MW
Operating Generator Capacity
PSSE Bus Count
Operating Generator Count
Plant Name
Generator Code EIA
Activity Code (1 = Adder, 2 = Change record, 3 = Retire)
Commercial Date
Generator Key Match
Generator Key
Non-unique Generator Name
ARC Year
ARC Week
CC32 – PSSE No match table for PSSE
Cols 1-5
Cols 6-12
Cols 13-15
Cols 16-24
Cols 25-29
Cols 30-36
Cols 37-46
Cols 47-49
Control Code (32)
PSSE Bus Number
PSSE Bus ID
PSSE Bus Name
PSSE Max MW
PSSE Area Name
Load Model Code
cc32 Type
CC33 – PJM Gens table (or all Operating Generators) for PSSE
Cols 1-5
Cols 6-27
Cols 28-37
Cols 38-70
Cols 71-73
Cols 74-78
Cols 79-83
Cols 84-89
Cols 90-95
Cols 96-97
Cols 98-103
Cols 104-106
Cols 107-170
Cols 171-182
Cols 183-190
Cols 191
Cols 192-201
Cols 202-205
Cols 206-207
Cols 208-214
Cols 215-220
Control Code (33)
Operating Generators Plant Name
Generator Code EIA
Zone Name
Sub-zone Number
Summer Capability
Winter Capability
EFORd
EEFORd
POF
Variance
Gen Class Key
Owner
GEBGE Name
GEMARS Name
Activity Code (1 = Adder, 2 = Change record, 3 = Retire)
Commercial Date
ARC Year
ARC Week
PSSE Bus ID
XEFORd
CC9999 – Separator
Cols 1-5
Control Code (9999)
© PJM Interconnection 2011. All rights reserved
Page 132 of 136
MARS
Table Category
Name
*
# Inputs
Description
General Study Data
GEN-CASE
GEN-POOL
GEN-AREA
GEN-ARGP
GEN-UNTY
GEN-PRNT
GEN-TIME
CNV-CRIT
GEN-OPTN
MNT-AOPT
INT-ONLY
11
2
6
3
2
4
9
13
5
6
12
Case Identification and General Options
Pool Identification
Area Identification
Area Group Identification
Unit Summary Type Identification
General Print Information
Years to Study and Annual Output Options
Simulation Convergence Criteria
Program Options
Annual Output Options
Infrequently Used Programs and Output Options
6
4
4
Annual Area Load Data
Annual Month-to-Annual Ratios
Load Forecast Uncertainty
7
General Unit Data
2
2
4
4
3
4
Capacity States
Thermal Unit Maximum Capacity
Thermal Unit Capacity Derations
Transition Rate Data
Forced Outage Rates
Number of State Transitions
6
3
4
2
7
2
2
2
2
2
Unit Maintenance Data Options
Thermal Unit Fixed Daily Maintenance
Non-Thermal Unit Fixed Daily Maintenance
ES and DSM Capacities for Maintenance Scheduling
Maintenance Scheduling Options
Cycles for Per Unit Planned Outage Rates
Cycles for Weeks of Planned Maintenance
Annual Peaks for Maintenance Load Model
Weekly Ratios for Maintenance Load Model
Capacity Adjustments for Maintenance Scheduling
Load Data
LOD-DATA
LOD-MTAR
LOD-UNCY
General Unit Data
UNT-DATA
Thermal Unit Data
UNT-CAPS
UNT-MXCP
UNT-DERT
UNT-TRNS
UNT-FORS
NUM-TRNS
Maintenance Data
MNT-UNOP
MNT-FIXD
MNT-FDMD
MNT-MDCD
MNT-OPTN
MNT-PLOR
MNT-WEEK
MNT-ANPK
MNT-RATI
MNT-FIXC
Type 1 Energy-Limited Unit
ELU-DIST
3
Unit Outage Distribution
Non-Thermal Units and Load Modifiers
MOD-ELMW
MOD-ELDD
5
3
Ratings for Type 2 Energy-Limited Unites
Loads for Scheduling Type 2 Energy-Limited Units
© PJM Interconnection 2011. All rights reserved
Page 133 of 136
Table Category
Name
MOD-CGMW
MOD-MDMW
MOD-SHAP
MOD-PENE
MOD-PRIO
# Inputs*
Description
2
2
3
2
1
Cogeneration Unit Hourly Load Demand
ES and DSM Net Hourly Load Modification
DSM Hourly Shapes
DSM Hourly Shape Penetration Factors
Priority Order for Scheduling Non-Thermal Units and Load Modifiers
7
Contract Data
4
4
3
5
11
Definition of Area Interfaces
Definition of Interface Groups
Identification of Closed Interface Groups
Interface Transfer Limits
Dynamic Interface Transfer Limits
1
4
Priority Order for Allocating Resources among Areas
Pool Reserve Sharing Data
Contracts
FCT-DATA
Area Interfaces
INF-DATA
INF-GRPS
INF-CLSD
INF-TRLM
INF-DYLM
Resource Allocation
RES-PRIO
RES-POOL
Emergency Operating Procedures (EOPs)
EOP-DATA
EOP-DLAY
11
2
Margin State Data for Modeling Emergency Operating Procedures
Data for Delayed Implementation of Emergency Operating Procedures
Histogram Post-Processor Data
HST-DATA
Total Parameters
7
Indices to be processed by HISTO
225*
*
The number of inputs indicated is an average estimate, for typical modeling. However they can vary due to a given
assessment model requirements. The full MARS User Manual is the best source for determination of input requirements.
The MARS User Manual is available to: 1) Licensed MARS users 2) parties that have executed a signed Use and nonDisclosure Agreement (NDA) with GEII.
In total, MARS has 50 input tables that allow approximately 225 parameters to be entered. The commonly
used tables are highlighted in yellow, which total 22 Tables and 121 parameters. There are a few additional
files necessary for analysis to start including the in02 file that stores the 8760 hourly loads, the in17 file that
has a resource hourly energy profile (MOD-SHAP), a control file, and a batch file.
In the proprietary MARS user manual the input parameter requirements are detailed over approximately 70
pages of text and figures. This manual is an excellent and detailed reference, extremely useful in
manipulating the various options that MARS allows the user to choose.
Taking advantage of MARS’ strengths requires using other Table inputs such as: a) the MOD-SHAP table to
specific wind unit profiles, b) MOD-ELMW and MOD-ELDD to investigate hourly pattern impacts, c) INFDYLM to adjust imports over several paths and d) HST-DATA to investigate and report underlying model
calculations and details.
Of the 50 MARS tables, there are several tables that are populated automatically via the ARC Java-based
GUI.
© PJM Interconnection 2011. All rights reserved
Page 134 of 136
Input parameter comparison comments.
Due to the length of time that the PJM Staff has used PRISM, there are many more automated facilities and
tools for PRISM data inputs than there are for MARS inputs. Many of the MARS tables require manual
intervention to establish correct entries. Both PRISM and MARS store the final input data file in a text-based
format which does allow for quick access to make changes.
MARS has more categories of input parameters; however both PRISM and MARS have load and capacity
inputs that enable many different assessments. MARS typically needs Emergency Operational Procedure
(EOP) data, reserve sharing, contract data, transmission pipe data, and load forecast uncertainty values for
each area modeled.
The detailed input requirements for PRISM are shown above. The MARS detailed input requirements are
covered in 70 pages of the MARS User Manual. PRISM has much detailed input requirements available in
its GUI and with approximately the same level of detailed as given in the MARS User Manual. There are 50
categories for MARS Tables and 23 categories for PRISM’s Control Cards (CC).
Both PRISM and MARS are part of the Application for Reliability Calculations (ARC) in the PJM Production
environment. The data is fed into the input tables via the database schemas and established process
options of the GUI.
© PJM Interconnection 2011. All rights reserved
Page 135 of 136
Appendix G – Items for possible future investigation
There are a number of future investigation activities that can supplement and augment this PRISM-MARS
comparison report. These items not only enhance the existing complementary tasks between both
programs but can also increase stakeholder understanding of each program.
The following potential future PRISM-MARS study activities are listed in priority order with deference to
resources and time needed to complete each task:
1.
While currently dedicated to using PRISM for resource adequacy studies, PJM also uses several
key MARS assessment modules to supplement this work (discussed in the “complementary”
section of this report.) For the benefit of PJM stakeholders, PJM Staff should make greater efforts
to clarify and document how MARS is already used in conjunction with PRISM for LOLE
assessments.
2.
With ISO-NE Staff, PJM Staff should more fully explore the use of PRISM’s probabilistic load
modeling (PLOTS) in assisting the development of the MARS Load Forecast Uncertainty table
(LOD-UNCY).
3.
MARS can assist the evaluation of high ambient conditions on generation resources and
performance by using its UNT-DERT table. (This is related to the item shown in the 2010 RRS
report, shown in section RRAWG activities, 4th bullet [page 48].)
4.
Interpret and develop analytical processes for reporting MARS output data (such as OT6, OT7,
OT8, OT9 and OT11 output files); including processes for MARS post-processing tools such as the
Histogram application.
5.
Compare and analyze the MARS hourly load shape file (in02 file) for reporting differences seen
with calendar year 2002, 2005 and 2006 – years identified in the industry as possible hourly load
shapes. Investigate a process for developing a “design year”.
6.
Investigate the NYSRC’s “Unified Methodology” procedure that coordinates and links the New York
Control Area (NYCA) IRM requirements with the Locational Capacity Requirements (LCRs)
developed by the NYISO. (LCRs are developed separately for the NYCA Zones J (New York City)
and K (Long Island) and are differentiated from the NYCA Rest of State (ROS).)
7.
Investigate the NYISO and MISO’s use of reporting IRM with MARS.
8.
Consider items in the PJM neighboring region’s MARS model and assessment methods. If needed
enter into a Non Disclosure agreement (NDA), so that any investigation is comprehensive and
measurable. These MARS modeling methods should initially be focused on the development of the
following MARS tables and coordinated within the NPCC CP-8 WG, MISO LOLE WG, IPSAC, and
the SERC Corporation RAWG. A long term goal of these efforts could be the development of a
process to transfer the following MARS data into the PJM database schemas.
o
o
o
o
o
o
o
LOD-UNCY values - Load uncertainty
In02 hourly load file values
EOP-DATA values - Emergency Operating Procedures levels
INF-TRLM values - Transmission pipe sizes
Specific Path related contracts
Severe case assumptions or sensitivity assumptions
Evaluation to aggregate supplied details to match the PJM data model needs.
© PJM Interconnection 2011. All rights reserved
Page 136 of 136
End: Comparison of PRISM and MARS
February 17, 2011
DOCs #599848
PJM Resource Adequacy Analysis Subcommittee (RAAS)
© PJM Interconnection 2011. All rights reserved
Last Page