Uploaded by vrindavan00

Boylan20Syntetos20and20Karakostas20200820-20JORS

advertisement
See discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/32021906
Classification for forecasting and stock control: A case study
Article in Journal of the Operational Research Society · April 2008
DOI: 10.1057/palgrave.jors.2602312 · Source: OAI
CITATIONS
READS
118
3,937
3 authors, including:
John Boylan
Buckinghamshire New University
73 PUBLICATIONS 2,483 CITATIONS
SEE PROFILE
All content following this page was uploaded by John Boylan on 25 June 2014.
The user has requested enhancement of the downloaded file.
Journal of the Operational Research Society (2008) 59, 473 --481

2008 Operational Research Society Ltd. All rights reserved. 0160-5682/08 $30.00
www.palgrave-journals.com/jors
Classification for forecasting and stock control:
a case study
JE Boylan1∗ , AA Syntetos2 and GC Karakostas3
1 Buckinghamshire
and
3 Eurometal
Chilterns University College, Buckinghamshire, UK; 2 University of Salford, Salford, UK;
SA, Greece
Different stock keeping units (SKUs) are associated with different underlying demand structures, which in
turn require different methods for forecasting and stock control. Consequently, there is a need to categorize
SKUs and apply the most appropriate methods in each category. The way this task is performed has significant
implications in terms of stock and customer satisfaction. Therefore, categorization rules constitute a vital
element of intelligent inventory management systems. Very little work has been conducted in this area and,
from the limited research to date, it is not clear how managers should classify demand patterns for forecasting
and inventory management. A previous research project was concerned with the development of a theoretically
coherent demand categorization scheme for forecasting only. In this paper, the stock control implications of
such an approach are assessed by experimentation on an inventory system developed by a UK-based software
manufacturer. The experimental database consists of the individual demand histories of almost 16 000 SKUs.
The empirical results from this study demonstrate considerable scope for improving real-world systems.
Journal of the Operational Research Society (2008) 59, 473 – 481. doi:10.1057/palgrave.jors.2602312
Published online 18 October 2006
Keywords: categorization; forecasting; inventory; intermittent demand; case study
Introduction
Inventory management is a very complex problem area and
it is this complexity that has necessitated the development
of intelligent systems to assist decision making. Ultimately,
ordering decisions must be made at the level of the individual
item or product. Thus, operational rules for the classification of stock keeping units (SKUs) are essential components
of inventory systems. They allow the appropriate amount
of managerial attention and the right forecasting and stock
control methods to be used on the right products. The classification of SKUs has significant implications for stock-holdings
and customer service levels (CSLs).
Demand categorization has attracted very limited academic
interest. There has been a considerable amount of research
on classifying inventory and forecasting methods, but not on
distinguishing between different demand patterns to guide
forecasting and stock control. Notable exceptions are the
papers by Williams (1984) and Eaves and Kingsman (2004).
Although some work has been done on the implications of
forecasting on stock control (eg, Watson, 1987; Gardner,
1990; Strijbosch et al, 2000), more theoretical and empirical
research is much needed in this area.
∗ Correspondence: JE Boylan, Buckinghamshire Business School,
Buckinghamshire Chilterns University College, Chalfont St Giles,
Buckinghamshire, HP8 4AD, UK.
E-mail: john.boylan@bcuc.ac.uk
The selection of forecasting and stock-control methods is
generally based on the values of parameters that define the
current state of the SKU. The parameters most commonly
applied for this purpose relate to the underlying demand
patterns. Other relevant parameters, not considered in this
research, include discount break-points and foreign exchange
rates (see eg, Kobbacy and Liang, 1999).
This paper offers some empirical evidence on demand categorization based on underlying demand patterns. The forecasting and stock-control classification system employed by a
UK-based software manufacturer is examined. The company
sought to improve the classification capability of its software
by (i) considering alternative classification parameters and
(ii) removing hard-coded classification parameter values and
enabling users to select their own values from appropriate
menus. It is the standard practice of the manufacturer to test
all new approaches extensively, prior to adoption, to establish
‘proof of concept’. The empirical database supplied by the
company for this purpose consisted of the individual demand
histories of approximately 16 000 SKUs from the automotive,
aerospace and chemical industries.
The remainder of the paper is structured as follows. Firstly,
a categorization framework is presented. Secondly, the
company’s old method of classification is summarized. A literature review follows. The objectives of the empirical investigation are stated. These are informed both by the company’s
old classification method and recent theoretical developments.
474
Journal of the Operational Research Society Vol. 59, No. 4
The next sections explain the empirical findings, from forecasting and stock-control perspectives. Unresolved issues are
discussed as part of a ‘research agenda’ for demand categorization. Conclusions are summarized at the end of the paper.
Categorization framework
Infrequent demand occurrences or irregular demand sizes,
when demand occurs, do not allow lead time demand to be
represented by the normal distribution. In these cases, demand
will be called non-normal. By using this nomenclature, it is
not meant that all regular, non-sporadic, non-lumpy demands
are necessarily normally distributed. A different distribution
may be more appropriate, although the normality assumption
should reasonably represent many real-world cases (see eg,
Silver et al, 1998).
A rather modest part of the operational research literature
has been devoted to exploring non-normal demand patterns.
Nevertheless, there are inconsistencies in definitions of nonnormal demand patterns. For example, Williams (1984) argues
that the variances of the number of orders, order sizes and
lead-times should be taken into account, whereas Gelders and
Van Looy (1978) define non-normal patterns by the magnitude
of demand over a calendar year.
A categorization framework has been developed (see
Figure 1), to link definitions of non-normal demand patterns
with factors that have been proposed in the literature. The
first column contains three of these factors. Of course, the
framework can be extended to take other factors into account,
such as the variance of lead-time. The second column defines
‘intermittent’ and ‘erratic’ demand, according to the factors
proposed in the forecasting-categorization of Syntetos et
al (2005). In the final column, ‘lumpy’ demand is defined
according to Syntetos et al (2005), while the ‘clumped’
category contains more regular demand patterns, including
fixed clump-sizes (Ritchie and Kingsman, 1985). The ‘slow’
category is defined according to demand volume. This allows
the ‘slow’ category to be the ‘C’ class in an A–B–C classification by volume. However, in the remainder of the paper,
the company’s usage is followed, where the ‘slow’ category
is restricted to intermittent demand patterns.
In summary, according to this framework:
• An intermittent demand item is an item with infrequent
demand occurrences.
• A slow moving item is an item whose average demand
per period is low. This may be due to infrequent demand
occurrences, low average demand sizes or both.
• An erratic demand item is an item whose demand size is
highly variable.
• A lumpy demand item is an intermittent item for which
demand, when it occurs, is highly variable.
• A clumped demand item is an intermittent item for which
demand, when it occurs, is constant (or almost constant).
The framework is conceptual, rather than operational, since
‘high’ and ‘low’ are not quantified. The conceptual framework was used to guide discussions with the company and
experimentation with their software.
Case study: old method of classification
The empirical investigation presented here is based on the
categorization scheme employed by the forecasting and
stock control software manufacturer, as shown pictorially in
Figure 2. (The nomenclature differs slightly from Figure 1
since an item must be non-intermittent to be ‘erratic’, and
intermittent to be ‘slow’.) The criteria used in this categorization scheme reflect, according to the manufacturer, an
approach that is not untypical within this specialized software
sector.
In the old system, the number of zero demand time
periods (r ), during the last n time periods (n = 13), is
High
Intermittent
Mean inter demand interval
Non-intermittent
Low
Mean demand size
Low
AND/OR
Slow
High
Erratic
Lumpy
AND
Coefficient of variation
of demand sizes
Non-erratic
Low
Figure 1
Clumped
AND
Categorization of non-normal demand patterns.
JE Boylan et al—Classification for forecasting and stock control
Forecasting
475
Stock control
Average demand
Low
Slow
Lumpy
Fast
Erratic
Intermittent
Order frequency
Non-intermittent
High
Low
High
Variability of demand
Figure 2
A software package categorization scheme.
compared against a corresponding break-point to determine
intermittence. (A break-point of 7 has been used traditionally.
If r 7, then demand is classified as intermittent; if r < 7, then
demand is non-intermittent.) In this paper, the term ‘breakpoint’ will be used to denote a critical value of any parameter
that determines the category of an SKU.
The average demand per unit time period (over the last
13 periods) is used as a criterion for further distinguishing
between lumpy and slow-moving SKUs in the intermittent
demand category. (The average demand break-point is set to
0.5.) If demand has been classified as non-intermittent, the
variability of the demand per period is compared against a
specified break-point to further distinguish between fast and
erratic SKUs.
Intermittent demand requirements are estimated using a
13-period simple moving average (SMA (13)). Nonintermittent demand is estimated using various forms of
exponential smoothing, depending on the trend and seasonal
characteristics of the demand series. In the next stage, stock
control applications are selected based on a sub-categorization
scheme. For slow demand, the Poisson distribution is used
to calculate the order quantity. For lumpy demand, a proprietary empirical method is used instead. For non-intermittent
demand, continuous order level or periodic order-up-to-level
methods are used depending on the client’s requirements (with
demand standard deviation adjustments for the erratic SKUs).
Literature review
Williams (1984) analysed the demand data of a public utility
and proposed a method of demand pattern categorization
based on an idea called variance partition. According to
this method, the variance of the demand during lead time is
split into its constituent parts: variance of the order sizes,
transaction variability and variance of the lead times. The
categorization scheme meets various theoretical and practical
requirements that were proposed in Williams’ work.
Nevertheless, the break-points were chosen arbitrarily, so
that they make sense only for the particular situation that was
analysed in the 1984 paper. This raises some doubts about the
potential applicability of the proposed categorization scheme
in other contexts. Eaves and Kingsman (2004) analysed
demand data from the Royal Air Force and concluded that
Williams’ classification scheme did not adequately describe
the observed demand structure. Consequently, a revised classification scheme was proposed (for more details, see Eaves,
2002). The break-point values assigned to the categorization
parameters were chosen subjectively, based on management decisions, thereby restricting the generalization of their
results. Of course, the production of universally applicable
rules was not the objective of either academic paper. In both
papers, a case-study-based solution was developed.
In both categorization schemes discussed above, forecasting methods and stock control models are selected once
the demand patterns have been categorized. This is also true
for the software solution summarized in the previous section.
Johnston and Boylan (1996) re-conceptualized the term
‘intermittence’ by considering the mean inter-demand intervals for which Croston’s (1972) method, specifically designed
for intermittence, outperformed single exponential smoothing
(SES). The authors recommended a rule that if the mean interdemand interval ( p) is greater than 1.25 forecast revision
periods, then Croston’s method should be used, rather than
SES (based on the simulated mean squared error (MSE)). This
form of rule is based on the premise that it is preferable to
identify conditions for superior forecasting performance, and
then to categorize demand based on these results, rather than
the other way round. The essence of the re-conceptualization
lies in this approach and the identification of the mean interdemand interval as a categorization parameter, rather than the
specification of an exact break-point value. Indeed, it seems
more logical to work in the following way: (i) compare
alternative estimation procedures, (ii) identify the regions of
superior performance for each one of them and (iii) define
Journal of the Operational Research Society Vol. 59, No. 4
476
Low
p = 1.32(cut-off value)
High
High
Erratic
(Syntetos & Boylan)
Lumpy
(Syntetos & Boylan)
CV = 0.49
2
(cut-off value)
‘Smooth’
(Croston)
Intermittent
(Syntetos & Boylan)
Low
Figure 3 Demand-based categorization for forecasting (after
Syntetos et al, 2005).
the demand patterns based on the methods’ comparative
performance, rather than arbitrarily defining demand patterns
and then testing which estimation procedure performs best
on each particular demand category.
Johnston and Boylan (1996) compared SES and Croston’s
method, but did not question the validity of Croston’s method
itself. Croston’s (1972) method works in the following way:
estimates of mean demand sizes and intervals are updated
using SES, after a demand occurrence, and the ratio of the size
to interval estimates is used as an estimate of the mean demand
per unit time period. Croston’s estimator was found by
Syntetos and Boylan (2001) to be biased. The Syntetos–
Boylan approximation (SBA) (Syntetos and Boylan, 2005,
2006a) adjusts the biased estimates by applying a multiplicative factor of (1 − /2) to the Croston estimate, where is
the smoothing constant used to update estimates of the mean
inter-demand interval.
Syntetos et al (2005) used the three-stage approach
discussed above to develop a theoretically coherent demand
categorization scheme based on the mean inter-demand
interval ( p) and the squared coefficient of variation of
demand sizes (CV2 ). The definition of the alternative demand
patterns resulted from a direct comparison between the theoretical MSE performances of Croston’s method, SES and the
SBA. Both the parameters and their break-point values were
the outcome of this formal comparison, and the validity of
the scheme was empirically tested and confirmed on 3000
SKUs from the automotive industry. The contribution of this
work lies in the identification of the CV2 as an additional
categorization parameter for demand forecasting purposes.
Nevertheless, inventory control issues were not addressed.
The scheme is presented pictorially in Figure 3.
Purpose of empirical investigation
The old classification scheme used by the software manufacturer was found to have some limitations, in the light of the
theoretical considerations summarized in the previous section.
The categorization of ‘intermittent’ and ‘non-intermittent’
demand was based on the number of zero demand time periods
(r ), during the last n time periods (n = 13). This is a sound
approach, being equivalent to the mean inter-demand interval
recommended in the literature. However, the break-point of
r = 7 was determined subjectively. Therefore, the first objective of the empirical analysis is to compare the forecasting and
stock-control implications of different choices of the breakpoint for the number of zeroes in the last 13 periods. To
conduct this analysis, the following estimators are considered:
SES and SMA (13) for ‘non-intermittent’ demand; Croston’s
method and the SBA for ‘intermittent demand’.
The inclusion of CV2 as a further parameter to distinguish
intermittent demand was discussed with the company. They
were interested in its inclusion at a later stage of development
but did not wish to change too many aspects of their system
at the first stage. Consequently, the CV2 parameter was not
included as a forecasting classification parameter in this study.
In the old system, the categorization of ‘slow’ and ‘lumpy’
demand was based on the average demand. This does not
appear to capture the essence of ‘lumpiness’, as described in
the conceptual framework of Figure 1. This argument was
accepted by the company, since it was known that, in the
old system, low-volume high-variation items had insufficient
stock and high-volume low-variation items had excess stock.
Two categorization variables were discussed, namely the ratio
of variance to mean of demand and the squared coefficient
of variation of demand size (CV2 ). The first measure is an
index of dispersion for the Poisson distribution. The second
measure was preferred by the company, since it is dimensionless and would allow the same variable to be used in distinguishing ‘intermittent’ from ‘non-intermittent’ and ‘slow’ and
‘lumpy’, if CV2 is included in the former categorization at a
later stage of development. Therefore, the second objective of
the empirical analysis is to assess the stock-control implications for slow and lumpy demand, if new forecasting methods
are used and the company replaces average demand by CV2
as the categorization parameter.
In the following sections, the forecasting implications of the
‘intermittent’/‘non-intermittent’ categorization and the stockcontrol implications of the ‘slow’/‘lumpy’ categorization are
summarized, in accordance with the design of the software
package (see Figure 2).
Empirical investigation: forecasting
Summary characteristics of the demand data sets used in this
research are presented in the following table.
The demand data series have been divided into three
blocks: (i) initialization, (ii) calibration and (iii) performance
measurement. The lengths of these blocks are shown in the
final column of Table 1. The ‘initialization block’ is used
to initialize values required for methods based on recursive formulae (such as the mean inter-demand interval for
Croston’s method). In the ‘calibration block’, the optimal
JE Boylan et al—Classification for forecasting and stock control
Table 1
477
Empirical demand data sets
SKUs
Periods
Time bucket
Period breakdown
Motor spare parts
Aerospace spare parts
Chemical products
12 881
3025
53
26
52
60
Monthly
Bi-monthly
Monthly
13/7/6
18/18/16
24/24/12
smoothing constants are identified, based on MSE. Finally,
the optimal smoothing constants are used to update forecasts
in the ‘performance block’, in which performance statistics
are calculated.
The three data sets do not contain items categorized by the
company as ‘obsolete’, ‘dying’ or ‘new’. The data sets contain
only those SKUs with greater than or equal to two zeroes and
less than or equal to eleven zeroes in the last thirteen periods,
excluding the performance block. Also, the data sets do not
include SKUs with strictly less than seven zeroes in the last 13
(excluding the performance period) that were found to have
a 13-period global trend that was statistically significant (5%
level, two-tailed test). The restriction to a minimum of two
non-zero values ensures that all initial values can be calculated by the end of the calibration block, thereby allowing the
recursive formulae to proceed in the performance block. The
other restrictions were in keeping with the broadest range of
definitions of intermittence that the software company found
acceptable.
The forecast error statistics used for reporting results are:
the geometric root mean squared error (GRMSE) and the
average mean absolute error (MAE). As its name implies, the
GRMSE is based on squared errors and taking the appropriate
geometric mean as the summary measure. This summarization may be across time, across series or both. In this study,
the following approach is adopted: (i) calculate mean square
forecast error for each individual series (SKU) and then (ii)
multiply these values together for all n series and take the 2nth
root. This ensures that, when comparing two methods, the
effect of high errors for outlying SKUs cancel out. The properties of the GRMSE measure are discussed further by Boylan
and Syntetos (2006). The MAE averages errors, ignoring their
sign; it is summarized across series by its mean value. For
both measures, GRMSE and MAE, performance is measured
on all points in time and issue points only. The former corresponds to a re-order interval and the latter to a re-order level
system.
As discussed in the section on the purpose of the empirical
investigation, the estimators that are considered for the intermittent demand category (ie demand patterns associated with
a ‘low’ order frequency as determined by the break-point)
are Croston’s method and the SBA. For the non-intermittent
demand category (ie demand patterns with a ‘high’ order
frequency), SES and the SMA (13) are considered. Every
combination of those methods, for each possible break-point,
is tested. In Figure 4, the effect of break-point values is shown
for every pair of the estimators examined. These GRMSE-
Geometric Root Mean Square Error (GRMSE)
Industry
0.86
CR-SMA
CR-SES
SBA-SMA
SBA-SES
0.84
0.82
0.8
0.78
0.76
0.74
0.72
2
3
4
5
6
7
8
9
10
11
Break point (number of zeroes)
Figure 4 Categorization break-points for intermittent demand
(Data set 1).
based results were generated from the first data set, using
issue points only.
Results were also generated for the two smaller data sets
and using the mean MAE. The best break-points varied
slightly by data set and by error measure.
The recommended ranges of break-points are presented in
the following table. The ranges are consistent with the plateaus
shown in Figure 4 for Data set 1, but were recommended on
the basis of their consistency in yielding accurate forecasts
across all data sets and error measures.
Results were also generated for ‘all points in time’.
However, in this case, the results were consistent neither
across data sets nor, in some cases, across error measures.
This is an important element of the research agenda for
demand categorization.
The results in Table 2, for issue points only, indicate the
sensitivity of break-points to the choice of estimators. The
recommended range of break-points relating to the comparison of SES and the Syntetos–Boylan Approximation (SBA) is
consistent with theoretical expectations (Syntetos et al, 2005).
This is not true for the comparison of SES and Croston’s
estimator, in which case lower break-points for mean interdemand intervals were expected (Johnston and Boylan, 1996).
Nevertheless, consideration of the shapes of the curves in
Figure 4 results in an important operational conclusion, not
revealed in previous theoretical investigations. The curves are
478
Journal of the Operational Research Society Vol. 59, No. 4
Table 2
Suggested rules for identifying intermittence (Data sets 1, 2 and 3)
Intermittent
1.
2.
3.
4.
Non-intermittent
Recommended range of
break-point values
(number of zeroes)
Equivalent range of
mean inter-demand
intervals
SMA
SES
SMA
SES
5–7
6–8
2–3
2–4
1.63–2.17
1.86–2.60
1.18–1.30
1.18–1.44
Croston
Croston
SBA
SBA
quite flat for break-points from r = 2 to r = 6 showing that
forecast accuracy is insensitive to the choice of break-point in
this region. In the region from r = 7 (the old company breakpoint) to r = 13, forecast accuracy is highly sensitive to the
break-point value. This conclusion is confirmed by analysis
of mean MAE, where similar results are obtained.
This analysis strengthened the company’s resolve to change
this parameter from being hard-coded to being menu-driven
and to recommend r = 3 as an initial setting. This breakpoint consistently yielded accuracy that was close to the best,
for ‘issue points’, even when the optimal break-point was
somewhat higher. For ‘all points in time’, the setting of r = 3
performed well for Data set 1, but less so for Data sets 2 and
3. The need to customize this setting, based on client data,
remained for ‘all points in time’.
Empirical investigation: stock control
In the next stage of this investigation, stock control related
issues were considered. A demand forecasting system, such
as the one discussed in the previous section, may perform
well in terms of a given accuracy measure, but this does not
necessarily translate into excellent stock-control performance.
Simulation design and performance criteria
At this stage, the results were based mainly on one data
set (Data set 1, motor spare parts). The chemical data set
(Data set 3) was not used at all for stock simulations because
inventory-related data could not be made available. Moreover,
the aerospace spare parts data set (Data set 2) was of limited
use because, for many SKUs, the demand status switched very
frequently across different categories over time, leaving only
a small number (fewer than 30) that were consistently defined
as slow or lumpy. This also affected Data set 1, although to a
much lesser degree. Overall, many SKUs had to be discarded
in order to examine the characteristics of slow and lumpy
data. However, this enabled a more systematic investigation
on the remaining series.
A positive trend of about 6% over 26 months was detected
at the aggregate level for the first data set. This trend was not
detectable at SKU level, but it is possible that it may have
affected the results.
For the purposes of this research, only the intermittent
demand category was evaluated, as improvements in that area
constituted the main objective of the software manufacturer.
The software features the continuous re-order point, order
quantity (s, Q) system, the periodic order-up-to-level (T, S)
system and the periodic order point order-up-to-level (T, s, S)
system. (For a discussion of these systems, see Silver et al,
1998.) In this research, detailed results were obtained only for
an (s, Q) system, in which orders of quantity Q are placed
when the stock falls below the re-order point (s). The simulations were not replicated on the other control systems since
preliminary results indicated no significant differences. This
conclusion is also supported by the findings of the empirical study conducted by Sani and Kingsman (1997) on the
combined performance of statistical estimators and periodic
stock control models.
The order-quantity, Q, was determined by the cumulative
forecast over the lead-time, and s by using an appropriate
distribution. For slow items, a Poisson distribution is assumed.
This is a natural choice for slow movers and reflected the
company’s existing practice. For lumpy items, the proprietary method developed by the company was investigated, in
addition to calculations based on the negative binomial
distribution (NBD). The NBD is included since it satisfies
both theoretical and empirical criteria (Syntetos and Boylan,
2006b). The forecasting methods considered, for both slow
and lumpy items, were: SBA and SMA(13). SMA(13) has
been included as it is the estimator currently employed in
the system. Croston’s method and SES were excluded on the
basis of a further forecasting analysis, the details of which
are beyond the scope of this paper.
The following measures were recorded: (i) individual
SKU time average stock-holding (in units and in value); (ii)
average of (i) over all SKUs; (iii) Customer service level
(CSL) achieved. The ‘CSL’ is defined as the ratio of the
fulfilled demand (total demand minus backorders) to the total
demand, calculated in terms of units. Default options for the
simulation experiment included: backorders carried forward
(ie no lost sales); time series are treated as demand rather
than sales; optimization of forecasting parameters over lead
time is based on the MSE.
To conduct this analysis, ‘intermittent demand’ was defined
as those SKUs with three or more zero demand periods over
the last 13 periods, following the outcome of the forecasting
investigation reported earlier. To distinguish between slow and
lumpy demand, the squared coefficient of variation of demand
size (CV2 ) was used. The evaluation of break-points, which
clients would use in practice, was problematic. Clients operate
JE Boylan et al—Classification for forecasting and stock control
Estimator
SMA(13)
SBA
Resulting inventory for slow demand-proposed scheme
(Data set 1)
Avg. stock
volume/units
Total stock
value (index)
CSL achieved (%)
1.454
1.325
100.0
88.3
96.75
93.37
differential service-level targets for slow and lumpy items,
and so the choice of break-point would largely depend on
service-mix requirements. In Data set 1, though, almost 50%
of the SKUs would be classified as slow, regardless of client
preferences, since these items had zero variance of demand
size. Classification of this subset of SKUs as slow, and the
remainder as lumpy, would provide a useful benchmark for
future analyses, taking client service-mix requirements into
account.
98.0
96.0
94.0
92.0
90.0
88.0
93.0
94.0
95.0
96.0
97.0
Target Customer Service Level %
Stock control implications of forecasting for slow items
SMA(13)
SBA
Target Service Level
Figure 5 Target and achieved service level—proposed scheme
(Data set 1).
Slow demand
1350
1300
Total Stock Volume (units)
The main objective of this analysis was to extend the
forecasting results to a stock control context, rather than
to conduct a detailed investigation of alternative standard
statistical demand distributions, as this issue has been often
addressed in the literature. The target CSL was initially set
to a commonly employed target, 95%, but was later treated
as a simulation parameter, with variation from 93 to 97%.
The SMA(13)–SBA comparison results (95% target) are
shown in Table 3. The figures that relate to the total inventory
value are presented as index numbers because the relevant
information is confidential.
Under the proposed scheme, the resulting stock from the
application of the SBA method is 8.9% less than that associated with SMA(13). Such a decrease in the volume of
stock translates to an 11.7% reduction in the total inventory
value. Nevertheless, the inventory-related savings occur at the
expense of a reduction in the achieved CSL, which drops
from 96.75 to 93.37%. These results conform to theoretical
expectations. Achieving a CSL higher than the target was
expected for SMA(13), owing to the bias (over-forecast) associated with this method after an issue-point ((LS, Q) system).
Similarly, some under-achievement of the target for the SBA
can be attributed to the fact that it slightly under-forecasts the
mean demand. The service level target was then varied, using
the same definitions of demand categories, and the same forecasting methods. This enabled the effect on the achieved CSL
to be observed (Figure 5).
Figure 5 shows that SMA consistently attains a service
level above the target. Although this is desirable, it is at the
expense of high inventory levels (see Figure 6). The SBA
attains service levels that are close to targets above 95%. For
lower targets, there is a wider discrepancy. The effect of target
service levels on inventory is shown in Figure 6.
Figure 6 shows marked savings in inventory arising from
the application of the SBA for all service level targets. This is
Slow demand
100.0
Achieved Customer Service Level %
Table 3
479
1250
1200
1150
1100
1050
1000
950
900
850
93.0
94.0
95.0
96.0
97.0
Target Customer Service Level %
SMA(13)
Figure 6 Target service level and
(volume)—proposed scheme (Data set 1).
SBA
resulting
inventory
consistent with theoretical expectations and with the empirical findings of Eaves and Kingsman (2004). The software
company was greatly encouraged by these results for slow
items. Their view was that clients would prefer small undershoots to overshoots of the target CSL if large inventory
savings could be achieved.
480
Journal of the Operational Research Society Vol. 59, No. 4
Stock control implications of forecasting for lumpy items
Conclusions
The lumpy demand category is evaluated in terms of the
impact of forecasting methods and of subsequent calculations
for stock-control. The target CSL for this demand category is
70%. All combinations of methods examined failed to achieve
this target. In terms of the subsequent calculations, the possibility of using the NBD instead of the current empirically
driven approach was assessed, as discussed earlier. Comparison results between the use of NBD and the currently used
procedure indicated a very similar performance.
The philosophy of categorization by comparison of
methods depends upon having good methods for each of
the categories. In this evaluation, the broadest definition of
lumpy demand was used, including SKUs with very modest
variability in demand size. Nevertheless, it was not possible
to attain the target service level or to get close to it. In the
absence of a good approach for lumpy demand, it is not
possible to determine the appropriateness of the companypreferred classification variable (squared coefficient of variation of demand size) in a stock-control context.
Since it was not possible to improve on the current proprietary method, no changes were recommended to the system,
although the NBD may be made available as an alternative
for lumpy demand.
The development of demand categorization schemes has not
received as much academic attention as it deserves. Demand
categorization rules dictate the forecasting and stock control
methods to be used for different SKUs. Consequently, these
rules have significant implications in terms of stock and
customer satisfaction.
In the empirical investigation, it has been found that the
company’s old break-point for intermittence (seven or more
zeroes in the last 13) was not the best and that, for forecasts
after ‘issue points’, a parameter value between two and six
would be better. Within this range, forecast accuracy is not
sensitive to the exact break-point value. The company were
persuaded by a priori arguments that the average demand
should not be used to distinguish slow from lumpy demand.
To replace this parameter, they preferred the squared coefficient of variation of demand size, since it is dimensionless
and would fit in with potential developments in categorization of intermittence. Inventory assessments showed substantial savings by using the SBA for slow demand, with a slight
undershoot of the target CSL. The software manufacturer has
changed from hard-coding to a menu-driven approach for classification according to the number of zeroes, and dropped the
old rule of at least seven zeroes out of 13. Although more work
remains to be carried out on lumpy demand, the company
has also introduced the CV2 parameter to distinguish between
slow and lumpy demand, accompanied by facilities for client
experimentation.
Research agenda
In this paper, the effect of demand categorization on both forecasting and stock-control has been evaluated empirically. In
previous research, based on forecasting only, it was found that
categorization schemes should take into account the number
of zero demand periods as well as the coefficient of variation of demand size (Syntetos et al, 2005). In this research,
the number of zero demand periods has been confirmed as
an effective categorization parameter for forecasting, at least
for forecasts after an issue-point. The results show forecast
accuracy to be relatively insensitive to the exact choice of the
break-point value for the number of zero periods. The results
on forecasts for ‘all points in time’ were less conclusive, and
further research is required to assess the sensitivity of forecast
accuracy to categorization methods in this case. The interaction of forecasts at ‘all points in time’ with re-order interval
systems also requires detailed examination.
The results for lumpy data show that, for all combinations
of methods, the target CSLs were not achieved. Further work
in this area should include a wider selection of forecasting and
inventory approaches and analysis of alternative categorization variables, such as the ratio of variance to mean of demand.
An overall optimization of forecasting and stock-control
categories has not been considered. It is important to undertake such an analysis, to understand the effect of overall
optimization on inventory system performance. Moreover,
extension of this work to other inventory methods would
allow the interaction between forecasting and stock-control
methods to be understood more fully.
Acknowledgements — We acknowledge financial support for this project
from the company involved and the DTI. The empirical findings of
the paper emerged from a Knowledge Transfer Partnership between the
company and Buckinghamshire Chilterns University College. Also, we
thank the participants in the Intelligent Management Systems in Operations
(IMSIO) III conference (Salford, June 28–29, 2005) for their comments
on an earlier draft of this paper.
References
Boylan JE and Syntetos AA (2006). Accuracy and accuracyimplication metrics for intermittent demand. Foresight: Int J Appl
Forecasting 4: 39–42.
Croston JD (1972). Forecasting and stock control for intermittent
demands. Opl Res Q 23: 289–304.
Eaves AHC (2002). Forecasting for the ordering and stock holding
of consumable spare parts. PhD thesis: Lancaster University.
Eaves AHC and Kingsman BG (2004). Forecasting for the ordering
and stock-holding of spare parts. J Opl Res Soc 55: 431–437.
Gardner ES (1990). Evaluating forecast performance in an inventory
control system. Mgmt Sci 36: 490–499.
Gelders LF and Van Looy PM (1978). An inventory policy for slow
and fast movers in a petrochemical plant: a case study. J Opl Res
Soc 29: 867–874.
Johnston FR and Boylan JE (1996). Forecasting for items with
intermittent demand. J Opl Res Soc 47: 113–121.
Kobbacy KAH and Liang Y (1999). Towards the development of an
intelligent inventory management system. Integrated Manuf Syst
10: 354–366.
JE Boylan et al—Classification for forecasting and stock control
Ritchie E and Kingsman BG (1985). Setting stock levels for
wholesaling: Performance measures and conflict of objectives
between supplier and stockist. Eur J Opl Res 20: 17–24.
Sani B and Kingsman BG (1997). Selecting the best periodic inventory
control and demand forecasting methods for low demand items.
J Opl Res Soc 48: 700–713.
Silver EA, Pyke DF and Peterson R (1998). Inventory Management
and Production Planning and Scheduling, 3rd ed. John Wiley &
Sons: New York.
Strijbosch LWG, Heuts RMJ and van der Schoot EHM (2000). A
combined forecast-inventory control procedure for spare parts.
J Opl Res Soc 51: 1184–1192.
Syntetos AA and Boylan JE (2001). On the bias of intermittent demand
estimates. Int J Production Econ 71: 457–466.
Syntetos AA and Boylan JE (2005). The accuracy of intermittent
demand estimates. Int J Forecasting 21: 303–314.
View publication stats
481
Syntetos AA and Boylan JE (2006a). Comments on the attribution of
an intermittent demand estimator. Int J Forecasting 22: p. 201.
Syntetos AA and Boylan JE (2006b). On the stock-control
performance of intermittent demand estimators. Int J Production
Econ 103: 36–47.
Syntetos AA, Boylan JE and Croston JD (2005). On the categorization
of demand patterns. J Opl Res Soc 56: 495–503.
Watson RB (1987). The effects of demand–forecast fluctuations on
customer service and inventory cost when demand is lumpy. J Opl
Res Soc 38: 75–82.
Williams TM (1984). Stock control with sporadic and slow-moving
demand. J Opl Res Soc 35: 939–948.
Received October 2005;
accepted July 2006 after one revision
Download