Project Analytics
© Dan Trietsch
June 2019
Reference
Kenneth R. Baker and Dan Trietsch (2019),
Principles of Sequencing and Scheduling,
2nd edition, Wiley. (Ch. 18 and Appendix
A, with earlier work cited within.)
Summary
• PERT is useless because it is based on
an invalid assumption that the triplet
method works, AND that activity times
are statistically independent.
• A meritorious recommendation—to
calibrate—has been practiced in the
breach, AND would not be sufficient.
Summary
• We’ll validate that the lognormal
distribution works for projects.
• By cross-validation, we’ll show that
linear association—perhaps the simplest
model of statistical dependence—works
for projects, and makes possible
prediction for future projects.
Summary
• While doing that, we will show that
subjective estimates are inferior, and
that the independence assumption is
indefensible.
• And we will show that calibration helps
but not enough: it downplays the effects
of between-projects variation.
Summary
• Among many other future opportunities
now open, first and foremost is software
development.
Summary
• Mainly for theoreticians, we need
dedicated software for specialized
statistical tests. That will support further
research in Project Analytics.
Summary
• For practitioners, I believe we have
enough results already to support the
incorporation of software for Project
Analytics within project management
DSSs.
Summary
• Some of the practically-oriented software
can also support further research,
especially in partitioning.
Preliminary Notes
This talk is describes the state of the art of
Project Analytics. As such, it is also an
invitation to push said state of the art
forward.
Preliminary Notes
As the title suggests, we’ll focus on analytics.
But the context is Lognormal Scheduling for
PERT 21. Lognormal Scheduling is Safe
Scheduling under the assumption processing
times are lognormal. Safe Scheduling takes
safety time into account explicitly. (If I could,
I would have called it “Robust Scheduling,”
but that term had been hi-jacked by then.)
Preliminary Notes
Largely because it incorporates Lognormal
Scheduling, PERT 21 is a version of PERT
that does NOT rely on the beta distribution (as
well as other invalid assumptions), considers
restricted capacity (by sequencing), and
provides optimal safety time (by scheduling
release dates).
Preliminary Notes
Think about it:
NOBODY EVER VALIDATED PERT
And incidentally, nobody ever validated CC
(or the so called “Theory” of Constraints)
either.
If I have time, I’ll show that they are both
invalid.
Preliminary Notes
The next two slides are the title and the first
slide of a talk I presented at Tel Aviv
University and at the Technion (Haifa) in
2005. A revised version was presented at
Leuven in 2007 (with the same two slides
at its top).
Stochastic Economic Balance
Principles for Projects Scheduling
Feeding buffers, Crashing and
Sequencing
Stochastic Economic Balance
• Any practicable and robust project scheduling
and control framework should:
• (1) be based on sound theory;
• (2) yield economical results;
• (3) be intuitively acceptable to decision
makers;
• (4) require modest information inputs from
users and provide easy-to-utilize outputs.
Preliminary Notes
In particular, a stochastic ‘engine’ must satisfy
(1) and (4): Be valid and easy to use. But an
engine can be ‘easy to use’ and yet powerful:
we can and should employ powerful models
within DSSs. (3)—intuitive acceptance—does
not require that the user be able to do the job
without (possibly complex) software.
Preliminary Notes
In particular, a stochastic ‘engine’ must satisfy
(1) and (4): Be valid and easy to use.
The simplest possible input requirement is a
point estimate for each activity (as in CC). We
use only that plus relevant history. (Such
history is always available, even if it is
apparently missing. Ask me later.)
Preliminary Notes
Simplicity aside, one can think about PERT
21 as a version of PERT/CPM that requires
validated theory. The validation is what
Project Analytics is all about. That is, PERT
21 would not exist without Project
Analytics.
Some general observations
To elaborate, the need to avoid GIGO
(garbage in, garbage out) applies to any
operations research model, not just to
simulation models.
Some general observations
By definition, using a model without valid
data is invalid. But for some reason this
message has typically been ignored or
practiced in the breach.
Some general observations
Where money is clearly involved, valid
data is more likely to be used, to wit,
Financial Engineering. But are we there
with respect to project scheduling and
budgeting?
Some general observations
Project Analytics is necessary because the
answer is, unfortunately, no! So the crux is
to validate assumptions and obtain valid
data for project scheduling and budgeting.
(At present, however, we have much more
evidence regarding scheduling.)
Some general observations
Validation is akin to empirical science: you
start with a hypothesis (theory) and apply
trial and error.
This talk covers the major developments so
far, including (but not limited to) work at
American University of Armenia and at
Ghent University.
Some general observations
Another point of view:
Models are never perfect:
(1) they do not fit reality perfectly
(including imprecision and the use of
substitute objectives), and
(2) they may not have tractable analytic
solutions.
Some general observations
When we write a mathematical paper we
can ignore the fit issue, and focus on
achieving the best possible analytic
solution. Regarding fit, all we have to do is
assume it is good.
Some general observations
However, scheduling is an engineering
subject. At least if should be treated as an
engineering subject. (And even if you don’t
agree, that’s our assumption today.)
Some general observations
Stereotypically, mathematicians require
analytic solutions but may be willing to
accept unrealistic models. Engineers accept
approximate solutions but require realistic
models. A good engineering model
achieves the right balance between fit and
solution quality.
Some general observations
Project Analytics involves approximations
but for more realistic models than those
published before.
Some history
As a practicing engineer, long ago, I had
access to project scheduling data but no
motivation to use it. Later, I became an
academic and I acquired the motivation, but
no longer had access! Field data is in short
supply for academics (unless their
consulting is relevant to their research).
Some history
Then I had the opportunity to supervise a
masters student—Lilit Gevorgyan—who
had access to two project managers. For her
thesis, she collected data on 5 + 9 projects,
and her task was to check whether the
lognormal distribution would provide a
good fit.
Some history
But Lilit decided to use a statistical
package to find the best fit. Four or five
distributions looked okay. The lognormal
was second or third in the list. Accordingly,
she suggested we consider the first one in
the list first.
Some history
Nonetheless, I stuck to the lognormal.
Why? Because it was the only one that
made sense theoretically (which is why it
was hypothesized to fit in the first place).
Some history
Letting a small sample steer you away from
a hypothesis even though it does not
contradict it is the wrong approach. Of
course one can check new hypotheses in
future. And if the hypothesis would be
rejected our next step should be to
construct a new or corrected hypothesis.
Some history
To elaborate, as taught by Deming, we
must start with a hypothesis. Otherwise,
random data will drive us all over the place.
Until a hypothesis is rejected, there is no
compelling need to look for a different one.
Some history
Regarding the lognormal hypothesis in
particular, a decade later, it has not yet been
rejected (although corrections in the model
did take place). Nobody claims that it is
necessarily the only valid distribution we
could use, just that it is valid. It also ticks
several theoretical boxes….
Theory/Hypothesis
The most important engineering deviations
from plan are not additive, but proportional.
Theory/Hypothesis
With that in mind, if a positive deviation
may be 100%, that does not mean that a
negative deviation is –100%.
Ergo, don’t say 0.0 to 2.0, say 0.5 to 2.0.
Theory/Hypothesis
For instance, suppose our estimate is 0.5 but
we err by ‘–100%,’ then the true value is
0.5 x 2.0 = 1.
Likewise, if the estimate is 2.0 but we err
by ‘+100%,’ in which case 2.0 x 0.5 = 1
Theory/Hypothesis
Suppose we have an expected error of 0.
Then our range is from 1 to 1 (that is, [1–0]
to [1+0]). By contrast, if our expected error
is ∞ then our range is from ‘1/∞’ = 0 to ∞;
so 0 corresponds to –∞, 1 corresponds to 0,
and ∞ corresponds to ∞.
Theory/Hypothesis
Translating {0, 1, ∞} to {–∞, 0, ∞} ‘begs’
for the logarithmic transformation!
(Of course, some of us are so ‘addicted’ to
the additive model that the fact that
logarithms of products are additive would
be sufficient to convince us that this
particular transformation is the one.)
Theory/Hypothesis
IF the post-transformation distribution is
normal, then the original distribution must
be lognormal. So that’s the hypothesis we
start with.
Theory/Hypothesis
Hypothesis: Activity Times and processing
times follow the lognormal distribution.
Theory/Hypothesis
Hypothesis: Activity Times and processing
times follow the lognormal distribution.
And because deviations are proportional,
the lognormal distribution would apply to
ratios.
Theory/Hypothesis
The lognormal is especially attractive for
four reasons at least: (1) it is strictly
positive (w.p.1), (2) its cv is not restricted,
(3) it can approximate sums of positive
random variables (using the Lognormal
Sum Approximation), and (4) it can
represent the relationship between capacity
and activity time (for stochastic crashing).
46
Lognormal Scheduling
0
0.4
0.8
1.2
1.6
2
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
x
f(x)
exp
sigma = 0.25
sigma = 0.5
sigma = 1
sigma = 2
Drilling times in minutes (Clyde Dam, NZ)
0
20
40
60
80
100
120
140
160
180
0 25 55 85 115 145 175 205 235 280
Some Skewness Implications
The last column indicates the probability of
falling below the mode. Compare it to the third
column. And note: SL > 0.5 does not guarantee
that the mean will be supported!
Theory/Hypothesis
BTW, the ubiquitous variance statistic is
implicitly much more suited to the additive
world than it is for the multiplicative world.
Theory/Hypothesis
For instance, an erroneous (but frequent)
assumption is that a deviation that exceeds
3σ is out of control (As it would be for the
normal r.v.). As a result, an “in control”
deviation of the core normal r.v. within the
lognormal can often be misinterpreted as an
“out of control” signal! To continue,
Theory/Hypothesis
Concerning ratios, if pj denotes processing
time of activity j and ej is the actual time,
we hypothesize that ln(pj/ej) is normal.
Theory/Hypothesis
But first it is highly recommended to make
sure that activities are of similar planned
duration, with a ratio of 4-5 at most. That
may necessitate a partition.
The following example is from a
construction project in Armenia, and the
first step is to partition activities by size:
ln(pj/ej) as a function of planned duration
-1.2
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
1 2 4 8 16 32 64
On Rounding
Three parts are required. As a rule, each is then
analyzed separately.
In particular, there are 61 activities (or ‘points’)
with planned durations between 4 and 61, but
only 23 are visible. That implies that there are
multiple points that ‘cover’ each other.
This could not happen without rounding. But the
question is what to do about it.
On Rounding
The next slide shows a Q-Q chart where such
duplicate points receive consecutive (but
different) Blom scores. (Blom’s Scores were
developed for the normal distribution. The kth
smallest sample value receives the score
zk = Φ−1[(k − 0.375) / (n + 0.25)],
where n is the sample size.)
Q-Q chart of 61 activities with planned durations
between 4 and 61
y = 0.3718x - 0.1619
R² = 0.976
-1.2
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
-3 -2 -1 0 1 2 3
On Rounding
Let m and s be the mean and the standard
deviation of lnX, where X is a lognormal random
variable. In such a Q-Q chart, the intercept
provides an estimate of m and the slope estimates
s; in particular, above we have m = −0.1609 and
s = 0.3718. R = 0.9879 (the square root of
0.976). R can be compared to tabulated values to
test for normality. Indeed, normality cannot be
rejected here: the p-value is almost 0.25 (> 0.05).
On Rounding
The left hand side of the following slide is
identical to the previous slide, but the right hand
side shows a Q-Q chart where such duplicate
points receive an appropriate average score
instead; that is, we take the average of all
duplicated points as the score that applies to all
of them.
Q-Q charts before and after correcting for rounding
y = 0.3718x - 0.1619
R² = 0.976
-1.2
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
-3 -2 -1 0 1 2 3
y = 0.3768x - 0.1619
R² = 0.9892
-1.2
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
-3 -2 -1 0 1 2 3
On Rounding
In this case, the p-value increases to slightly over
0.5, but we didn’t need that to avoid erroneous
rejection. Often, however, this simple rounding
correction spells the difference between
accepting and rejecting the hypothesis that the
distribution is normal.
On Rounding
Royston (1993) proposed averaging the indices
to calculate a single Blom’s score for them with
the result replacing k. However, experimentation
revealed that the previous way yields slightly
higher R2 values. On the one hand, my sample
was small, so it may have been a fluke. On the
other hand, I checked because I suspected (i.e.,
hypothesized) that would be the case in advance.
The Parkinson Distribution
Parkinson’s Law states that work expands to fill
out the time allotted to it (q). Empirically,
however, it may just be that work that finishes
early is REPORTED on time. Therefore, the time
we OBSERVE, X is:
X = max{q, Y}, where q is the allotted time and
Y is the real underlying random variable.
The Parkinson Distribution
We introduced this distribution in Baker and
Trietsch (2009), Principles of Sequencing and
Scheduling. But empirical evidence collected
later and reported in Trietsch et al. (2012),
suggests that sometimes some early points are
reported correctly, whereas others are reported
‘on time.’
The Parkinson Distribution
If some early points are reported correctly,
whereas others are reported ‘on time,’ then the
general form of the Parkinson distribution
assumes early points are reported on-time with a
probability PP, and reported correctly otherwise.
The Parkinson Distribution
If some early points are reported correctly,
whereas others are reported ‘on time,’ then the
general form of the Parkinson distribution
assumes early points are reported on-time with a
probability PP, and reported correctly otherwise.
How would that look on a Q-Q chart?
Eurasia Foundation data, w/ and w/o rounding
y = 0.5664x + 0.2985
R² = 0.9105
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
2.5
-4 -2 0 2 4
y = 0.5961x + 0.2985
R² = 0.9581
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
2.5
-4 -2 0 2 4
Now assume that 9 points are Parkinsonian
y = 0.7469x + 0.1474
R² = 0.9665
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
2.5
-2 0 2 4
y = 0.7592x + 0.1436
R² = 0.9823
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
2.5
-2 0 2 4
The Parkinson Distribution
Compare the parameters, such as slope
(higher, why?), the horizontal range
(indicating Blom’s scores, an issue we’ll
return to), and R2 (higher, why?).
The Parkinson Distribution
BTW, the late Eliyahu Goldratt, a highly
intuitive thinker, opined that what we ascribe
to the Parkinson effect is actually due to what
he called the “Student’s Syndrome.” But if
that were the case the tails would not likely
have been that straight (because, likely, we’d
have been dealing with a mixture of
distributions based on the allowed lead time).
The Parkinson Distribution
Altogether, there are 15 points reported on-time,
and yet the figure assumes only 9 Parkinsonian
points. The issue is that when a point is reported
on-time, we can’t tell if it finished approximately
on-time and then rounded (so it should be treated
as ‘true’) or finished early and then reported on-
time. The figure shows the case where 6 points
are assumed to be approximately on-time. But
how do we decide that it should be 9 + 6?
The Parkinson Distribution
But how do we decide that it should be 9 + 6?
First, notice that we cannot prove that any
number we might select is correct. Validation
only shows that the lognormal Parkinson
distribution can work with 9 Parkinsonian
points. It cannot prove that no other number
could work too and what is the correct number.
The Parkinson Distribution
First, notice that we cannot prove that any
number we might select is correct.
Selecting the number of Parkinsonian points is
based on an obvious heuristic: minimize SEY (the
regression standard error).
The Parkinson Distribution
Selecting the number of Parkinsonian points is
based on an obvious heuristic: minimize SEY (the
regression standard error).
Notice that this is not equivalent to maximizing
R2. (It does so approximately, however.)
The Parkinson Distribution
Notice that this is not equivalent to maximizing
R2. (It does so approximately, however.)
For multiple regression, minimizing SEY
maximizes the adjusted-R2. For that reason,
initially I thought it was equivalent. However,
Jordy Batselier and Mario Vanhoucke
demonstrated to me I was wrong.)
The Parkinson Distribution
When we use the SEY–minimization heuristic to
decide how many points should be considered
Parkinsonian, however, we also have to adjust
the slope of the early points accordingly. That is
why the horizontal scale was different on the left.
However, we will revisit this issue once again,
still later.
The Parkinson Distribution
It is certainly possible that the reason the
lognormal has not been validated for projects
earlier is the combination of rounding and the
Parkinson effect: both mask the true distribution.
The Parkinson Distribution
It is certainly possible that the reason the
lognormal has not been validated for projects
earlier is the combination of rounding and the
Parkinson effect: both mask the true distribution.
To wit, Trietsch et al. (2012) tested a handful of
project data collected from various sources and
all of them yielded acceptable Q-Q charts once
rounding and the Parkinson effect were
considered. (Without the SEY heuristic, BTW.)
The Parkinson Distribution
Colin and Vanhoucke (2015) tested the results of
Trietsch et al. (2012) on a dataset collected by
Batselier and Vanhoucke (2015), and in some
instances lognormality was not supported.
Resolving those is our next topic.
Mixtures and Partitioning
We have already mentioned partitioning briefly
when we discussed the need to keep the ratio
between planned durations within a ratio range
of up to 4 or 5 (that is, the maximum divided by
the minimum should not exceed 4 (preferably) or
5 (if necessary).
Mixtures and Partition
It turned out that although Trietsch et al. (2012)
briefly mentioned the need for partition to meet
the ratio requirement between the maximal and
minimal planned duration, they (i.e., we) failed
to actually do it! As a result, Colin and
Vanhoucke (2015)—who followed the same
procedure—did not do it either.
Mixtures and Partition
Indeed, analysis of the offending instances in
Colin and Vanhoucke (2015) revealed that they
involved outliers or mixtures. Most of them were
resolved by the basic partition. Few required
removal of clear outliers (always a potential
issue—not everything under the sun is in
statistical control).
Mixtures and Partitioning
Sometimes, however, just ensuring a ratio of up
to 4 or 5 is not enough. For instance, in
construction, some activities are sensitive to the
weather and others are not. Such instances
should not be addressed as outliers. Instead, we
have to find a way to partition, either in advance
based on project information or during post-
mortem analysis (to learn lessons for the future).
Project C2014-03, ej = 1, with the on time activities
included and after correcting for the Parkinson effect
Project C2014-03, ej = 1, but partitioned instead of
assuming the Parkinson effect
Project Analytics
Mixtures and Partitioning
The current state of the art for post-mortem
partition is essentially by the SEY heuristic. The
idea is that if we pool all points together, SEY
will be excessive. If so, it is highly likely that we
can find at least one point whose removal will
reduce SEY. Now just repeat until no such
reductions are possible. The remaining points
form one part. Next, repeat from the start for all
rejected points. (Usually, there is no need to
repeat more than once.)
Mixtures and Partitioning
Recall that we postponed detailed discussion of
the way the SEY heuristic helps decide how many
points should be considered Parkinsonian. For
this purpose we only consider candidate points
with a ratio of actual to planned of unity. But as
we consider whether to call one more point
Parkinsonian (i.e., ‘remove’ it) we must also
adjust the slope of the remaining early points.
That is achieved by adjusting their Blom scores.
Mixtures and Partitioning
As we consider whether to call one more point
Parkinsonian (i.e., ‘remove’ it) we must also
adjust the slope of the remaining early points.
That is achieved by adjusting their Blom scores.
Essentially, suppose there are nP Parkinsonian
points and nE points reported early. Then we
estimate PP = nP / (nP + nE).
Mixtures and Partitioning
Essentially, suppose there are nP Parkinsonian
points and nE points reported early. Then we
estimate PP = nP / (nP + nE).
Because PP is the fraction of early activities that
cannot be used in the Q-Q chart (their true
duration is unknown), we modify Blom’s scores
for the remaining early activities. We use zk =
Φ−1((k − 0.375)/(n(1 – pP) + 0.25)).
Mixtures and Partitioning
Because PP is the fraction of early activities that
cannot be used in the Q-Q chart (their true
duration is unknown), we modify Blom’s scores
for the remaining early activities. We use zk =
Φ−1((k − 0.375)/(n(1 – pP) + 0.25)).
The adjustment reflects the effective reduction in
sample size that applies to early activities. But
we only use it for the nE strictly early activities.
Mixtures and Partitioning
Again, we use adjusted scores of
zk = Φ−1((k − 0.375)/(n(1 – pP) + 0.25))
for the nE strictly early activities. We use
unadjusted scores,
zk = Φ−1((k − 0.375)/(n + 0.25))
for all on-time and ‘rounded-to-one’ activities.
That leaves out the nP Parkinsonian observations,
which are simply omitted from the chart.
Linear Association and Prediction
Whereas mixtures is an issue only recently
addressed and not yet even properly published,
linear association—which is essentially an
intentionally simple way to model statistical
dependence—has been at the core of the earlier
developments (started at American University of
Armenia).
On Subjective Estimates
The Hill et al. dataset shows linear association,
and demonstrates the above result as well.
Linear Association and Prediction
Linear association—or some other validated
dependence model—is ESSENTIAL for
prediction. So in a sense it is the most important
part of Project Analytics.
It is essential because ignoring dependence leads
to a very high probability of missing deadlines.
Linear Association and Prediction
As these projects demonstrate, each project has
its own mean deviation from plan (inc. budget)
Linear Association and Prediction
Because it’s only 12 projects—a small number as
Q-Q charts go—in both of these normality
cannot be rejected. But more importantly, each
point here represents the sum total of a full
project, and if we had independence than such
sums would exhibit low variation. Here it
appears that the hypothesis all project have the
same mean is rejectable. That can be more
formally shown by ANOVA.
Linear Association and Prediction
In other words, each project has a different bias.
For project k, denote that bias by bk. Let B denote
the random variable from which bk is sampled.
Linear Association and Prediction
Suppose that a set of random variables is defined
by Yj = XjB > 0 for a set of independent
nonnegative Xj and an independent positive
common factor B. Then we say that the members
of the set {Yj} are linearly associated.
Linear Association and Prediction
If we assume linear association, then to simulate
a new project we can generate a value for bk as
well as realizations xj for logarithmic ratios and
then obtain linearly associated realizations yj = xj
bk. If we repeat this in r rows (where r should be
a large number, e.g., 1000), then the variation of
B will be taken into account, as well as the
variation of Xj. Let’s look at the result with or
without linear association.
Linear Association and Prediction
In this case we used nonparametric bootstrap
instead. It yields similar results, but is simpler.
Linear Association and Prediction
If, instead, we only calibrate the PERT estimates,
it’s better, but not good enough.

More Related Content

PDF
MUMS Opening Workshop - Panel Discussion: Inexact Computer Model Calibration,...
PDF
Pert techniques
PDF
Size-Based Disciplines for Job Scheduling in Data-Intensive Scalable Computin...
PPTX
Knowledge representation in AI
PDF
Is it important to explain a theorem? A case study in UML and ALCQI
PPTX
Final prescription vs prediction
PPTX
Understanding data dfm_1_yogi_schulz_2017_05
ODP
Story Points considered harmful – a new look at estimation techniques
MUMS Opening Workshop - Panel Discussion: Inexact Computer Model Calibration,...
Pert techniques
Size-Based Disciplines for Job Scheduling in Data-Intensive Scalable Computin...
Knowledge representation in AI
Is it important to explain a theorem? A case study in UML and ALCQI
Final prescription vs prediction
Understanding data dfm_1_yogi_schulz_2017_05
Story Points considered harmful – a new look at estimation techniques

Similar to Project Analytics (20)

PPTX
Prescriptive Analytics.pptx
PDF
Why are data transformations a bad choice in statistics
PDF
Taguchi and the tablecloth trick | Insight, issue 5
PDF
Free Lunch or No Free Lunch: That is not Just a Question?
PDF
Scientific Benchmarking of Parallel Computing Systems
PDF
ON AVERAGE CASE ANALYSIS THROUGH STATISTICAL BOUNDS : LINKING THEORY TO PRACTICE
PDF
On average case analysis through statistical bounds linking theory to practice
PDF
On average case analysis through statistical bounds linking theory to practice
DOCX
internship project1 report
PDF
Introductory Statistics for Engineering Experimentation 1st Edition Peter R. ...
PDF
Resource Allocation Using Metaheuristic Search
PPTX
Bps managing dissertation
PDF
Do we measure functional size or do we count thomas fehlmann
PDF
Weapons of Math Instruction: Evolving from Data0-Driven to Science-Driven
PPT
cs1538.ppt
PDF
Design thinking unpacked: an evolutionary algorithm?
PDF
Production operations management
PDF
Introductory Statistics for Engineering Experimentation 1st Edition Peter R. ...
PPTX
Statistics in the age of data science, issues you can not ignore
PDF
The potential role of ai in the minimisation and mitigation of project delay
Prescriptive Analytics.pptx
Why are data transformations a bad choice in statistics
Taguchi and the tablecloth trick | Insight, issue 5
Free Lunch or No Free Lunch: That is not Just a Question?
Scientific Benchmarking of Parallel Computing Systems
ON AVERAGE CASE ANALYSIS THROUGH STATISTICAL BOUNDS : LINKING THEORY TO PRACTICE
On average case analysis through statistical bounds linking theory to practice
On average case analysis through statistical bounds linking theory to practice
internship project1 report
Introductory Statistics for Engineering Experimentation 1st Edition Peter R. ...
Resource Allocation Using Metaheuristic Search
Bps managing dissertation
Do we measure functional size or do we count thomas fehlmann
Weapons of Math Instruction: Evolving from Data0-Driven to Science-Driven
cs1538.ppt
Design thinking unpacked: an evolutionary algorithm?
Production operations management
Introductory Statistics for Engineering Experimentation 1st Edition Peter R. ...
Statistics in the age of data science, issues you can not ignore
The potential role of ai in the minimisation and mitigation of project delay
Ad

Recently uploaded (20)

PPTX
statsppt this is statistics ppt for giving knowledge about this topic
PPTX
DS-40-Pre-Engagement and Kickoff deck - v8.0.pptx
PDF
OneRead_20250728_1808.pdfhdhddhshahwhwwjjaaja
PDF
Best Data Science Professional Certificates in the USA | IABAC
PDF
Navigating the Thai Supplements Landscape.pdf
PDF
Loose-Leaf for Auditing & Assurance Services A Systematic Approach 11th ed. E...
PPTX
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
PPTX
ai agent creaction with langgraph_presentation_
PPTX
CYBER SECURITY the Next Warefare Tactics
PDF
CS3352FOUNDATION OF DATA SCIENCE _1_MAterial.pdf
PPTX
SET 1 Compulsory MNH machine learning intro
PPT
statistics analysis - topic 3 - describing data visually
PPTX
chrmotography.pptx food anaylysis techni
PDF
Votre score augmente si vous choisissez une catégorie et que vous rédigez une...
PPTX
Statisticsccdxghbbnhhbvvvvvvvvvv. Dxcvvvhhbdzvbsdvvbbvv ccc
PPTX
recommendation Project PPT with details attached
PPT
lectureusjsjdhdsjjshdshshddhdhddhhd1.ppt
PPT
statistic analysis for study - data collection
PPTX
IMPACT OF LANDSLIDE.....................
PDF
Tetra Pak Index 2023 - The future of health and nutrition - Full report.pdf
statsppt this is statistics ppt for giving knowledge about this topic
DS-40-Pre-Engagement and Kickoff deck - v8.0.pptx
OneRead_20250728_1808.pdfhdhddhshahwhwwjjaaja
Best Data Science Professional Certificates in the USA | IABAC
Navigating the Thai Supplements Landscape.pdf
Loose-Leaf for Auditing & Assurance Services A Systematic Approach 11th ed. E...
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
ai agent creaction with langgraph_presentation_
CYBER SECURITY the Next Warefare Tactics
CS3352FOUNDATION OF DATA SCIENCE _1_MAterial.pdf
SET 1 Compulsory MNH machine learning intro
statistics analysis - topic 3 - describing data visually
chrmotography.pptx food anaylysis techni
Votre score augmente si vous choisissez une catégorie et que vous rédigez une...
Statisticsccdxghbbnhhbvvvvvvvvvv. Dxcvvvhhbdzvbsdvvbbvv ccc
recommendation Project PPT with details attached
lectureusjsjdhdsjjshdshshddhdhddhhd1.ppt
statistic analysis for study - data collection
IMPACT OF LANDSLIDE.....................
Tetra Pak Index 2023 - The future of health and nutrition - Full report.pdf
Ad

Project Analytics

  • 1. Project Analytics © Dan Trietsch June 2019
  • 2. Reference Kenneth R. Baker and Dan Trietsch (2019), Principles of Sequencing and Scheduling, 2nd edition, Wiley. (Ch. 18 and Appendix A, with earlier work cited within.)
  • 3. Summary • PERT is useless because it is based on an invalid assumption that the triplet method works, AND that activity times are statistically independent. • A meritorious recommendation—to calibrate—has been practiced in the breach, AND would not be sufficient.
  • 4. Summary • We’ll validate that the lognormal distribution works for projects. • By cross-validation, we’ll show that linear association—perhaps the simplest model of statistical dependence—works for projects, and makes possible prediction for future projects.
  • 5. Summary • While doing that, we will show that subjective estimates are inferior, and that the independence assumption is indefensible. • And we will show that calibration helps but not enough: it downplays the effects of between-projects variation.
  • 6. Summary • Among many other future opportunities now open, first and foremost is software development.
  • 7. Summary • Mainly for theoreticians, we need dedicated software for specialized statistical tests. That will support further research in Project Analytics.
  • 8. Summary • For practitioners, I believe we have enough results already to support the incorporation of software for Project Analytics within project management DSSs.
  • 9. Summary • Some of the practically-oriented software can also support further research, especially in partitioning.
  • 10. Preliminary Notes This talk is describes the state of the art of Project Analytics. As such, it is also an invitation to push said state of the art forward.
  • 11. Preliminary Notes As the title suggests, we’ll focus on analytics. But the context is Lognormal Scheduling for PERT 21. Lognormal Scheduling is Safe Scheduling under the assumption processing times are lognormal. Safe Scheduling takes safety time into account explicitly. (If I could, I would have called it “Robust Scheduling,” but that term had been hi-jacked by then.)
  • 12. Preliminary Notes Largely because it incorporates Lognormal Scheduling, PERT 21 is a version of PERT that does NOT rely on the beta distribution (as well as other invalid assumptions), considers restricted capacity (by sequencing), and provides optimal safety time (by scheduling release dates).
  • 13. Preliminary Notes Think about it: NOBODY EVER VALIDATED PERT And incidentally, nobody ever validated CC (or the so called “Theory” of Constraints) either. If I have time, I’ll show that they are both invalid.
  • 14. Preliminary Notes The next two slides are the title and the first slide of a talk I presented at Tel Aviv University and at the Technion (Haifa) in 2005. A revised version was presented at Leuven in 2007 (with the same two slides at its top).
  • 15. Stochastic Economic Balance Principles for Projects Scheduling Feeding buffers, Crashing and Sequencing
  • 16. Stochastic Economic Balance • Any practicable and robust project scheduling and control framework should: • (1) be based on sound theory; • (2) yield economical results; • (3) be intuitively acceptable to decision makers; • (4) require modest information inputs from users and provide easy-to-utilize outputs.
  • 17. Preliminary Notes In particular, a stochastic ‘engine’ must satisfy (1) and (4): Be valid and easy to use. But an engine can be ‘easy to use’ and yet powerful: we can and should employ powerful models within DSSs. (3)—intuitive acceptance—does not require that the user be able to do the job without (possibly complex) software.
  • 18. Preliminary Notes In particular, a stochastic ‘engine’ must satisfy (1) and (4): Be valid and easy to use. The simplest possible input requirement is a point estimate for each activity (as in CC). We use only that plus relevant history. (Such history is always available, even if it is apparently missing. Ask me later.)
  • 19. Preliminary Notes Simplicity aside, one can think about PERT 21 as a version of PERT/CPM that requires validated theory. The validation is what Project Analytics is all about. That is, PERT 21 would not exist without Project Analytics.
  • 20. Some general observations To elaborate, the need to avoid GIGO (garbage in, garbage out) applies to any operations research model, not just to simulation models.
  • 21. Some general observations By definition, using a model without valid data is invalid. But for some reason this message has typically been ignored or practiced in the breach.
  • 22. Some general observations Where money is clearly involved, valid data is more likely to be used, to wit, Financial Engineering. But are we there with respect to project scheduling and budgeting?
  • 23. Some general observations Project Analytics is necessary because the answer is, unfortunately, no! So the crux is to validate assumptions and obtain valid data for project scheduling and budgeting. (At present, however, we have much more evidence regarding scheduling.)
  • 24. Some general observations Validation is akin to empirical science: you start with a hypothesis (theory) and apply trial and error. This talk covers the major developments so far, including (but not limited to) work at American University of Armenia and at Ghent University.
  • 25. Some general observations Another point of view: Models are never perfect: (1) they do not fit reality perfectly (including imprecision and the use of substitute objectives), and (2) they may not have tractable analytic solutions.
  • 26. Some general observations When we write a mathematical paper we can ignore the fit issue, and focus on achieving the best possible analytic solution. Regarding fit, all we have to do is assume it is good.
  • 27. Some general observations However, scheduling is an engineering subject. At least if should be treated as an engineering subject. (And even if you don’t agree, that’s our assumption today.)
  • 28. Some general observations Stereotypically, mathematicians require analytic solutions but may be willing to accept unrealistic models. Engineers accept approximate solutions but require realistic models. A good engineering model achieves the right balance between fit and solution quality.
  • 29. Some general observations Project Analytics involves approximations but for more realistic models than those published before.
  • 30. Some history As a practicing engineer, long ago, I had access to project scheduling data but no motivation to use it. Later, I became an academic and I acquired the motivation, but no longer had access! Field data is in short supply for academics (unless their consulting is relevant to their research).
  • 31. Some history Then I had the opportunity to supervise a masters student—Lilit Gevorgyan—who had access to two project managers. For her thesis, she collected data on 5 + 9 projects, and her task was to check whether the lognormal distribution would provide a good fit.
  • 32. Some history But Lilit decided to use a statistical package to find the best fit. Four or five distributions looked okay. The lognormal was second or third in the list. Accordingly, she suggested we consider the first one in the list first.
  • 33. Some history Nonetheless, I stuck to the lognormal. Why? Because it was the only one that made sense theoretically (which is why it was hypothesized to fit in the first place).
  • 34. Some history Letting a small sample steer you away from a hypothesis even though it does not contradict it is the wrong approach. Of course one can check new hypotheses in future. And if the hypothesis would be rejected our next step should be to construct a new or corrected hypothesis.
  • 35. Some history To elaborate, as taught by Deming, we must start with a hypothesis. Otherwise, random data will drive us all over the place. Until a hypothesis is rejected, there is no compelling need to look for a different one.
  • 36. Some history Regarding the lognormal hypothesis in particular, a decade later, it has not yet been rejected (although corrections in the model did take place). Nobody claims that it is necessarily the only valid distribution we could use, just that it is valid. It also ticks several theoretical boxes….
  • 37. Theory/Hypothesis The most important engineering deviations from plan are not additive, but proportional.
  • 38. Theory/Hypothesis With that in mind, if a positive deviation may be 100%, that does not mean that a negative deviation is –100%. Ergo, don’t say 0.0 to 2.0, say 0.5 to 2.0.
  • 39. Theory/Hypothesis For instance, suppose our estimate is 0.5 but we err by ‘–100%,’ then the true value is 0.5 x 2.0 = 1. Likewise, if the estimate is 2.0 but we err by ‘+100%,’ in which case 2.0 x 0.5 = 1
  • 40. Theory/Hypothesis Suppose we have an expected error of 0. Then our range is from 1 to 1 (that is, [1–0] to [1+0]). By contrast, if our expected error is ∞ then our range is from ‘1/∞’ = 0 to ∞; so 0 corresponds to –∞, 1 corresponds to 0, and ∞ corresponds to ∞.
  • 41. Theory/Hypothesis Translating {0, 1, ∞} to {–∞, 0, ∞} ‘begs’ for the logarithmic transformation! (Of course, some of us are so ‘addicted’ to the additive model that the fact that logarithms of products are additive would be sufficient to convince us that this particular transformation is the one.)
  • 42. Theory/Hypothesis IF the post-transformation distribution is normal, then the original distribution must be lognormal. So that’s the hypothesis we start with.
  • 43. Theory/Hypothesis Hypothesis: Activity Times and processing times follow the lognormal distribution.
  • 44. Theory/Hypothesis Hypothesis: Activity Times and processing times follow the lognormal distribution. And because deviations are proportional, the lognormal distribution would apply to ratios.
  • 45. Theory/Hypothesis The lognormal is especially attractive for four reasons at least: (1) it is strictly positive (w.p.1), (2) its cv is not restricted, (3) it can approximate sums of positive random variables (using the Lognormal Sum Approximation), and (4) it can represent the relationship between capacity and activity time (for stochastic crashing).
  • 46. 46 Lognormal Scheduling 0 0.4 0.8 1.2 1.6 2 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 x f(x) exp sigma = 0.25 sigma = 0.5 sigma = 1 sigma = 2
  • 47. Drilling times in minutes (Clyde Dam, NZ) 0 20 40 60 80 100 120 140 160 180 0 25 55 85 115 145 175 205 235 280
  • 48. Some Skewness Implications The last column indicates the probability of falling below the mode. Compare it to the third column. And note: SL > 0.5 does not guarantee that the mean will be supported!
  • 49. Theory/Hypothesis BTW, the ubiquitous variance statistic is implicitly much more suited to the additive world than it is for the multiplicative world.
  • 50. Theory/Hypothesis For instance, an erroneous (but frequent) assumption is that a deviation that exceeds 3σ is out of control (As it would be for the normal r.v.). As a result, an “in control” deviation of the core normal r.v. within the lognormal can often be misinterpreted as an “out of control” signal! To continue,
  • 51. Theory/Hypothesis Concerning ratios, if pj denotes processing time of activity j and ej is the actual time, we hypothesize that ln(pj/ej) is normal.
  • 52. Theory/Hypothesis But first it is highly recommended to make sure that activities are of similar planned duration, with a ratio of 4-5 at most. That may necessitate a partition. The following example is from a construction project in Armenia, and the first step is to partition activities by size:
  • 53. ln(pj/ej) as a function of planned duration -1.2 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 2 4 8 16 32 64
  • 54. On Rounding Three parts are required. As a rule, each is then analyzed separately. In particular, there are 61 activities (or ‘points’) with planned durations between 4 and 61, but only 23 are visible. That implies that there are multiple points that ‘cover’ each other. This could not happen without rounding. But the question is what to do about it.
  • 55. On Rounding The next slide shows a Q-Q chart where such duplicate points receive consecutive (but different) Blom scores. (Blom’s Scores were developed for the normal distribution. The kth smallest sample value receives the score zk = Φ−1[(k − 0.375) / (n + 0.25)], where n is the sample size.)
  • 56. Q-Q chart of 61 activities with planned durations between 4 and 61 y = 0.3718x - 0.1619 R² = 0.976 -1.2 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 -3 -2 -1 0 1 2 3
  • 57. On Rounding Let m and s be the mean and the standard deviation of lnX, where X is a lognormal random variable. In such a Q-Q chart, the intercept provides an estimate of m and the slope estimates s; in particular, above we have m = −0.1609 and s = 0.3718. R = 0.9879 (the square root of 0.976). R can be compared to tabulated values to test for normality. Indeed, normality cannot be rejected here: the p-value is almost 0.25 (> 0.05).
  • 58. On Rounding The left hand side of the following slide is identical to the previous slide, but the right hand side shows a Q-Q chart where such duplicate points receive an appropriate average score instead; that is, we take the average of all duplicated points as the score that applies to all of them.
  • 59. Q-Q charts before and after correcting for rounding y = 0.3718x - 0.1619 R² = 0.976 -1.2 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 -3 -2 -1 0 1 2 3 y = 0.3768x - 0.1619 R² = 0.9892 -1.2 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 -3 -2 -1 0 1 2 3
  • 60. On Rounding In this case, the p-value increases to slightly over 0.5, but we didn’t need that to avoid erroneous rejection. Often, however, this simple rounding correction spells the difference between accepting and rejecting the hypothesis that the distribution is normal.
  • 61. On Rounding Royston (1993) proposed averaging the indices to calculate a single Blom’s score for them with the result replacing k. However, experimentation revealed that the previous way yields slightly higher R2 values. On the one hand, my sample was small, so it may have been a fluke. On the other hand, I checked because I suspected (i.e., hypothesized) that would be the case in advance.
  • 62. The Parkinson Distribution Parkinson’s Law states that work expands to fill out the time allotted to it (q). Empirically, however, it may just be that work that finishes early is REPORTED on time. Therefore, the time we OBSERVE, X is: X = max{q, Y}, where q is the allotted time and Y is the real underlying random variable.
  • 63. The Parkinson Distribution We introduced this distribution in Baker and Trietsch (2009), Principles of Sequencing and Scheduling. But empirical evidence collected later and reported in Trietsch et al. (2012), suggests that sometimes some early points are reported correctly, whereas others are reported ‘on time.’
  • 64. The Parkinson Distribution If some early points are reported correctly, whereas others are reported ‘on time,’ then the general form of the Parkinson distribution assumes early points are reported on-time with a probability PP, and reported correctly otherwise.
  • 65. The Parkinson Distribution If some early points are reported correctly, whereas others are reported ‘on time,’ then the general form of the Parkinson distribution assumes early points are reported on-time with a probability PP, and reported correctly otherwise. How would that look on a Q-Q chart?
  • 66. Eurasia Foundation data, w/ and w/o rounding y = 0.5664x + 0.2985 R² = 0.9105 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.5 -4 -2 0 2 4 y = 0.5961x + 0.2985 R² = 0.9581 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.5 -4 -2 0 2 4
  • 67. Now assume that 9 points are Parkinsonian y = 0.7469x + 0.1474 R² = 0.9665 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.5 -2 0 2 4 y = 0.7592x + 0.1436 R² = 0.9823 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 2.5 -2 0 2 4
  • 68. The Parkinson Distribution Compare the parameters, such as slope (higher, why?), the horizontal range (indicating Blom’s scores, an issue we’ll return to), and R2 (higher, why?).
  • 69. The Parkinson Distribution BTW, the late Eliyahu Goldratt, a highly intuitive thinker, opined that what we ascribe to the Parkinson effect is actually due to what he called the “Student’s Syndrome.” But if that were the case the tails would not likely have been that straight (because, likely, we’d have been dealing with a mixture of distributions based on the allowed lead time).
  • 70. The Parkinson Distribution Altogether, there are 15 points reported on-time, and yet the figure assumes only 9 Parkinsonian points. The issue is that when a point is reported on-time, we can’t tell if it finished approximately on-time and then rounded (so it should be treated as ‘true’) or finished early and then reported on- time. The figure shows the case where 6 points are assumed to be approximately on-time. But how do we decide that it should be 9 + 6?
  • 71. The Parkinson Distribution But how do we decide that it should be 9 + 6? First, notice that we cannot prove that any number we might select is correct. Validation only shows that the lognormal Parkinson distribution can work with 9 Parkinsonian points. It cannot prove that no other number could work too and what is the correct number.
  • 72. The Parkinson Distribution First, notice that we cannot prove that any number we might select is correct. Selecting the number of Parkinsonian points is based on an obvious heuristic: minimize SEY (the regression standard error).
  • 73. The Parkinson Distribution Selecting the number of Parkinsonian points is based on an obvious heuristic: minimize SEY (the regression standard error). Notice that this is not equivalent to maximizing R2. (It does so approximately, however.)
  • 74. The Parkinson Distribution Notice that this is not equivalent to maximizing R2. (It does so approximately, however.) For multiple regression, minimizing SEY maximizes the adjusted-R2. For that reason, initially I thought it was equivalent. However, Jordy Batselier and Mario Vanhoucke demonstrated to me I was wrong.)
  • 75. The Parkinson Distribution When we use the SEY–minimization heuristic to decide how many points should be considered Parkinsonian, however, we also have to adjust the slope of the early points accordingly. That is why the horizontal scale was different on the left. However, we will revisit this issue once again, still later.
  • 76. The Parkinson Distribution It is certainly possible that the reason the lognormal has not been validated for projects earlier is the combination of rounding and the Parkinson effect: both mask the true distribution.
  • 77. The Parkinson Distribution It is certainly possible that the reason the lognormal has not been validated for projects earlier is the combination of rounding and the Parkinson effect: both mask the true distribution. To wit, Trietsch et al. (2012) tested a handful of project data collected from various sources and all of them yielded acceptable Q-Q charts once rounding and the Parkinson effect were considered. (Without the SEY heuristic, BTW.)
  • 78. The Parkinson Distribution Colin and Vanhoucke (2015) tested the results of Trietsch et al. (2012) on a dataset collected by Batselier and Vanhoucke (2015), and in some instances lognormality was not supported. Resolving those is our next topic.
  • 79. Mixtures and Partitioning We have already mentioned partitioning briefly when we discussed the need to keep the ratio between planned durations within a ratio range of up to 4 or 5 (that is, the maximum divided by the minimum should not exceed 4 (preferably) or 5 (if necessary).
  • 80. Mixtures and Partition It turned out that although Trietsch et al. (2012) briefly mentioned the need for partition to meet the ratio requirement between the maximal and minimal planned duration, they (i.e., we) failed to actually do it! As a result, Colin and Vanhoucke (2015)—who followed the same procedure—did not do it either.
  • 81. Mixtures and Partition Indeed, analysis of the offending instances in Colin and Vanhoucke (2015) revealed that they involved outliers or mixtures. Most of them were resolved by the basic partition. Few required removal of clear outliers (always a potential issue—not everything under the sun is in statistical control).
  • 82. Mixtures and Partitioning Sometimes, however, just ensuring a ratio of up to 4 or 5 is not enough. For instance, in construction, some activities are sensitive to the weather and others are not. Such instances should not be addressed as outliers. Instead, we have to find a way to partition, either in advance based on project information or during post- mortem analysis (to learn lessons for the future).
  • 83. Project C2014-03, ej = 1, with the on time activities included and after correcting for the Parkinson effect
  • 84. Project C2014-03, ej = 1, but partitioned instead of assuming the Parkinson effect
  • 86. Mixtures and Partitioning The current state of the art for post-mortem partition is essentially by the SEY heuristic. The idea is that if we pool all points together, SEY will be excessive. If so, it is highly likely that we can find at least one point whose removal will reduce SEY. Now just repeat until no such reductions are possible. The remaining points form one part. Next, repeat from the start for all rejected points. (Usually, there is no need to repeat more than once.)
  • 87. Mixtures and Partitioning Recall that we postponed detailed discussion of the way the SEY heuristic helps decide how many points should be considered Parkinsonian. For this purpose we only consider candidate points with a ratio of actual to planned of unity. But as we consider whether to call one more point Parkinsonian (i.e., ‘remove’ it) we must also adjust the slope of the remaining early points. That is achieved by adjusting their Blom scores.
  • 88. Mixtures and Partitioning As we consider whether to call one more point Parkinsonian (i.e., ‘remove’ it) we must also adjust the slope of the remaining early points. That is achieved by adjusting their Blom scores. Essentially, suppose there are nP Parkinsonian points and nE points reported early. Then we estimate PP = nP / (nP + nE).
  • 89. Mixtures and Partitioning Essentially, suppose there are nP Parkinsonian points and nE points reported early. Then we estimate PP = nP / (nP + nE). Because PP is the fraction of early activities that cannot be used in the Q-Q chart (their true duration is unknown), we modify Blom’s scores for the remaining early activities. We use zk = Φ−1((k − 0.375)/(n(1 – pP) + 0.25)).
  • 90. Mixtures and Partitioning Because PP is the fraction of early activities that cannot be used in the Q-Q chart (their true duration is unknown), we modify Blom’s scores for the remaining early activities. We use zk = Φ−1((k − 0.375)/(n(1 – pP) + 0.25)). The adjustment reflects the effective reduction in sample size that applies to early activities. But we only use it for the nE strictly early activities.
  • 91. Mixtures and Partitioning Again, we use adjusted scores of zk = Φ−1((k − 0.375)/(n(1 – pP) + 0.25)) for the nE strictly early activities. We use unadjusted scores, zk = Φ−1((k − 0.375)/(n + 0.25)) for all on-time and ‘rounded-to-one’ activities. That leaves out the nP Parkinsonian observations, which are simply omitted from the chart.
  • 92. Linear Association and Prediction Whereas mixtures is an issue only recently addressed and not yet even properly published, linear association—which is essentially an intentionally simple way to model statistical dependence—has been at the core of the earlier developments (started at American University of Armenia).
  • 93. On Subjective Estimates The Hill et al. dataset shows linear association, and demonstrates the above result as well.
  • 94. Linear Association and Prediction Linear association—or some other validated dependence model—is ESSENTIAL for prediction. So in a sense it is the most important part of Project Analytics. It is essential because ignoring dependence leads to a very high probability of missing deadlines.
  • 95. Linear Association and Prediction As these projects demonstrate, each project has its own mean deviation from plan (inc. budget)
  • 96. Linear Association and Prediction Because it’s only 12 projects—a small number as Q-Q charts go—in both of these normality cannot be rejected. But more importantly, each point here represents the sum total of a full project, and if we had independence than such sums would exhibit low variation. Here it appears that the hypothesis all project have the same mean is rejectable. That can be more formally shown by ANOVA.
  • 97. Linear Association and Prediction In other words, each project has a different bias. For project k, denote that bias by bk. Let B denote the random variable from which bk is sampled.
  • 98. Linear Association and Prediction Suppose that a set of random variables is defined by Yj = XjB > 0 for a set of independent nonnegative Xj and an independent positive common factor B. Then we say that the members of the set {Yj} are linearly associated.
  • 99. Linear Association and Prediction If we assume linear association, then to simulate a new project we can generate a value for bk as well as realizations xj for logarithmic ratios and then obtain linearly associated realizations yj = xj bk. If we repeat this in r rows (where r should be a large number, e.g., 1000), then the variation of B will be taken into account, as well as the variation of Xj. Let’s look at the result with or without linear association.
  • 100. Linear Association and Prediction In this case we used nonparametric bootstrap instead. It yields similar results, but is simpler.
  • 101. Linear Association and Prediction If, instead, we only calibrate the PERT estimates, it’s better, but not good enough.

Editor's Notes

  • #2: We will focus on Project Analytics for scheduling, but the technical analysis holds for budgeting as well.
  • #15: But they used to be yellow
  • #17: For example, CC and the PERT beta are unsound. But CC excels in (3) and (4). We must keep those in PERT 21. Project Analytics is the way to satisfy (1), (3), and (4)
  • #18: (3) Be intuitively acceptable to decision makers
  • #26: Examples: Using the wrong distribution leads to imprecision (and that’s our focus today). Minimizing maximal regret involves a substitute objective, but more benign examples involve the use of easy to measure substitutes; e.g., minimizing turnover to maximize job satisfaction. Or take Safe Scheduling: ignoring safety time is a quintessential example of lack of fit.
  • #27: Examples: Minimizing maximal regret.
  • #29: Examples of approximations: Fenton-Wilkinson and Linear Association. In the wider PERT 21 context, add the use of stored samples as well as sequencing heuristics. Note: Solution quality includes precision, accuracy, AND tractability.
  • #30: Examples of approximations: Fenton-Wilkinson and Linear Association. In the wider PERT 21 context, add the use of stored samples as well as sequencing heuristics.
  • #31: My practical experience was as an irrigation engineer. Most of my consulting experience was about SMED, with a focus on BNs.
  • #34: As taught by Deming, we must start with a hypothesis or random data will drive us all over the place. Until a hypothesis is rejected, there is no compelling need to look for a different one. Regarding the lognormal hypothesis in particular, a decade later, it has not yet been rejected (although corrections in the model did take place). NOBODY claims that it is necessarily the ONLY valid distribution we could use, just that it IS valid and ticks many more theoretical boxes.
  • #35: As taught by Deming, we must start with a hypothesis or random data will drive us all over the place. Until a hypothesis is rejected, there is no compelling need to look for a different one. Regarding the lognormal hypothesis in particular, a decade later, it has not yet been rejected (although corrections in the model did take place). NOBODY claims that it is necessarily the ONLY valid distribution we could use, just that it IS valid and ticks many more theoretical boxes.
  • #46: The lognormal sum approximation matches the mean and the variance of the sum. When the components are lognormal, it is aka the Fenton-Wilkinson approximation. Conclusion: The LN is good for products AND for sums!
  • #47: The figure depicts lognormal PDFs with mean = 1. We call them “basic.” Neither their modes nor their medians match 1, however. In red, we show the exponential, for comparison. The “blue” lognormal approximates the exponential to within 5% or so. The purple approximates the normal. The black goes with a higher standard deviation of 2, and it is visibly “sharper” than the exponential.
  • #48: Based on data from Premachandra and Gonzales (1996) from the Clyde Dam in NZ. Drilling time is in minutes. Both the beta and the lognormal would work here (but for high cv the PERT-beta would fail).
  • #54: Figure 18.1: Construction project in Armenia. Note, each point may represent several activities with the same planned duration and the same actual duration. In particular, all points on the X-axis have a REPORTED deviation of zero (that is, their actual duration, as REPORTED, matches their planned duration). Quite a few of them represent more than one activity. Negative values imply earliness whereas positive ones imply tardiness. NOTE: There are 61 activities between 4 and 15, but only 23 are visible. In the next slide we focus on those 61 points.
  • #57: Figure 18.2: Construction project in Armenia. There are 61 activities with planned durations between 4 and 15. On the left, we can see all 61 of them (but only 23 on the right). Even the left side, where values with the same ln(pj/ej) correspond to different Blom scores, the regression line has R2 = 0.976. Taking the square root, 0.988, we find that it passes the normality test with a probability of almost 0.25. That is, we cannot reject normality. Four points corresponding to the ratio 12/7 (for which ln(pj/ej) = 0.539) and at the bottom, two points corresponding to the ratio 1/3 (for which ln(pj/ej) = −1.099). In between, most ln(pj/ej) values appear more than once, and in particular, the ratios 8/7, 5/6, 4/5, and 1/2 each appear at least five times.
  • #60: Figure 18.2: Construction project in Armenia. There are 61 activities with planned durations between 4 and 15. On the left, we can see all 61 of them (but only 23 on the right). Even the left side, where values with the same ln(pj/ej) correspond to different Blom scores, the regression line has R2 = 0.976. Taking the square root, 0.988, we find that it passes the normality test with a probability of almost 0.25. That is, we cannot reject normality. Four points corresponding to the ratio 12/7 (for which ln(pj/ej) = 0.539) and at the bottom, two points corresponding to the ratio 1/3 (for which ln(pj/ej) = −1.099). In between, most ln(pj/ej) values appear more than once, and in particular, the ratios 8/7, 5/6, 4/5, and 1/2 each appear at least five times.
  • #66: We should expect a horizontal segment representing the Parkinsonian points followed by a steep drop representing the early points that were reported correctly. Those are pushed to the left and thus matched with too low Blom scores. Rounding will not change that dramatically except that the horizontal segment will be compressed to a single point. We can also expect an inverted S shape.
  • #67: Figure 18.3: A development project from the Eurasia Foundation in Armenia. The project had 52 activities with planned durations between 4 and 48 weeks. Preliminary analysis suggests that we should analyze the 41 activities of up to 15 weeks separately. Of those, 15 are reported as precisely on time. The left-side chart fails the normality test with a probability of about 0.005. Rounding helps, as the right-side chart demonstrates, but the test remains marginal (a low pass). Note the inverted S shape still visible on the right.
  • #68: Figure 18.4: Now assume 9 Parkinsonian points (so 6 points rounded to 1). Although we don’t have an automated test yet, clearly both sides are better than the previous RHS. Note shifted scale (due to the Parkinson correction of the early points, which now represent an effectively smaller sample). EXPLAIN WHY WE CHOSE PRECISELY 9 PARKINSONIAN POINTS. Note that we did not remove the inverted S shape completely but it’s much better (and it’s just a visual cue, not the way we decide how many Parkinsonian points to assume).
  • #84: Figure 18.4: Now assume 9 Parkinsonian points (so 6 points rounded to 1). Although we don’t have an automated test yet, clearly both sides are better than the previous RHS. Note shifted scale (due to the Parkinson correction of the early points, which now represent an effectively smaller sample). EXPLAIN WHY WE CHOSE PRECISELY 9 PARKINSONIAN POINTS. Note that we did not remove the inverted S shape completely but it’s much better (and it’s just a visual cue, not the way we decide how many Parkinsonian points to assume).
  • #85: Figure 18.4: Now assume 9 Parkinsonian points (so 6 points rounded to 1). Although we don’t have an automated test yet, clearly both sides are better than the previous RHS. Note shifted scale (due to the Parkinson correction of the early points, which now represent an effectively smaller sample). EXPLAIN WHY WE CHOSE PRECISELY 9 PARKINSONIAN POINTS. Note that we did not remove the inverted S shape completely but it’s much better (and it’s just a visual cue, not the way we decide how many Parkinsonian points to assume).
  • #102: The remaining problem is that variance is underestimated, and therefore we tend to have too many points near the top or near the bottom. Those reflect projects that are either much shorter or much longer than average.