Download Applying Statistical Methods in Business and Industry

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Sales process engineering wikipedia , lookup

Industrial design wikipedia , lookup

Bayesian inference in marketing wikipedia , lookup

Product lifecycle wikipedia , lookup

Control chart wikipedia , lookup

Predictive engineering analytics wikipedia , lookup

Official statistics wikipedia , lookup

Transcript
Deliverable 8.7 Report of the state of the art in
Applying Statistical Methods in Business and
Industry
Mission Statement for the report

To promote the wider understanding and application of contemporary and
emerging statistical methods

For Students

For people working in business and industry worldwide

And for the professional development of statistical practitioners (statistical
practitioner is any person using statistical methods whether formally trained or
not),

To foster best practice for the benefit of business and industry
Readership
All members of ENBIS
Contents
1. Introduction to the report, to ENBIS and pro-ENBIS
2. Industrial statistics - history and background
3. Statistical Consultancy
4. From Process improvement to Quality by design
5. Management statistics
6. Service quality
7. Risk, Finance and Insurance
8. Applying data mining methods to business and industry
9. Process Monitoring, Improvement and Control
10. Measurement system analysis
11. Design and analysis of experiments (DoE)
12. Safety and reliability engineering
13. Multivariate analysis focussing on multiscale modelling
14. Simulation
15. Communication
16. Summary, cross-fertilisation and future directions
Chapter 1
Introduction to the report, to ENBIS and pro-ENBIS
Shirley Coleman, Tony Fouweather and Dave Stewardson
The report was conceived as a deliverable in the original description of the proENBIS project proposal as a natural consequence of the three year working
partnership between many of the foremost players in this field. Contributions have
been volunteered from each of the work package leaders and from most of the other
project partners and members. Many of the contributions are collaborative efforts not
only ensuring that the subjects covered are dealt with from more than one perspective,
but also carrying on the initial drive of the project in promoting collaborations to
further the subject as a whole. The resulting report provides an overview of the state
of the art in business and industrial statistics in Europe.
The subject areas are dealt with in a logical manner, starting with a brief overview of
the history of industrial statistics, before tackling a variety of subjects which have
become vital to many areas of business and industry throughout the world. The
subjects and methods covered by this report are essential for any company wishing to
compete in an increasingly efficient market place.
With the current popularity of the six sigma movement, which seeks to implement
statistical methods and tools into businesses and industry, there is a growing need for
a comprehensive overview of the subject. The applications of these techniques are far
reaching, and can be extended to improve every aspect of the way a company is
managed. Often the use of these methods is the only reason that a business can gain
an edge over their competitors, and increasingly their use is crucial to the survival of
the business.
The report covers methodology and techniques currently in use and those about to be
launched and includes references to miscellaneous sources of information including
web-sites.
Such a collection of experts working together in partnership has been possible through
the support of the European Commission’s fifth framework growth programme and
the thematic network pro-ENBIS, and also via the recently established ENBIS webbased society.
For completeness, this introductory chapter gives a brief description of ENBIS
followed by an overview of the pro-ENBIS project.
In August 1999 a group of about 20 statisticians and statistical practitioners met in
Linköping, Sweden at the end of the First International Symposium on Industrial
Statistics to initiate creating a European Society for Industrial and Applied Statistics
(ESIAS). As a first step, a list of interested members was drawn up and dialogue
started. The internet was identified as a feasible platform to cheaply and efficiently
co-ordinate the activities of any body formed out of these meetings. A network of
people from industry and academia from all European nations interested in applying,
promoting and facilitating the use of statistics in business and industry was created to
address 2 main observations


As with many disciplines, applied statisticians and statistical practitioners
often work in professional environments where they are rather isolated from
interactions and stimulation from like minded professionals.
Statistics is vital for the economic and technical development and improved
competitiveness of European industry.
By February 2000 an executive committee was formed which held a founding
meeting at EURANDOM in Eindhoven, the Netherlands, on February 26th and 27th.
The name ENBIS (European Network for Business and Industrial Statistics) was
adopted as the definitive name for the society. It was decided to have a founding
conference on December 11, 2000 in Amsterdam and during that meeting, ENBIS
was formally launched. The conference was accompanied by a 3-day workshop on
design of experiments lead by Søren Bisgaard, a renowned European statistician.
The mission of ENBIS:
 Foster and facilitate the application and understanding of statistical methods to
the benefit of European business and industry,
 Provide a forum for the dynamic exchange of ideas and facilitate networking
among statistical practitioners (a statistical practitioner is any person using
statistical methods whether formally trained or not),
 Nurture interactions and professional development of statistical practitioners
regionally and internationally.
ENBIS has adopted the subsequent points as its vision.








To promote the widespread use of sound science driven, applied statistical
methods in European business and industry,
That membership consists primarily of statistical practitioners from business
and industry,
To emphasise multidisciplinary problem solving involving statistics,
To facilitate the rapid transfer of statistical methods and related technologies
to and from business and industry,
To link academic teaching and research in statistics with industrial and
business practice,
To facilitate and sponsor continuing professional development,
To keep its membership up to date in the field of statistics and related
technologies,
To seek collaborative agreements with related organisations.
ENBIS is a web based society and the activities can be found on www.enbis.org
ENBIS has so far arranged 5 annual conferences at various locations around Europe
which have allowed the showcasing of a broad spectrum of applications and generated
discussion about the use of statistics in a wide range of business and industrial areas.
ENBIS also plans to start a mid-year workshop meeting as well. Several ideas for new
projects are being developed and the group of partners and members continue to seek
ways to further their working relationships. Pro-ENBIS encouraged these
collaborations.
The thematic network, pro-ENBIS sought to build on the success of ENBIS and to
develop partnerships within Europe to support selected projects at the forefront of
industrial and business statistics, with the specific mission “to promote the widespread
use of sound science driven, applied statistical methods in European business and
industry.”
Pro-ENBIS was contracted for three years until 31st December 2004 by the European
Commission with a budget of about 800 000 Euros. The project was co-ordinated by
the University of Newcastle upon Tyne (UK) and had contractors and members from
across Europe.
The thematic network was funded so that it could achieve specific outcomes, such as
promoting industrial statistics through workshops, industrial visits and through the
publishing of papers and articles. These activities relate to the network’s aim to
provide a forum for the dissemination of industrial statistical methodology directly
from statisticians and practitioners to European industry.
The deliverables were grouped around statistical themes, with 8 work packages.
 WP1 Design of experiments
 WP2 Data mining/warehousing
 WP3 General statistical modelling, process modelling and control
 WP4 Reliability, safety and quality improvement
 WP5 Discovering European resources and expertise
 WP6 Drafting and initiating further activity
 WP7 Management statistics
 WP8 Network management
Outputs can be viewed on the project web-site http://www.enbis.org/pro-enbis/
Achievements and the future
 ENBIS Magazine within Scientific Computing World posted free to ENBIS
members every 2 months
 George Box medal for exceptional contributions to industrial statistics
established and awarded annually at ENBIS conference
 Prizes for best presentation, young statistician and supporting manager
awarded annually at ENBIS conference
 Establish local networks in most European countries.
 Have ENBIS members in the top 10 companies in each European country
 Continue Pro-ENBIS type workshop and research activities.
 Promote ENBIS membership widely
Presidents of ENBIS:
Henry Wynn (2000), Dave Stewardson (2001), Tony Greenfield (2002), Poul
Thyregod (2003), Shirley Coleman (2004), Fabrizio Ruggeri (2005).
Chapter 2
Industrial Statistics – History and Background
Jeroen de Mast, Ronald Does
Industrial statistics in a historical perspective
The quality situation pre 1980:
Before the industrial revolution, quality was assured by the guild system and
craftmanship. The Industrial revolution led to standardization, specifications and the
invention of management as discipline. There followed Taylorism / Fordism which
became the standard way of working in the West from 1910 till 1980. Mass
fabrication spread across Europe and the western world. Quality was seen as a tradeoff with productivity. There was an inward-perspective with the focus firmly on
efficiency and volume. The usual organization was command-and-control.
Early statistical contributions to industry:
The earliest well documented work was by Gosset during his time in the Guinness
brewery in Dublin. It was there that he proposed the t-test, which he famously
attributed to ‘Student’. Walter Shewhart developed control charts and statistical
process control during his time in munitions. He promoted statistics as a catalyst to
empirical learning. Dodge perfected inspection plans for use in production.
Statistical contributions post World War II:
George Box worked extensively on experimentation and response surface
methodology for process optimisation. W.Edwards Deming went to Japan and created
the movement towards company wide quality improvement.
Deming’s work ignited the Japanese revolution in which quality and productivity are
seen as synergetic. This was accompanied by a new outward-in perspective (what and
how to produce was not determined by a company’s tradition, technical specialists,
etc., but by what the customer needs and wants and how much he is willing to pay for
it). This led to decentralization and empowerment and the involvement of topmanagement. It was the start of a race for operational efficiency and effectiveness: the
Japanese have a manufacturing system (lean manufacturing) which is far superior to
the Taylor/Ford system. Western companies started to embark on a race to make up
the difference. This resulted in heavily increased competitiveness over issues like
operational efficiency, innovativeness and quality.
Statistical quality improvement
Amongst the early statistical quality improvement methodologies, two of the most
influential were the Taguchi method and the Shainin System. Taguchi has become a
by-word for experimental design and the sequential step-wise approach of Shainin has
been adopted by a number of major companies.
From a focus on statistical tools, currently the emphasis has moved on to statistical
thinking. This is exemplified by the notion that it is important that statistical tools and
inferential procedures are generally mastered by employees as members of
organisations.
Progress in the last 30 years was marked by a steady institutionalisation of approaches
to quality control and improvement, for example the ISO 9000 series. Professional
statisticians stand aside and watch the quality movement overemphasize the
qualitative aspects of quality management.
The quality improvement initiative Six Sigma was the next powerful influence in
these developments. Six sigma draws from previous innovations, for example
Taylor’s focus on efficiency, SPC’s use of data and statistical methods, Japan’s lean
manufacturing and aggressive defect reduction, customer focus and decentralized
project structure. But it is also a step forward, being the first programme to combine
statistical and non-statistical tools, a project methodology, an organizational structure
for projects, a metric system for performance management, and more aspects into a
single integrated programme. It is also innovative in prescribing a taskforce of
dedicated project leaders: the investments in time spent on process improvement in
Six Sigma companies are unprecedented.
Industrial statistics as a discipline
Historically there has always been a strong mathematical background to industrial
statistics, in particular the relationship with probability and mathematical statistics.
Statistics can be seen as a reconstruction of inferential procedures. Probabilistic
reasoning underpins much statistical problem solving.
Statistics has always had strong relationships with other disciplines: economics,
management science, philosophy of science. This is even more so the case with
industrial statistics which requires a genuine interest in the processes under
investigation. Industrial statistics has now emerged as a discipline in itself. It is
motivated by its own programme of development and progress including scientific
research of methodologies and analysis techniques.
References
Box GEP, Hunter WG and Hunter JS (1978), Statistics for Experimenters, Wiley,
New York.
Deming WE (1986) Out of the Crisis. MIT, Cambridge (MA).
Drucker PE (1954) The Practice of Management, Harper & Row, New York.
Fisher Box J (1987) “Guinness, Gosset, Fisher, and small samples” Statistical Science
2(1) 45–52.
Ishikawa K (1982) Guide to Quality Control. Asian Productivity Organization.
Juran JM (1989) Juran on Leadership for Quality: An Executive Handbook. Free
Press, New York.
Shewhart WA (1931) Economic Control of Quality of Manufactured Product. Van
Nostrand Reinhold, Princeton.
Student (WS Gosset) (1908), On the probable error of a mean, Biometrika 6, pp. 1-25.
Chapter 3
Statistical Consultancy
Ronald Does
The statistician in industry
The role of statisticians in industry has for many years been a source of anxiety to
statisticians. There are many papers about this topic, and many colleagues have
considerable experience as SPC and Six Sigma experts and statistics consultant.
Skills for a statistician working in consultancy
Caulcutt (2001) describes Six Sigma by listing the characteristics shared by the
organisations that have successfully adopted the Six Sigma approach. His list
includes “Black Belts” and he emphasises the important contribution that Black Belts
make in the pursuit of business improvement. There must be few, if any, companies
that have successfully pursued the objectives of Six Sigma without using employees
in the Black Belt role. Caulcutt’s paper focuses on Black Belts and describes what
they do and what skills they require if they are to perform this important role.
Statistical Consultancy units
Statistical consulting units can be successful and much thought has gone into how
they can be established to enhance their chances of being successful. They can be setup in Departments of Statistics & Mathematics, Business Studies and in other bodies.
Good examples exist of European statistical consulting units at universities.
Running a commercial consultancy bureau
As a contrast to being located within a university, it is possible to run a consultancy
unit on a commercial basis. Such a consultancy bureaus exist by providing statistical
services in exchange for financial remuneration.
Statistical Structuring
Statistical Consultancy is, by nature, based on the two fundamental disciplines of
applied statistics: ‘data analysis’ and ‘probability calculus’. But too often a statistical
consultant, especially in industry, reviewing a finished project is confused. Although
his/her contribution is recognized as a break-through added value, he/she probably
only used simple statistical methods. What made the contribution so valuable for the
customer? The real added value of a statistical consultant is his power to structure a
problem (one might also say, reality) in a specific way. This distinguishes our view on
problems from all other disciplines. This is what is meant by ‘Statistical Structuring’.
References
R. Caulcutt (2004), Black Belt types, Quality and Reliability Engineering International 20,
pp. 427-432.
R.J.M.M. Does & A. Trip (2001), The impact of statistics in industry and the role of
statisticians, Austrian Journal of Statistics 30, pp. 7-20.
R.J.M.M. Does & A. Zempléni (2001), Establishing a commercial statistical consulting unit at
universities, Kwantitatieve Methoden. 67, pp. 51-63.
Chapter 4
From Process Improvement to Quality by Design
Ron Kenett
Keywords: Improvement teams, process improvement, quality by design, robust
design, optimization, Six Sigma, DMAIC, Design for Six Sigma, problem solving,
new product development, quality ladder.
The Quality Ladder
Modern industrial organizations in manufacturing and services are subjected to
increased competitive pressures and rising customer expectations. Management teams
on all five continents are attempting to satisfy and even “delight” their customers while
simultaneously cutting costs. An increasing number of organizations have shown that
the apparent conflict between high productivity and high quality can be resolved
through improvements in work processes and quality of designs. The different
approaches to the management of industrial organizations can be summarized and
classified using a four step Quality Ladder. The four approaches are
1) Fire Fighting,
2) Inspection,
3) Process Control and
4) Quality by Design. To each management approach there corresponds a particular set
of statistical methods. Managers involved in reactive fire fighting need to be exposed to
basic statistical thinking. Managers attempting to contain quality and inefficiency
problems through inspection and 100% control can have their tasks alleviated by
implementing sampling techniques. More proactive managers investing in process
control and process improvement are well aware of the advantages of control chart and
process control procedures. At the top of the quality ladder is the quality by design
approach where up front investments are secured in order to run experiments designed
to impact product and process specifications. Efficient implementation of statistical
methods requires a proper match between management approach and statistical tools.
There are many well-known companies who have gone up the quality ladder and are
now enjoying increased efficiency and profitability.
Improvement teams
Revolutionary improvement of quality within an organization involves improvement
projects. A project by project improvement strategy relies on successful employee
participation in project identification, analysis and implementation. Factors relating
to a project's success or maximizing success are numerous, but are often overlooked
and probably unknown. A project by project quality improvement program must be
supported by quality principles, analysis techniques, effective leaders and facilitators,
and extensive training. It is feasible to identify specific factors that affect the success
or maximise the success of the quality improvement project team.
The Juran Trilogy and the Six Sigma DMAIC steps
It is feasible to review the Juran Trilogy of Improvement, Planning and Control and
map it into the DMAIC process.
The Taguchi approach and Design For Six Sigma
It is feasible to review the Taguchi approach of robust design and disucss its
implementation within the DFSS context.
Practical Statistical Efficiency
Practical Statistical Efficiency is an attempt to measure the full impact of statistical
tools on real life problems. We define Practical Statistical Efficiency (PSE) as:
PSE = E{R} x T {I} x P {I} x P {S} x V {PS} x V {P} x V {M} x V {D}
Where:
V{D} = value of the data actually collected.
V{M} = value of the statistical method employed.
V{P} = value of the problem to be solved.
V{PS} = value of the problem actually solved.
P{S} = probability that the problem actually gets solved.
P{I} = probability the solution is actually implemented.
T{I} = time the solution stays implemented.
E{R} = expected number of replications.
Conclusion
It can be shown how “going up the quality ladder” – from process improvement to
quality by design – can increase practical statistical efficiency.
References
Aubrey, C. A. and P. K. Felkins, Teamwork: Involving People in Quality and
Productivity Improvement, 1988, Quality Press.
Bickel, P. and K. Doksum, Mathematical Statistics: Basic Ideas and Selected
Topics, 1977, Holden-Day: San Francisco.
Chambers, P.R.G, J.L. Piggott and S.Y.Coleman, “SPC – a team effort for process
improvement across four area control centers”, J. Appl. Stats, 2001, 28,3: 307-324.
Coleman, S.Y and D.J. Stewardson “Use of measurement and charts to inform
management decisions”, Managerial Auditing Journal, 2002, 17,1/2: 16-19.
Coleman, S.Y., A. Gordon and P.R. Chambers, “SPC – making it work for the gas
transportation industry”, J. Appl. Stats, 2001, 28,3: 343-351.
Coleman, S.Y., G. Arunakumar, F. Foldvary and R. Feltham, “SPC as a tool for
creating a successful business measurement framework”, J. Appl. Stats, 2001, 28,3:
325-334.
Godfrey, A. Blanton, "Statistics, Quality and the Bottom Line," Part 1, ASQC
Statistics Division Newsletter, 1988, 9, 2, Winter: 211-213.
Godfrey, A. Blanton, "Statistics, Quality and the Bottom Line," Part 2, ASQC
Statistics Division Newsletter, 1989, 10, 1, Spring: 14-17.
Hoadley, B. “A Zero Defect Paradigm”, ASQC Quality Congress Transaction,
Anaheim, 1986.
Juran, J.M., Juran on Leadership for Quality, 1989, Free Press.
Kenett, R. S. and D. Albert, “The International Quality Manager: Translating quality
concepts into different cultures requires knowing what to do, why it should be done
and how to do it”, Quality Progress, 2001, 34, 7: 45-48.
Kenett, R.S., S. Y. Coleman and D. Stewardson, “Statistical Efficiency: The Practical
Perspective”, Quality and Reliability Engineering International, 2003, 19: 265-272.
Kenett, R. S. and S. Zacks, Modern Industrial Statistics: Design and Control of
Quality and Reliability, 1988, Duxbury Press: San Francisco.
Chapter 5
Management Statistics
Irena Ograjenšek, Ron Kenett, Jeroen de Mast, Ronald Does and Soren Bisgaard
Rationale for Evidence-Based Management
Following the logic of the WP7 research programme, it can be shown that the key
question of what is the contribution of statistics to management processes in
organisations can be addressed from many different perspectives, such as historical,
economic and the perspective of quality management.
Data Sources for Evidence-Based Management
Organisations make use of both internal and external data sources (ranging from
customer transaction and survey data to official statistics). These can be used to
support decision-making under uncertainty (the so-called informed decision-making).
Data from different sources are evaluated according to criteria such as cost, accuracy,
timeliness, etc.
Role of Key Performance Indicators in Evidence-Based Management
Three complementary approaches to set up Evidence-Based Key Performance
Indicators in an organisation exist:



Integrated Models used to map out cause and effect relationships,
the Balanced Scorecard that provides management with a navigation panel
Economic Value Added (EVA), an economic model measuring the creation of
wealth.
A first example of an integrated model was implemented by Sears Roebuck and Co.
into what they call the employee-customer-profit model. Another example of an
Integrated Model is the American Customer Satisfaction Index (ACSI) that is based
on a model originally developed in Sweden. ACSI is a Structural Equations Model
with six endogenous variables measuring perceived and expected quality, perceived
value, satisfaction level, customer complaints and customer retention. In order to
assess the customer satisfaction level, a Partial Least Squares analysis is conducted on
exogenous data gathered through telephone surveys.
The Balanced Scorecard was created Kaplan and Norton to translate vision and
strategy into objectives. The balanced scorecard is meant to help managers keep their
finger on the pulse of the business. It usually has four broad categories, such as
financial performance, customers, internal processes and learning and growth.
Typically, each category will have two to five measures. If the business strategy is to
increase market share and reduce operating costs, the measures may include market
share and cost per unit.
EVA is the one measure that is used to monitor the overall value creation in a
business. EVA is not the strategy; it is the way we measure the results. There are
many value drivers that need to be managed, but there can be only one measure that
demonstrates success. A single measure is needed as the ultimate reference of
performance to help managers balance conflicting objectives.
Role of Statistical Programmes in Evidence-Based Management
In order to introduce statistical methods in a coherent and operational form in
business and industry, numerous statistical programmes have been proposed, such as
Total Quality Management (TQM), Statistical Process Control (SPC), Six Sigma,
Taguchi’s method, the Shainin System. These programmes would benefit from a
detailed overview and cost-benefit comparison along with practical examples of
successful or failed applications (there are numerous references to these in the
literature).
Case for European Six Sigma Academy
The exact content of Six Sigma courses varies between institutions and countries
across Europe. Some institutions tailor their Six Sigma courses to individual
companies. A number of the institutions have accredited courses and certify people
who pass as Black Belts and Green Belts. A few of the members have also arranged
for students to complete the exams set by the ASQ to gain the ASQ Black Belt
certification in addition to their own certification.
Companies wishing to send staff on Six Sigma courses often want to know if the
course has been approved by any outside organisations. Having ENBIS accreditation
for their Six Sigma courses would be of benefit. It would also be sensible if Six Sigma
exams could count towards university qualifications.
There is a case for agreeing the core curriculum of a Six Sigma course and this should
be the basis for accrediting courses. There must be a common standard or the
accreditation will have no value.
Companies employing Black Belts and Green Belts have difficulty knowing whether
the person is indeed trained to this level, ENBIS accreditation would show that the
person had taken a course covering the core areas and meeting the set standard. This
would also be of use to the employee, as their Black Belt / Green Belt qualification
would be of a certain recognisable standard should they wish to move companies.
The best form of certification for a Black Belt is working for a company known to
have made good savings and improvements using Six Sigma.
Rather than sign up to any other supplier, ENBIS could give accreditation to courses
which it believes trains students up to a standard where they would pass the ASQ
exams.
It is sensible to take a broader long term view and not concentrate solely on Six Sigma
as Six Sigma will most likely have a limited life span. Instead the academy or
accreditation scheme could be for Six Sigma and other up coming subject areas; it
could be a general academy for all the business areas of ENBIS.
ENBIS already has a procedure relating to endorsements. At present the information
on this procedure is not widely available. It was suggested that the existing procedure
should be looked at, updated if necessary and made available (perhaps through the
website). It is also noteworthy that as Six Sigma covers statistics and management
issues, ENBIS cannot cover the management training aspects.
References
Albright, S.C., W.L. Winston and C.J. Zappe (2000): Managerial Statistics. Duxbury, Pacific
Grove.
Albright, S.C., W.L. Winston and C.J. Zappe (1999): Data Analysis and Decision Making
with Microsoft Excel. Duxbury, Pacific Grove.
Anderson, E.W. and C. Fornell (2000): The Customer Satisfaction Index as a Leading
Indicator. In Swartz, T.A. and D. Iacobucci, eds., Handbook of Services Marketing &
Management. Sage Publications, Thousand Oaks, 255-267.
Anderson, E.W. and C. Fornell (2000): Foundations of the American Customer Satisfaction
Index. Total Quality Management, 7, 869-882.
Anderson, E.W., C. Fornell and D.R. Lehmann (1994): Customer Satisfaction, Market Share
and Profitability: Findings from Sweden. Journal of Marketing, 3, 53-66.
Anderson, E.W. and M.W. Sullivan (1993): The Antecedents and Consequences of Customer
Satisfaction for Firms. Marketing Science, 2, 125-143.
Beauregard, M.R., R.J. Mikulak and B.A. Olson (1992): A Practical Guide to Statistical
Quality Improvement. Opening up the Statistical Toolbox. Van Nostrand Reinhold, New
York.
Bendell, T., J. Kelly, T. Merry and F. Sims (1993): Quality: Measuring and Monitoring.
Century Business, London.
Bisgaard, S. (2000): The Role of Scientific Method in Quality Management. Total Quality
Management, 3, 295-306.
Breyfogle, Forrest W. III (1999): Implementing Six Sigma. Smarter Solutions Using
Statistical Methods. John Wiley & Sons, New York.
Cole, W.E. and J.W. Mogab (1995): The Economics of Total Quality Management: Clashing
Paradigms in the Global Market. Blackwell Publishers, Cambridge (MA).
Czarnecki, M.T. (1999): Managing by Measuring. How to Improve Your Organisation’s
Performance Through Effective Benchmarking. AMACOM, American Management
Association, New York.
Dransfield, S.B., N.I. Fisher and N.J. Vogel (1999): Using Statistics and Statistical Thinking
to Improve Organisational Performance. International Statistical Review, 2, 99-150.
Drummond, H. (1994): The Quality Movement. Nichols Publishing, New Jersey.
Easton, G.S. (1995): A Baldrige Examiner’s Assessment of U.S. Total Quality Management.
In Cole, R.E., ed., The Death and Life of the American Quality Movement. Oxford University
Press, New York, 11-41.
Easton, G.S. and S.L. Jarrell (2000): The Effects of Total Quality Management on Corporate
Performance. An Empirical Investigation. In Cole, R.E. and W.R. Scott, eds., The Quality
Movement & Organization Theory. Sage Publications, Inc., Thousand Oaks, 237-270.
Easton, G.S. and S.L. Jarrell (2000a): Patterns in the Deployment of Total Quality
Management. An Analysis of 44 Leading Companies. In Cole, R.E. and W.R. Scott, eds., The
Quality Movement & Organization Theory. Sage Publications, Inc., Thousand Oaks, 89-130.
Fornell, C. (1992): A National Customer Satisfaction Barometer: The Swedish Experience.
Journal of Marketing, 1, 6-21.
Fornell, C., M.D. Johnson, E.W. Anderson, J. Cha and B.E. Bryant (1996): The American
Customer Satisfaction Index: Nature, Purpose and Findings. Journal of Marketing, 4, 7-18.
Garvare, R. and H. Wiklund (1997): Facilitating the Use of Statistical Methods in Small and
Medium Sized Enterprises. Proceedings of the 41st Congress of the European Organization
for Quality, Vol. 3. European Organization for Quality, Trondheim, 211-220.
Hoerl, R.W. (1998): Six Sigma and the Future of the Quality Profession. Quality Progress,
June, 35-69.
Hogg, R.V. (1997): The Quality Movement: Where It Stands and the Role of Statisticians in
Its Future. In Ghosh, S., W.R. Schucany and W.B. Smith, eds., Statistics of Quality. Marcel
Dekker, Inc., New York, 11-20.
Kaplan, R.S. and D.P. Norton (2004): Strategy Maps. Harvard Business School Press.
Kaplan, R.S. and D.P. Norton (1996): The Balanced Scorecard. Harvard Business School
Press.
Kim, J.S. and M.D. Larsen (1997): Integration of Statistical Techniques into Quality
Improvement Systems. Proceedings of the 41st Congress of the European Organization for
Quality, Vol. 2. European Organization for Quality, Trondheim, 277-284.
Ograjenšek, I. and P. Thyregod (2004): Qualitative vs. quantitative methods. Quality
Progress, Jan. 2004, Vol. 37, 1, 82-85.
Snee, R.D. (1999): Statisticians Must Develop Data-Based Management and Improvement
Systems as Well as Create Measurement Systems. International Statistical Review, 2, 139144.
Thyregod, P. and K. Conradsen (1999): Discussion. International Statistical Review, 2, 144146.
Chapter 6
Service Quality
Irena Ograjenšek
Rationale for Statistical Quality Control and Improvement in the Service Sector
Service industries embraced the basic quality improvement ideas simultaneously with
the manufacturing sector, but have been neglecting the use of statistical methods in
quality improvement processes even more than their manufacturing counterparts. One
of the major arguments against the use of statistical methods is given by differences in
the nature of services and manufactured goods. These differences have always been
emphasised in the literature, especially with regard to measurability of service quality
attributes and, consequently, characteristics of the measurement process. Whether
such emphasis is still valid in the 21st century is a moot point, perhaps it is now
possible to talk about a paradigm change.
Approaches to Statistical Quality Control and Improvement in the Service
Sector
There are important differences among the transcendent or philosophic approach, the
manufacturing-based approach, the product-based approach, the user-based approach
and the value-based approach to quality improvement in the service sector. The tools
used in the framework of each approach are specific in terms of their focus (internal
or external) and practical applicability.
Statistical Toolbox of the Manufacturing-Based Approach to Service Quality
There is a contrast between observational and inferential studies in service operations:
on the one hand there is the application of the basic 7 tools as defined by Ishikawa
and on the other hand the design of experiments, conjoint analysis, Markov chains,
risk analysis, etc.
Statistical Toolbox of the Product-Based Approach to Service Quality
The analysis of quality attributes is particularly important in this approach. There are
pros and cons for the application of rating scales (e.g. SERVQUAL, SERVPERF,
etc.), penalty/reward analysis and the vignette method, etc. Survey research and
structural equation modelling (SEM) also are important features of the statistical
toolbox.
Statistical Toolbox of the User-Based Approach to Service Quality
The technicalities of mystery shopping fall into the user-based approach, along with
methods and techniques of the analysis of overall quality (critical incident technique,
analysis of complaints, analysis of contacts /service blueprinting).
Statistical Toolbox of the Value-Based Approach to Service Quality
Tools such as e.g. cost of quality analysis, or analysis of willingness to pay, are
important to the value-based approach.
References
Babakus, E. and G.W. Boller (1992): An Empirical Assessment of the SERVQUAL Scale.
Journal of Business Research, 3, 253-268.
Babakus, E. and W.G. Mangold (1992): Adapting the SERVQUAL Scale to Hospital
Services: An Empirical Investigation. Health Services Research, 6, 767-786.
Babakus, E. and M. Inhofe (1991): Measuring Perceived Service Quality as a Multiattribute
Attitude. Journal of International Marketing.
Bisgaard, S. (2000a): Service Quality. In Belz, C. and T. Bieger, Dienstleistungskompetenz
und innovative Geschäftsmodelle. Verlag Thexis des Forschungsinstituts für Absatz und
Handel an der Universität St. Gallen, St. Gallen, 296-308.
Bolton R.N. and J.H. Drew (1994): Linking Customer Satisfaction to Service Operations and
Outcomes. In Rust R.T. and R.L. Oliver, eds., Service Quality. New Directions in Theory and
Practice. Sage Publications, Thousand Oaks, 173-200.
Bolton, R.N. and J.H. Drew (1991): A Multistage Model of Customers’ Assessments of
Service Quality and Value. Journal of Consumer Research, March, 375-384.
Boshoff, C., G. Mels and D. Nel (1995): The Dimensions of Perceived Service Quality: The
Original European Perspective Re-Visited. In Bergadaà, M., Proceedings of the 24th
European Marketing Academy Conference. ESSEC, Cergy-Pontoise, 161-175.
Brensinger, D.P. and D.M. Lambert (1990): Can the SERVQUAL Scale be Generalised to
Business-to-Business Services? In Enhancing Knowledge Development in Marketing.
American Marketing Association, Chicago, starting page 289.
Brown, S.W. and E.U. Bond III (1995): The Internal Market/External Market Framework and
Service Quality: Toward Theory in Services Marketing. Journal of Marketing Management,
1-3, 25-39.
Brown, T.J., G.A. Churchill and J.P. Peter (1993): Improving the Measurement of Service
Quality. Research Note. Journal of Retailing, 1, 127-139.
Buttle, F.A. (1995): What Future for SERVQUAL? In Bergadaà, M., Proceedings of the 24th
European Marketing Academy Conference. ESSEC, Cergy-Pontoise, 211-230.
Campanella, J. and F.J. Corcoran (1991): Principles of Quality Costs. In Drewes, W.F., ed.,
Quality Dynamics for the Service Industry. ASQC Quality Press, Milwaukee, 85-102.
Carman, J.M. (1990): Consumer Perceptions of Service Quality: An Assessment of the
SERVQUAL Dimensions. Journal of Retailing, 2, 33-55.
Carroll, J.D. and P.E. Green (1995): Psychometric Methods in Marketing Research: Part I,
Conjoint Analysis. Journal of Marketing Research, November, 385-391.
Cronin, J.J.Jr. and S.A. Taylor (1994): SERVPERF Versus SERVQUAL: Reconciling
Performance-Based and Perceptions-Minus-Expectations Measurement of Service Quality.
Journal of Marketing, 1, 125-131.
Cronin, J.J.Jr. and S.A. Taylor (1992): Measuring Service Quality: A Reexamination and
Extension. Journal of Marketing, 3, 55-68.
DeSarbo, W.S., L. Huff, M.M. Rolandelli and J. Choi (1994): On the Measurement of
Perceived Service Quality. A Conjoint Analysis Approach. In Rust R.T. and R.L. Oliver, eds.,
Service Quality. New Directions in Theory and Practice. Sage Publications, Thousand Oaks,
201-222.
Dolan, R.J. (1990): Conjoint Analysis: A Manager’s Guide. A Harvard Business School Case
Study, 9-590-059. Harvard Business School, Boston.
Drewes, W.F. (1991): The Cost of Delivering a Quality Product. In Drewes, W.F., ed.,
Quality Dynamics for the Service Industry. ASQC Quality Press, Milwaukee, 103-106.
Edvardsson, B. and B. Gustavsson (1991): Quality in Service and Quality in Service
Organizations. In Brown, S.W., E. Gummesson, B. Edvardsson and B. Gustavsson, eds.,
Service Quality. Multidisciplinary and Multinational Perspectives. Lexington Books,
Lexington, 319-340.
George, W.R. and B.E. Gibson (1991): Blueprinting. A Tool for Managing Quality in
Service. In Brown, S.W., E. Gummesson, B. Edvardsson and B. Gustavsson, eds., Service
Quality. Multidisciplinary and Multinational Perspectives. Lexington Books, Lexington, 7391.
Gummesson, E. (1996): Service Quality is Different, Yes Different! In Sandholm, L., ed.,
Quality Without Borders. Sandholm Associates, Djursholm.
Hallowell, R. and L.A. Schlesinger (2000): The Service Profit Chain. Intelectual Roots,
Current Realities and Future Prospects. In Swartz, T.A. and D. Iacobucci, eds., Handbook of
Services Marketing & Management. Sage Publications, Thousand Oaks, 203-221.
Maas, P. (2000): Transformation von Dienstleistungsunternehmen in Netzwerken Empirische Erkenntnisse im Bereich der Assekuranz. In Belz, C. and T. Bieger,
Dienstleistungskompetenz und innovative Geschäftsmodelle. Verlag Thexis des
Forschungsinstituts für Absatz und Handel an der Universität St. Gallen, St. Gallen, 52-74.
Ograjenšek, I. (2003): Use of Customer Data Analysis in Continuous Quality Improvement of
Service Processes. Proceedings of the Seventh Young Statisticians Meeting (Metodološki
zvezki 21), 51-69.
Ograjenšek, I. (2002): Applying Statistical Tools to Improve Quality in the Service Sector.
Developments in Social Science Methodology (Metodološki zvezki 18), 239-251.
Parasuraman, A., L.L. Berry and V. Zeithaml (1991): Understanding, Measuring and
Improving Service Quality. Findings from a Multiphase Research Program. In Brown, S.W.,
E. Gummesson, B. Edvardsson and B. Gustavsson, eds., Service Quality. Multidisciplinary
and Multinational Perspectives. Lexington Books, Lexington, 253-268.
Parasuraman, A., V. Zeithaml and L.L. Berry (1994): Reassessment of Expectations as a
Comparison Standard in Measuring Service Quality: Implications for Further Research.
Journal of Marketing, 1, 111-124.
Parasuraman, A., V. Zeithaml and L.L. Berry (1988): SERVQUAL: A Multiple-Item Scale
for Measuring Consumer Perceptions of Service Quality. Journal of Retailing, 1, 12-40.
Parasuraman, A., V. Zeithaml and L.L. Berry (1985): A Conceptual Model of Service Quality
and Its Implications for Future Research. Journal of Marketing, Fall, 41-50.
Smith, A.M. (1995): Measuring Service Quality: Is SERVQUAL Now Redundant? Journal of
Marketing Management, 1-3, 257-276.
Taylor, S.A. (1997): Assessing Regression-Based Importance Weights for Quality
Perceptions and Satisfaction Judgements in the Presence of Higher Order and/or Interaction
Effects. Journal of Retailing, Spring, 135-159.
Taylor, S.A. and T.L. Baker (1994): An Assessment of the Relationship Between Service
Quality and Customer Satisfaction in the Formation of Consumer’s Purchase Intentions.
Journal of Retailing, 2, 163-178.
Teas, R.K. (1994): Expectations as a Comparison Standard in Measuring Service Quality: An
Assessment of a Reassessment. Journal of Marketing, 1, 132-139.
Teas, R.K. (1993): Consumer Expectations and the Measurement of Perceived Service
Quality. Journal of Professional Services Marketing, 2, 33-54.
Teas, R.K. (1993a): Expectations, Performance Evaluation, and Consumers’ Perceptions of
Quality. Journal of Marketing, 4, 18-34.
Zahorik, A.J. and R.T. Rust (1992): Modeling the Impact of Service Quality on Profitability.
In Bowen, D.E., T.A. Swartz and S.E. Brown, Advances in Service Marketing and
Management. JAI, Greenwich, 247-276.
Zeithaml, V.A. (2000): Service Quality, Profitability and the Economic Worth of Customers:
What We Know and What We Need to Learn. Journal of the Academy of Marketing Science,
1, 67-85.
Zeithaml, V.A., L.L. Berry and A. Parasuraman (1996): The Behavioral Consequences of
Service Quality. Journal of Marketing, 2, 31-46.
Zeithaml, V.A. and M.J. Bitner (2000): Service Marketing. McGraw-Hill Education, New
York.
Zeithaml, V.A., L.L. Berry and A. Parasuraman (1993): The Nature and Determinants of
Customer Expectations of Service. Journal of the Academy of Marketing Science, 1, 1-12.
Chapter 7
Risk, Finance and Insurance
Henry Wynn
Introduction
Risk management and a number of specialist techniques constitute one of the faster
areas of growth in the application of statistical and data-analytic methods. These
range from what might be called soft risk assessment using simple scoring methods to
very advanced option pricing methods in finance. From the perspective of finance,
risk is usually divided into different categories: market risk, credit risk and
operational risk. The last of these is a catch-all, which covers everything normally
associated with areas such as reliability (covered separately in this report). A special
feature is that much of risk management is driven by new codes of practice in
corporate governance and financial regulation, such as the Basle I and II accords.
Risk metrics, risk scoring
The measurement of risk continues to be problematical. A principal reason is that
there is a strong element of judgement in areas not covered by adequate data capture.
Thus scoring methods are used as simple as a three point scale: red, amber green.
There is a long ranging debate about the different axes of risk: probability, effect
(loss) and other interesting concepts such as controllability. It is of considerable
interest that the more advanced risk metrics consider stochastic components such as
the probability of excess over a threshold, standard deviation as well as means, which
is very familiar in quality improvement (robust design, SPC, reliability). It is to be
hoped that there will be good cross-fertilization in the future.
Risk assessment and risk management
Broadly speaking risk assessment means assessing the level of risk for some
operation, product etc which lies in the future. Risk management is the continuous
control of risk during operations. A key component of risk management is the need to
take preventive or mitigating action and to periodically rescore.
Financial markets
There has been a very large increase in the use of mathematical and statistical
methods in finance driven by trading on the stock market and the design of special
financial derivatives. The core of this is the theory of option pricing which is now
embedded into much software used for trading, even automatic trading. There is a
continual search for advantage in trading through innovations in mathematical
technique and statistical analysis. One such area is “statistical arbitrage” where the
huge data available from trading are analysed to look for so-called “imperfect
markets” from which gains can be made.
Credit risk
More static analysis is used to assess the credit worthiness of consumers and
companies. Regression style models are used. For example, a yes/no response (allow/
do not allow credit) leads to the use of logistic regression, very familiar in medical
statistics (death/ survival).
Regulation
Most modern regulation relates to the annual audit function and to financial
management. It stresses openness, accountability and the realistic approach to risk,
including, importantly, operational risk. There is a greater appreciation of the
interdependence of the different activities of companies and organisations. This has
been given urgency by a number of notorious company failures or near failures.
Chapter 8
Applying data mining methods to business and industry
Paolo Giudici
Introduction
Data mining is a relatively new discipline that has developed mainly from studies
carried out in other disciplines such as in the field of computing, marketing, and
statistics. Many of the methodologies used in data mining come from two branches
of research, one developed in the machine learning community and the other
developed in the statistical community, particularly in multivariate and
computational statistics.
There are different perspectives on data mining from different scientific communities.
It is important to formalise how a data mining process should be entertained, to match
statistical quality criteria. There is a need for a “guide for the reader” that is often
confused about what data mining really does. Different methods are to be considered
for each application at hand
In many applications the aim is to gain value from customer data and investigate
business complexity. Techniques including decision trees, logistic regression, cluster
analysis and self-organising maps are used to measure customer value, segment
customer data, predict customer attrition and understand the uptake of new
technologies such as electronic purchasing.
Brief description of the most relevant business and industrial applications
Much work is underway in a number of important application areas of data mining,
from a business and industrial viewpoint. Some recent work in collaboration with a
bank has involved developing statistical models to predict the churn (abandon) of online clients of the bank. These ideas can be extended to other industries, such as
telephone companies, where churn is a key problem. Another current area of interest
is in comparison of predictive credit rating models. Further work in collaboration with
banks involves developing internal statistical models to predict credit ratings of
customers, in compliance with Basel II requirements.
Much research is underway in web mining, in particular looking at association models
for web usage mining and comparison of different models to understand the best
association model that can be used to describe web visit patterns.
Examples of research in text mining are predictive models developed to classify
documents and the comparison of different predictive models to classify text
documents.
REFERENCES
Akaike, H. (1974). A new look at statistical model identification. IEEE Transactions
on Automatic Control, 19, 716—723.
Bernardo, J.M. and Smith, A.F.M. (1994). Bayesian Theory, New York: Wiley.
Bickel, P.J. and Doksum, K. A. (1977). Mathematical Statistics, New York: Prentice
Hall.
Berry, M and Linoff, G. (1997). Data mining techniques for marketing. New York,
Wiley.
Brooks, S.P., Giudici, P. and Roberts, G.O. (2003). Efficient construction of
reversible jump MCMC proposal distributions. Journal of The Royal
Statistical Society series B, 1, 1-37..
Castelo, R. and Giudici, P. (2003). Improving Markov Chain model search for data
mining. Machine learning, 50, 127-158.
Giudici, P. (2001). Bayesian data mining, with application to credit scoring and
benchmarking. Applied Stochastic Models in Business and Industry, 17, 6981.
Giudici, P. (2003). Applied data mining, London, Wiley.
Hand, D.J., Mannila, H. and Smyth, P (2001). Principles of Data Mining, New York:
MIT Press.
Hand, D. (1997). Construction and assessment of classification rules. London: Wiley.
Han, J. and Kamber, M (2001). Data mining: concepts and techniques. New York:
Morgan and Kaufmann.
Hastie, T., Tibshirani, R., Friedman, J. (2001). The elements of statistical learning:
data mining, inference and prediction, New York: Springer-Verlag.
Heckerman, D., Chickering, D.M, Meek, C, Rounthwaite, R and Kadie, C.,
Dependency Networks for Inference, Collaborative Filtering and Data Visualization.
Journal of Machine Learning Research, 1, pp. 49{75, 2000.
Lauritzen, S.L. (1996). Graphical models, Oxford, Oxford University Press.
Mood, A.M., Graybill, F.A. and Boes, D.C. (1991). Introduction to the theory of
Statistics, Tokyo: Mc Graw Hill.
SAS Institute Inc. (2004). SAS enterprise miner reference
manual. Cary: SAS
Institute.
Schwarz, G. (1978). Estimating the dimension of a model. Annals of Statistics 62,
461-464.
Whittaker, J. (1990). Graphical models in applied multivariate Statistics, Wiley.
Witten, I. and Frank, E. (1999). Data Mining: practical machine learning tools and
techniques with Java implementation. New York: Morgan and Kaufmann.
Zucchini, Walter (2000). An Introduction to Model Selection. Journal of data mining,
inference and prediction. Springer-Verlag, Berlin.
Chapter 9
Process Monitoring, Improvement and Control
Oystein Evandt, Alberto Luceño, Ron Kenett, Rainer Goeb
Introductory comments
The main types of statistical processes used to model industrial processes are “Pure
Random Noise Processes” (in the sense of independent identically distributed random
deviations from a constant value), “Stationary Autocorrelated Processes” and
“Nonstationary Processes”. The differences between these processes can be
visualised when examples of data from such processes are plotted for comparison,
like in section 1.2 in Box and Luceño (1997). (A formal definition of a stationary
process is not given here, we just describe a stationary process as one that varies
around a mean value as time goes by, as in section 1.2 in Box and Luceño (1997). For
a formal definition of a stationary process, see The Oxford Dictionary of Statistical
Terms (2003).)
SPC originated in the parts industry, where large numbers of articles are produced,
and it is aimed at making these articles as equal as possible, with respect to important
characteristics, e.g. dimensions, weight and shape. Processes in this industry can at
best be modelled (approximately) by Pure Random Noise Deviations from the target
value of product characteristics. The deviations should preferably be as small as
possible. SPC is the natural tool for achieving this ideal state of a process, and for
maintaining this state. The Taguchi Loss Function helps to reinforce the implications
of the size of deviations from the target.
Engineering process control (EPC) came into existence in the process industry, where
a typical example is to keep the temperature under which a process is run as constant
as possible at a specified value, to keep the percentage conversion of chemicals in the
production process constant at a value as high as possible etc. To most people it will
seem obvious that time series of measurements of such quantities cannot be modelled
by Pure Random Noise Deviations from a constant value, but that the measurements
will be autocorrelated. (If a melting furnace is too warm at a certain time, it is likely
to be too warm shortly afterwards too.) It will seem reasonable to readers that one
should be aiming at having stationary processes, varying around a mean value as
equal as possible to the target value of the process, and with as little variation as
possible.
The economic gains of controlling processes, in the parts industry and the process
industry respectively, if the ideal states mentioned above are close to being obtained,
are often substantial.
SPC – Statistical Process Control
It is important to differentiate between retrospective analysis of data by means of
control charts, and real time applications of such charts. The most well-known and
probably well-used chart is the Xbar-R chart for grouped data, and variations of it,
and the XmR chart for individual values. Strictly speaking the only alarm rule of a
Shewhart Chart is the three-sigma limit rule. Three-sigma limits work well even if
data do not come from a normal distribution, see for example, the description in
Wheeler and Chambers (1992), section 4.2, “Why Three Sigma Limits?”.
Among most usual alarm rules, in addition to the three-sigma limits rule, are two
successive observations outside of the two sigma limits, 7 successive points
increasing or decreasing and 7 successive points either above or below the central line
of the chart. The application of such rules leads to a chart having memory. Also
applying too many of these rules may increase the tendency of false alarms
unacceptably. The four so-called Western Electric Zone Rules often provide a good
compromise between increased ability to detect out-of-control situations and avoiding
false alarms. See, for example Wheeler and Chambers (1992), section 5.4 “Four Rules
for Defining Lack of Control”.
The exponentially weighted moving average (EWMA) chart is particularly efficient at
detecting certain out-of- control situations, including those caused by underlying
trends. The CUSUM chart is also efficient at detecting these and other out-of- control
situations. Recently other types of control charts have been proposed, most of which
are good in particular situations and have their own their strengths and weaknesses.
Multivariate SPC
Many if not most situations are measured by more than one metric. When two or more
correlated characteristics are observed simultaneously in order to control them,
detection of out-of-control situations in the multivariate process cannot be properly
done by monitoring each characteristic separately by means of univariate control
charts. Further, multivariate SPC is usually based on Hotellings T2 statistic. For this
statistic to follow the T2 distribution, it is assumed that the simultaneously observed
characteristics come from independent observations from a multinormal distribution.
However, multivariate control charts based on T2 are very sensitive to deviations from
the assumption of multinormally distributed observations. If this assumption is not
sufficiently well fulfilled, a way out may be to use control charts based on principal
components. How this is done is accounted for in e.g. Ryan (2000).
On using software for SPC
Site specific SPC software is often a tempting alternative to buying a standard
software package which may need to be adjusted to give reports in the format most
useful for a company. It is difficult to review specific software packages for SPC,
since the packages are relatively frequently updated. Besides the larger,
comprehensive and well-known packages, such as e.g. Minitab, Statistica and
StatGraphics, there is also a plethora of smaller bespoke packages available on the
market. It is important to check each package to ensure that the control limits are
computed correctly, since for some control charts in some of the packages on the
market, this is done incorrectly.
Acceptance Sampling in Quality Control
The main focus here is product control (control of finished product) versus process
control. Lot inspection is a tool of product control and acceptance sampling is the
statistical version of lot inspection. The roots of statistical lot inspection schemes are
in the 1920s at Bell Telephone with the contribution of H. F. Dodge. An early
Military Standard was developed during World War II. Later there was further
development of the Military Standard and the movement to replace military and
national standards with ISO standards. Progress went alongside World Trade
Organization (WTO) agreements on tariffs and trade.
There has been a decline in statistical lot inspection in the quality movement over the
last decades. The emphasis has been on proactive as opposed to reactive quality
control, based on SPC, i.e. prevention versus inspection (sorting). This is generally a
healthy development. It does however not mean that statistical lot inspection is totally
obsolete. There are situatuions in which statistical lot inspection is the proper thing to
do. This is accounted for in a good way in Vardeman (1986).
Lot inspection and statistical lot inspection feature in ISO 9000-9004. The misleading
connotation in "acceptance" sampling is the sorting purpose (traditional) of statistical
lot inspection versus supporting purpose (modern) of statistical lot inspection. There
are now new ISO standards on the subject. These include ISO 13448 (allocation of
priorities principle), ISO 14560 (scheme for quality levels in nonconforming items per
million), ISO 18414 (accept-zero credit-based for outgoing quality), ISO 21247
(accept-zero single and continuous plans for attributes and variables), plus a few
sequential sampling plan standards. An important subset includes sampling plans
based on more than the traditional two states of product classification (primarily three
classes). Fundamental issues with the correct classification of products as
conforming, nonconforming, marginally conforming, etc. affect the operation of
sampling plans and schemes. There is now a movement properly to quantify
measurement uncertainty and include this in product conformity determinations (an
ISO standard was issued on this in 2003). Inspector error is another type of this
general issue. Sampling inspection can be based on new item and lot quality
indicators, e. g. squared or absolute deviation from target, one-sided deviation from
target. There are also corresponding applications in legal metrology and consumer
protection.
References
Box, George and Alberto Luceño (1997): Statistical Control by Monitoring and
Feedback Adjustment, John Wiley & Sons, Inc.
Ryan, Thomas P. (2000): Statistical Methods for Quality Improvement, John Wiley &
Sons, Inc.
The Oxford Dictionary of Statistical Terms (2003), Oxford University Press
Vardeman, Stephen B. (1986): The Legitimate Role of Inspection in Modern SQC.
The American Statistician, Vol. 40, No. 4.
Wheeler, Donald J. and David S. Chambers (1992): Understanding Statistical Process
Control, Second Edition, SPC Press, Inc., Knoxville, Tennesse
Chapter 10
Measurement Systems Analysis
Raffaello Levi
Statistical methods play a key role in the analysis of measurement and testing
work performed within the framework of quality systems. Measurement and testing
procedure organisation, as supported particularly by DOE, call for particular attention
to a fundamental quality index related to both cost and benefit, namely uncertainty
associated to measurement under consideration.
Indication of uncertainty is mandated by current quality standards, starting from
ISO 9001 and, more specifically, ISO 17025, governing management of testing and
measurement laboratories. Appraising the paramount importance of the subject led
major international organisations covering metrology and standardisation, such as
BIPM, IEC, IFCC, IUPAC, IUPAP, OIML, to draft and publish under the aegis of
ISO a fundamental reference text, “Guide to the expression of uncertainty in
measurement”, currently referred to as GUM and embodied into European standard
ENV 13005.
Statistical procedures dictated by GUM cover a broad range of applications.
Besides definition of such a delicate undertaking as evaluation of measurement
uncertainty with a clear set of rules accepted worldwide, by no means a minor
achievement, they cater for planning measurement and testing work aimed at specific
levels of uncertainty, in order to avoid both failure to reach mandated accuracy and
costly overdesign. These methods - covering both specific metrological work, such as
e.g. calibration of sensors and instrument systems, and generic testing work – deal
with three main items, namely:
 contributions to uncertainty belonging to A category, estimated according to
statistical methods. Besides managing random errors (always present in
measurements) with such tools as normal and Student’s distributions, they
enable detection, and estimation, of systematic effects (e.g. through tests of
normality, of linearity, and ANOVA), and furthermore proper treatment of
such outliers and mavericks typically associated to the ever increasing
diffusion of electronic instrumentation. Their inherently high sensitivity makes
these instrument systems often open to spurious signals due to electromagnetic
noise, leading to measurement incidents to be dealt with according to proper
exclusion procedures;
 contributions to uncertainty belonging to B category, assessed according to
non statistical methods (mainly according to technical expertise, and
accumulated experience), and transformed into equivalent variance
components, according to uniform, U shaped or triangular distributions
selected according to experience;
 composition of contributions mentioned above, and allotment of proper
degrees of freedom, leading to evaluation of overall uncertainty under the
form of a confidence interval, related to normal or Student’s distributions.
Categorisation and tabulation of terms entering measurement uncertainty
computation cater for assessment and comparative evaluation of contributions
pertaining to individual factors, according to an iterative computation routine (PUMA
method, ISO 14253-2). An awkward part of the process of uncertainty budget
management, namely that dealing with individual contributions to be associated with
every uncertainty factor, and with suggesting which factors are best acted upon in
order to cut costs and/or measurement uncertainty if need be, may thus be made
straightforward.
Chapter 11
Design and Analysis of Experiments (DOE)
Henry Wynn.
Introduction.
There has been a rapid growth in the application of experimental design to industry
from the early 1980s and largely as a component in quality improvement. This growth
was particularly rapid in the area associated with the name of Genichi Taguchi,
namely robust engineering design, but has spread to the use of experimental design
generally. Of course, DOE had already made major contributions in the areas of
factorial design and response surface with the work of G E P Box, N Draper and
others. The robust design ideas have now been absorbed into the mainstream
alongside factorial design, response surface design and to this mix must be added
developments in the optimal design of experiments, computer experiments and
simulation.
Factorial design.
This is the bedrock of the field. The basic idea is that to estimate effects (parameters)
within the context of regression it is not necessary to measure a response at every
level of every factor (explanatory variable). By careful choice of configurations, that
is to say an experimental design, much can be learned. Roughly, the more complex
the model the more complex and larger an experiment is needed.
Response surfaces: design and models.
This term is usually taken to mean designs which lie outside the somewhat rigid
structures of traditional factorial design and may be more adventurous in layout:
center points, star points, composites of different points, designs to guard against
hidden trends, designs for mixtures experiments and so on.
Optimal design.
A contribution from the more mathematical wing of statistics was to set up
experimental design as a decision problem and to optimize. J. Kiefer and co-workers
launched this field around 1960 but it took a number of years for the methods to get
out of academia into industry. Now algorithms for optimal design are available (as for
factorial design) in a number of packages. They are particularly useful for getting a
good experiment for a non standard region of experimentation.
Robust Engineering Design.
The major impact of this area has been to help reduce variability by paying attention
not only to the mean response but also to the variability. Special types of experimental
design can be used to estimate both the mean and the variability due to so-called
“noise” factors, and therefore need to be controlled in some way (using “control”
factors). This is sometimes called the dual response surface method. The modelbased version models all factors and then propagates the variability through the model
analytically or via simulation.
Computer experiments.
Many complex engineering products are designed using simulation packages. These
may for example solve banks of differential equations. Even though computers get
faster a single simulation may still take hours. This means that design of experiments
can be used to save computer runs. Also this has led to the growth of more complex
modelling including kernel methods, as opposed to the polynomial of the usual
response surface method.
Chapter 12
Safety and reliability engineering
Chris McCollin, Maria Ramalhoto
Introduction
The safety issue in engineering is very often a broad subject of a huge complex
nature. However, it might be substantially improved in many situations if we
understand better the stochastic aspects of reliability, maintenance, control, and its
interactions usually present in the equipment’s failure that caused the lack of safety.
Innovative ways to introduce that knowledge in the available safety frameworks
might prove useful.
Reliability analysis is a well-established part of most risk and safety studies. The
causal part of a risk analysis is usually accomplished by reliability techniques, like
failure mode and effects analysis (FMEA) and fault tree analysis. Reliability analysis
is often applied in risk analysis to evaluate the availability and applicability of safety
systems, ranging from single component safety barriers (e.g., valves) up to complex
computer based process safety systems.
A wide range of standards with respect to reliability and safety have been issued.
Among them are: British Standards BS 5760; MIL-STD (783, 882 D, 756, 105 D and
others,) IEEE Std. 352, IEC300 and so on. It is also interesting to notice the relation
between the ISO 9000 and the IEC 300 series of standards.
As part of the formation of the European Union a number of new EU directives have
been issued. Among these are the machine safety directive, the product safety
directive, major hazards directive and the product liability directive. The producers of
equipment must, according to these directives, verify that their equipment complies
with those requirements. Reliability analysis and reliability demonstration testing are
necessary tools in the verification process. It has already been reported that in the
USA courts, consultancy on these matters is becoming a booming business.
Engineering design reliability
The areas of design which mainly affect reliability are reliability specification; the
initial systems design process incorporating feasibility and concept design; the
development process incorporating the main CAD/CAE activities including FMEA,
circuit design and simulation packages; the development process incorporating
prototype build; components and training.
Recent papers on design for reliability are given in the references. General approaches
to reliability development cover four areas:
 Costing, Management and Specification
 Design Failure Investigation Techniques (DFITs)
 Development Testing
 Failure Reporting
Skills for a statistician/engineer working in safety and reliability engineering
Aviation, naval and offshore among other industries have already realized the
important connection between maintenance and reliability and implemented the
reliability centered maintenance (RCM) methodology. Identification of general and
specific industry targeted skills is important as is the presentation of a framework to
guarantee communication between different groups of stakeholders and to provide upto-date information on required skills.
Perspectives for safety and reliability engineering: trans-disciplinary issues
under the umbrella of the “Stochastics for the Quality Movement" Concept
Keeping in mind that Reliability is Quality over time and Maintenance is a Reliability
Function, an “Extended Quality Movement” integrating among others the Reliability
and the Maintenance Cultures makes sense. SQM (stochastics for the quality
movement), introduced in Ramalhoto (1999), embraces all the non-deterministic
quantitative methods relevant to industry and business practices. One of the aims is to
supply a complete body of trans-disciplinary and targeted up-to-date knowledge on
theory and practice, and facilitate their synergy development. That is a very
demanding and ambitious dream that has been already discussed among several
members of the Pro-ENBIS project. A vision for an ideal targeted framework could be
outlined in the context of maritime safety.
Signature analysis and predictive maintenance.
It is very advantageous if future wear and failure can be detected before it occurs.
This enables preventive action such as maintenance to be carried out, with cost
savings and the avoidance of possibly dangerous failure. Signature analysis is the
method of fault detection via a special pattern of, typically, a dynamic characteristic
such as noise, vibration, electrical output. Ideally every failure mode will have its own
special signature. Signal processing methods are typically used such as time
frequency plots using wavelet analysis.
References
Akao, Y. (1990) Quality Function Deployment Integrating Customer Requirements
into Product Design. Productivity Press.
Andrews, J.D. and Dugan, J.B. (1999) Advances in Fault tree analysis. Proceedings of
the SARSS. pp10/1-10/12.
Bartlett, L.M. and Andrews, J.D. (1999) Comparison of variable ordering
heuristics/algorithms for binary decision diagrams. Proceedings of the SARSS.
pp12/1-12/15.
Braglia, M. (2000) MAFMA: multi-attribute failure mode analysis. International
Journal of Quality and Reliability Management, Vol. 17 No. 9, pp1017-1033.
British Standards Institution. (2000) BS EN ISO 9001: Quality Management Systems
- Requirements. London: British Standards Institution.
Cini, P.F. and Griffith, P. (1999) Designing for MFOP: towards the autonomous
aircraft. Journal of Quality in Maintenance Engineering. Vol. 5 No. 4, pp 296-306.
DEF-STAN-0041 Parts 1 to 5: MOD Practices and Procedures for R & M. HMSO.
Drake Jr., P. (1999) Dimensioning and Tolerancing Handbook McGraw-Hill New
York.
Edwards, I.J. (2000) The impact of variable hazard-rates on in-service reliability. 14th
ARTS Conference University of Manchester.
Fagan, M.E. (1976) Design and Code inspections to reduce errors in program
development. IBM Systems Journal 15(3) pp182-211.
Feynman, R.P, (1989) "What do you care what other people think?" Harper Collins.
Gall, B. (1999) An advanced method for preventing business disruption. Proceedings
of the SARSS. pp1/1-1/12.
Gastwirth, J. L. (1991) The potential effect of unchecked statistical assumptions.
Journal of the Royal Statistical Society. Series A Vol 154 part 1, pp121-123.
Goldstein, H., Rasbash, J., Plewis, I., Draper, D., Browne, W., Yang, M., Woodhouse,
G. and Healy, M.J.R. (1998) A User's Guide to MLWin. London: Institute of
Education.
Goodacre, J. (2000) Identifying Current Industrial Needs from Root Cause Analysis
activity 14th ARTS Conference University of Manchester 28th-30th November 2000.
Gray, C., Harris, N., Bendell, A. and Walker, E.V. (1988) The Reliability Analysis of
Weapon Systems. Reliability Engineering and System Safety. 21. pp 245-269.
Harvey, A.C. (1993) Time Series Models Harvester Wheatsheaf 2nd edition p285.
Ke, H. and Shen, F. (1999) Integrated Bayesian reliability assessment during
equipment development. International Journal of Quality and Reliability
Management, Vol. 16 No. 9, pp892-902.
Kenett R., Ramalhoto M. F., Shade J. (2003). A proposal for management
knowledge of stochastics in the quality movement. In T.Bedford & P. H. A. L.
M. van Gelder (Eds), Safety & Reliability - ESREL 2003, Vol. 1, pp. 881-888,
Rotterdam: Balkema.
McCollin, C. (1999) Working around failure. Manufacturing Engineer, Volume 78
No. 1. February 1999. pp37-40.
Morrison, S.J. (2000) Statistical Engineering Design. The TQM magazine. Vol. 12
No. 1 pp26-30.
Morrison, S.J. (1957) Variability in Engineering Design. Applied Statistics, Vol 6.
Pp133-138.
O’Connor, P.D.T. (2000) Commentary: reliability – past, present and future. IEEE
Transactions on Reliability Vol. 49, issue 4, pp335-341.
Parry-Jones, R. (1999) Engineering for Corporate Success in the New Millennium.
Speech to the Royal Academy of Engineering, London 10th.
and A new way to teach statistics to engineers. To be reported in MSOR Connections.
Paulk, M. et al. (editors) (1995) The Capability Maturity Model Guidelines for
improving the software process. Addison-Wesley.
Ramalhoto M. F. (1999). A way of addressing some of the new challenges of
quality management. In: G. I. Schueller and P. Kafka (Eds) Safety & Reliability ESREL 1999, Vol. 2, pp. 1077-1082, Rotterdam: Balkema.
Reason, J. (1999) Managing the risks of organisational accidents Ashgate, Aldershot.
pp159.
Reunen, M., Heikkila, J. and Hanninen, S. (1989) On the Safety and Reliability
Engineering during Conceptual Design Phase of Mechatronic Products in Reliability
Achievement The Commercial Incentive Ed. T. Aven Elsevier.
Sankar, N.R. and Bantwal, S.P. (2001) Modified approach for prioritization of failures
in a system failure mode and effects analysis. International Journal of Quality and
Reliability Management, Vol. 18 No. 3, pp324-335.
Sexton, C.J., Lewis, S.M. and Please, C.P. (2001) Experiments for derived factors
with application to hydraulic gear pumps. Journal of the Royal Statistical Society.
Series C (Applied Statistics) Vol 50. Part 2. pp155-170.
Silverman, M. Why HALT cannot produce a meaningful MTBF number and why this
should not be a concern. http://www.qualmark.com/hh/01MTBFpaper.htm/
Strutt, J.E. (2000) Design for Reliability: A Key Risk Management Strategy in
Product Development. 14th ARTS Conference University of Manchester.
Swift, K.G., Raines, M and Booker, J.D. (1999) Designing for Reliability: a
probabilistic approach. Proceedings of the SARSS. pp3/1-3/12.
Thompson, G. A. (2000) Multi-objective approach to design for reliability. 14th ARTS
Conference University of Manchester.
Woodall, A., Bendell, A. and McCollin, C. Results of the Engineering Quality Forum
Survey to establish ongoing requirements for education and competency for engineers
in the field of quality management. Available from the IMechE. Website:
http://www.imeche.org.uk/manufacturing/quality%5Fand%5Fengineering%5Fsurvey.
htm/
Chapter 13
Multivariate analysis focussing on Multiscale modelling
Marco S. Reis and Pedro M. Saraiva
Classical modelling approaches, typically use (either explicit or implicitly) a single
particular time scale (usually that corresponding to lowest acquisition rate or that of
the most important set of variables), where they base the whole analysis, according to
the objectives to attain. This happens, e.g., in state-space, time-series, mechanistic and
regression models, and this limitation is passed on to other tasks that rely on them,
such as model parameter estimation (where such structures are used along with
properly generated data and a quality criteria of fitness that measures the adequacy of
predictions, usually at the lowest acquisition rate), as well as, for instance: data
rectification, fault detection and diagnosis, process monitoring, optimization and
design. However, in most of the application domains found in practice, the relevant
phenomena take place simultaneously at different time scales. In more technical
terms, this means that usually events occur at different locations and with different
localizations in the time (and frequency) domain. Thus, techniques that "zoom" the
analysis focus only at the finest scale, establishing their quality criteria in terms of the
fastest dynamics, are likely to neglect important information that corresponds to other
localizations in the time-frequency plan.
In this regard, tools derived from wavelet theory (Mallat, 1998) have been finding
wide been acceptance in applications where the data typically present multiscale
features, notably in signal de-noising and compression applications (Donoho and
Johnstone, 1992; Vetterli and Kovačević, 1995), but others can also be referred, such
as process monitoring (Bakshi, 1998; Ganesan et al., 2004) and system identification
(Tsatsanis and Giannakis, 1993). The wavelet-based methodologies do enable the
incorporation of the concept of scale right into the core of the data analysis tasks,
constituting the adequate mathematical language to describe multiscale phenomena.
Much of the success of wavelets arises from the efficiency they describe data
composed by events having different localization properties in time or frequency. In
fact, a proper analysis of these signals would require, for instance, a large number of
Fourier transform coefficients, meaning that it is not an “adequate” language for a
compact translation of the signal key features into the transform domain. This
happens because the form of the time/frequency windows (Vetterli and Kovačević,
1995; Mallat, 1998) associated with their basis functions does not change across the
time/frequency plane, in order to cover effectively and efficiently the localized high
energy zones of the several features present in the signal. Therefore, to cope with such
multiscale features, a more flexible tiling of the time/frequency space is required, and
can be provided by the wavelet basis functions, whose coefficients are called wavelet
transform. In practice, it is often the case that signals are composed of short duration
events of high frequency and low frequency events of long duration, and this is
exactly the kind of tiling that a wavelet basis does provide, since the relative
frequency bandwidth of these basis functions is a constant (i.e., the ratio between a
measure of the size of the frequency band and the mean frequency,   , is constant
for each wavelet function), a property also referred to as a “constant-Q” scheme.
Therefore, by developing approaches that integrate wavelet theory concepts with
classical multivariate methodologies, one can handle both the data features arising
from data dimensionality and correlation issues, with those arising from multiscale
characteristics. As some examples of efforts directed towards such goal, one can refer
the Multiscale Principal Component Analysis (MS-PCA; Bakshi, 1998), Multiscale
Regression (Depczynsky et al., 1999) and Multiscale Classification (Walczak et al.,
1996).
The comments above focus on the developments of multiscale (or multiresolution)
approaches for conducting classical data analysis tasks, which naturally integrate the
multiscale nature of the underlying phenomena, through the incorporation of a proper
scale variable right into the systems modelling, analysis and optimization paradigms.
There are several widely different application domains found in industry, where
multiscale approaches have been developed.
References
BAKSHI, B.R.  Multiscale Principal Component Analysis with Application to
Multivariate Statistical Process Monitoring. AIChE Journal. 44:7 (1998), p. 1596 –
1610.
DEPCZYNSKY, U.; JETTER, K.; MOLT, K.; NIEMÖLLER, A.—The Fast Wavelet
Transform on Compact Intervals as a Toll in Chemometrics – II. Boundary Effects,
Denoising and Compression. Chemometrics and Intelligent Laboratory Systems. 49
(1999): 151-161.
DONOHO, D.L.; JOHNSTONE, I.M. — Ideal Spatial Adaptation by Wavelet Shrinkage.
Department of Statistics, Stanford University (1992). Technical report.
GANESAN, R., DAS, T. K., AND VENKATARAMAN, V. — Wavelet Based Multiscale
Process Monitoring - A Literature Review. IIE Trans. on Quality and Reliability Eng.,
36 (9) (2004).
MALLAT, S. — A Wavelet Tour of Signal Processing. San Diego [etc.]: Academic
Press, 1998.
TSATSANIS, M.K.; GIANNAKIS,G.B. — Time-Varying Identification and Model
Validation Using Wavelets. IEEE Transaction on Signal Processing. 41:12, (1993):
3512-3274.
VETTERLI, M.; KOVAČEVIĆ, J. — Wavelets and Subband Coding. New Jersey:
Prentice Hall, 1995.
WALCZAK, B.; VAN DEN BOGAERT, V.; MASSART, D.L. — Application of Wavelet
Packet Transform in Patter Recognition of Near-IR Data. Anal. Chem. 68 (1996).
1742-1747.
Chapter 14
SIMULATION
David Rios Insua, Jorge Muruzabal, Jesus Palomo,
Fabrizio Ruggeri, Julio Holgado and Raul Moreno
Purpose
Once we have realised that a given system is not operating as desired, we would look
for ways to improving it. Sometimes it is possible to experiment with the real system
and, through observation and the aid of Statistics, reach valid conclusions towards
system improvement. Experimenting with a real system may entail ethical and/or
economical problems, which may be avoided by dealing with a prototype, a physical
model. However, sometimes we are not able, or it is not feasible or possible, to build
a prototype. Yet, we may obtain a mathematical model describing, through equations
and relations, the essential behaviour of the system. Its analysis may be done,
sometimes, through analytical or numerical methods. But the model may be too
complex to be dealt with. In such extreme cases, we may use Simulation. Essentially,
Simulation consists of (i) building a computer model that describes the behaviour of
the system; and (ii) experimenting with this model to reach conclusions that support
decision making.
There are several key concepts, methods and tools from the world of Simulation
which may prove useful to the industrial statistics practitioner. The basic four-step
process in any simulation experiment is as follows:
1.
2.
3.
4.
Obtain a source of random numbers.
Transform them into inputs to the model.
Transform these inputs into outputs of the model.
Analyse the outputs to reach conclusions.
These steps have been successfully applied to many practical situations to help solve
real problems, for example modelling a workflow line within the digitalisation
industry.
It is important to consider carefully the random number generation. It is not always
sensible to use the standard generators in standard software packages. Checks on the
randomness may show that there are significant patterns in the occurrence of
particular pairings of numbers, for example. Other properties are relevant as well:
computational efficiency (e.g. speed; little memory consumed), portability;
implementation simplicity, reproducibility, mutability and long enough period.
Good sources of random number generators are at Statlib at http://www.stat.cmu.edu/.
Other important sites in relation with random number generation are L'Ecuyer's page
at http://www.iro.umontreal.ca/lecuyer and http://random.mat.sbg.ac.at. A set of
statistical
tests
for
random
number
generators
are
available
at
http://csrc.nist.gov/rng/rng5.html.
The next step in a simulation experiment is to convert the random numbers into inputs
appropriate for the model at hand. The most popular traditional method for drawing
from a distribution F is based on inversion, i.e. in generating from a uniform
distribution and computing the inverse of F at the drawn value. In Bayesian statistics
the most important techniques used to simulate from a posterior distribution are the
Markov chain Monte Carlo (MCMC) ones.
As far as software is concerned, the Numerical Recipes (Press et al, 1992,or
http://www.nr.com) include code to generate from the exponential, normal, gamma,
Poisson and binomial distributions, from which many other distributions may be
sampled based on the principles outlined above. Many generators are available at
http://www.netlib.org/random/index.html.
WINBUGS and OpenBUGS are
downloadable from http://www.mrc-bsu.cam.ac.uk/bugs, facilitating MCMC
sampling in many applied settings. Another useful library is GSL, available at
http://www.gnu.org/software/gsl.
The third step in a simulation process consists of passing the inputs through the
simulation model to obtain outputs to be analysed later. Monte Carlo simulation is a
key method in Industrial Statistics. We may use MC methods for optimisation
purposes (say to a obtain an MLE or a posterior mode); for resampling purposes, as in
the bootstrap; within MC hypothesis tests and confidence intervals; for computations
in probabilistic expert systems, ... the key application being Monte Carlo integration,
especially within Bayesian statistics.
The final stage of a simulation experiment consists of the analysis of the output
obtained through the experiment. To a large extent, we may use standard estimation
methods, point estimates and precision measures, with the key observation that the
output might be correlated. Clearly, as we deal with stochastic models, each repetition
of the experiment will lead to a different result, provided that we use different seeds to
initialise the random number generators at each replication. The general issue here is
to provide information about some performance measure of our system, e.g.
unbiasedness, mean square error and consistency in point estimation. In MCMC
simulation, within Bayesian statistics, convergence detection is an important issue, for
which CODA http://www.mrc-bsu.cam.ac.uk/bugs/classic/coda04/readme.shtml has
been developed.
Finally, there are some tactical issues in simulation: how do we design simulation
experiments, how do we combine simulation with optimisation and the issue of
variance reduction.
References
Balci, O. (ed) (1994). Simulation and Modelling, Annals of Operations
Research, 53.
Banks, J. (1998). Handbook of Simulation, Wiley, New York.
Bays, C. and Durham, S. (1976). Improving a poor random number generator,
ACM Transactions in Mathematical Software, 2, 59-64.
Bratley, P., Fox, B. and Schrage, L. (1987). A Guide to Simulation, Springer,
New York.
Chick, S.E. (2001). Input Distribution Selection for Simulation Experiments:
Accounting for Input Uncertainty, Operations Research, 49, 744-758.
Extend, official web page. http://www.imaginethatinc.com
Fishman, G. S. (1996). Monte Carlo: Concepts, Algorithms and Applications,
Springer, New York.
French, S. and Rios Insua, D. (2000). Statistical Decision Theory, Arnold,
London.
Fu, M.C. (1994). Optimization via simulation, Annals of Operations Research, 53,
199-248.
Gamerman, D. (1997). Markov Chain Monte Carlo: Stochastic Simulation
for Bayesian Inference, Chapman & Hall, New York.
Imaginethatinc (2000). Extend Manual, Imagine That Inc.
Kleijnen, J.P.C. (1987). Statistical Tools for Simulation Practitioners, Dekker,
New York.
Knuth, D. (1981). The Art of Computer Programming, Vol. 2: Seminumerical
Algorithms, Addison Wesley, New York.
Law, A. M. and Kelton, W. D. (1991). Simulation Modeling and Analysis,
McGraw-Hill, New York.
L'Ecuyer, P. (1990). Random numbers for simulation, Communications
ACM, 33, 85-97.
L'Ecuyer, P. (1994). Uniform random number generators, Annals of Operations
Research, 53, 77-120.
L'Ecuyer (1998). Random Number Generators and Empirical Tests, Lecture
Notes in Statistics 127, Springer-Verlag, 1998, 124­138.
L'Ecuyer, P. (2001). Software for Uniform Random Number Generation:
Distinguishing the good and the bad, Proceedings of the 2001 Winter
Simulation Conference, IEEE Press, Dec, 95-105.
L'Ecuyer, P. and Granger-Piche, J. (2003). Combined Generators with
Components from Different Families, Mathematics and Computers in
Simulation, 62, 395-404.
L'Ecuyer, P., Simard, R., Chen, E.J. and Kelton, W.D. (2002). An objectoriented random number package with many long streams and substreams, Operation Research, 50, 1073-1075.
Lehmer, D. H. (1951). Mathematical methods in large-scale computing
units, Proceedings of the Second Symposium on Large Scale Digital
Computing Machinery, 141146, Harvard University Press, Cambridge.
Lewis, P.A., Goodman, A.S. and Miller, J.M. (1969). A pseudo-random
number generator for the system/360, IBM System's Journal, 8, 136143.
Matsumoto, M. and Nishimura, T. (1998). Mersenne twister: A 623dimensionally equidistributed uniform pseudo-random number generator, ACM
Transactions on Modelling and Computer Simulation, 8, 3-30..
Neelamkavil, F. (1987). Computer Simulation and Modelling, Wiley, New
York.
Niederreiter, H. (1992). Random Number Generation and Quasi-Monte
Carlo Methods, SIAM, Philadelphia.
Park, S. and Miller, K. (1988). Random number generators: good ones are
hard to find, Communications ACM, 31, 1192-1201.
Press, W.H., Teukolsky, S.A., Vetterling, W.T. and Flannery, B.P. (1992).
Numerical recipes in C, Cambridge University Press, Cambridge.
R, official web page. http://www.r-project.org
RAND Co. (1955). A Million Random Digits with 100000 Normal Deviates,
Free Press.
Rios Insua, D., Rios Insua, S. and Martin, J. (1997). Simulacion, Metodos
y Aplicaciones., RA-MA, Madrid.
Ripley, B. (1987). Stochastic Simulation, Wiley, New York.
Robbins, H. and Monro, S. (1951). A stochastic approximation method,
Annals of Mathematical Statistics, 22, 400-407.
Ross, S. (1991). A Course in Simulation, MacMillan, New York.
Rubinstein, R. and Mohamed, B. (1998). Modern Simulation and Modeling,
Wiley, New York.
Schmeiser, B. (1990). Simulation Methods, in Stochastic Models (Heyman
and Sobel eds), North Holland, Amsterdam.
Schrage, L. (1979). A more portable FORTRAN random number generator,
ACM Transactions on Mathematical Software, 5, 132-138.
Spiegelhalter, D., Thomas, A., Best, N. and Gilks, W. (1994). BUGS:
Bayesian inference using Gibbs sampling, version 0.30, MRC Biostatistics Unit,
Cambridge.
Tanner, M.A. (1996). Tools for Statistical Inference: Methods for the Exploration
of Posterior Distributions and Likelihood Functions, 3nd ed.,
Springer, New York.
Whitt, W. (1989). Planning queueing simulations, Management Science,
35, 1341-1366.
Chapter 15
Communication
Tony Greenfield and John Logsdon
You have just finished a project. Why did you start it? Was it just a topic that
grabbed your interest with no other interested people in mind? Did you come across a
problem, either published or in conversation with colleagues, on which you believed
you could cast more light? Did some colleagues offer you a problem, on which they
had been working, in their belief that you could take the solution further forward? Or
was it a problem that arose in a business or industrial context and for which managers
sought your help? Will the solution benefit a small special interest group, or will it be
of wider interest: to the whole industry, to the general public, to financiers, to
government?
The reason for your project’s existence and its origin will determine the style, the
length, the medium for its onward communication. Perhaps your purpose will specify
several styles, several lengths and several media for the promotion of your work.
Your first thought on completing a project may be of a peer-reviewed paper in an
academic journal: a publication to augment your CV, to tell other workers in your
field that you are with them, that you are as clever. But, in this report, we are not so
concerned with personal career enhancement as promoting the vision of ENBIS:
To promote the widespread use of sound science driven, applied statistical
methods in European business and industry.
Different styles and media are required for:
1. News items for popular newspapers
2. Feature articles for popular newspapers
3. News items for technical magazines
4. Feature articles for technical magazines
5. Articles for company house journals
6. Internal technical reports for companies
7. Promotional material for companies
8. Documentary radio scripts
9. Documentary television films
10. Company internal memoranda
11. Company training
12. Training course for wider industrial audiences
13. Short seminars
14. Public lectures
15. Posters for conferences
16. Platform presentations for conferences
17. Web pages
Each of these requires specific details in terms of the written and spoken word and of
printed and projected visual images, of 3-D models, of computer programs.
Throughout, it is vital to keep to the fore the questions:

Why was this work done?

For whom was it done?

To whom do you want to communicate information about the work?

Why would they be interested?

What information for what audiences?

Who may be the beneficiaries of the work?
It is also important to keep communication in mind from the start of a project:

Who originated it?

What exchange, style and content of memoranda were needed to clarify
the purpose of the project?

What communication measures were needed to establish high quality and
timely data collection?

What support was needed from colleagues or specialists?

What progress memoranda and reports were written and for whom
References
Altman D G, Gore S M, Gardner M J, Pocock S J, (1983), ‘Statistical guidelines for
contributors to medical journals’. BMJ, 286, 1489-1493.
Good guidance for presentation of statistics in all subjects, not only medicine.
Barrass R (1978) Scientists must write. Chapman and Hall, London.
Blamires H (2000) The Penguin guide to Plain English. Penguin , London.
Cooper B M (1975) Writing Technical Reports. Penguin , London.
Cresswell J (2000) The Penguin Dictionary of Clichés.
Penguin , London.
Ehrenberg A S C (1982) A Primer in Data Reduction. Wiley, London.
Guidance for presentation of data in tables.
Finney D J (1995), Statistical science and effective scientific communication. Journal of
Applied Statistics, Vol.22 (2), pp 193-308.
O’Connor M, Woodford F P (1978) Writing Scientific Papers in English, Pitman Medical,
London.
Kirkman J (1992) Good Style: Writing for Science. E & FN Spon, London.
Lindsay D (1997) A Guide to Scientific Writing. Longman, Melbourne.
Partridge E, (1962) A Dictionary of Clichés. Routledge and Keegan Paul , London
Tufte E R, (1997) Visual Explanations, Graphics Press, Connecticut.
Graphics for quantities, evidence and narrative.
Pocket Style Book. The Economist, London.
Guide to English Style and Usage. The Times, London.
Chapter 16
Summary, cross-fertilisation and future directions
Shirley Coleman and Tony Fouweather
Summary and Conclusion
The pro-ENBIS project has been enjoyable and energising. Experts from many fields,
institutions and locations have offered their knowledge of modern business and
industrial statistics to give a concise and easily readable summary of the current state
of their science.
Future directions
Following pro-ENBIS, the partners and members are in an excellent position to direct
their shared knowledge and vast experience towards improvements in the theory and
applications of statistics. Traditional barriers between many specialties have been
responsible for slow development in those specialties. The best example was the slow
uptake of agricultural-style designed experiments in manufacturing industries.
Parochial prejudices in different sectors are evident. For example, in the nonmanufacturing sector, designed experiments are called conjoint analysis.
Nomenclature is completely different. Their ‘part-worths’, for instance, are effectively
the same as ‘factor effects’ (van Lottum, 2003). The exploration of established
methods of experimental design in non-manufacturing and service sectors could
reveal powerful techniques for marketing, retail and other commercial processes.
There are many other examples where different sectors have different notations for
similar analysis methods. For example, semi-variograms used by soil scientists can be
just as useful in survey sampling in other specialties and in situations where time
replaces distance as the population dimension (Coleman et al, 2001). The crossfertilisation of these ideas will be very fruitful and will be aided by the close
partnerships developed during the pro-ENBIS project.
Applying the wider ideas of statistical thinking from industrial statistics to
service industries including health
Business and industry are increasingly aware of the benefits of statistical methods;
continuous improvement through six-sigma is popular; the opportunities to apply
these methods have become more common. Many people in service industries, as well
as in manufacturing industries, now recognise the huge benefits that can be gained.
These methods apply equally well to service industries as to those industries, such as
steel production, where they have been used for a long time.
The DMAIC method, promoted in six-sigma, is a logical procedure for using
statistical methods and tools. It allows any process, such as the issuing of invoices,
or the discharge of patients from a hospital, to be investigated and improved in the
same way as an industrial production line. Any process in a service industry can be
improved and reap the same benefits as a process in a factory. The DMAIC scheme
logically addresses any problem by first defining what the problem is, deciding on
what measure will adequately describe the problem and then measuring the extent of
the problem. The next logical steps are to analyse the results and to improve the
process before finally controlling the process to hold any gains.
Over recent years there has been an increasing readiness to use quantitative methods
to solve problems in new application areas as well as in the more traditional areas.
The benefits are becoming apparent to practitioners in many areas where these
methods have not previously been tried, such as in the financial markets.
Companies now realise the importance of good data collection, database management
and data mining methods to monitor their businesses and to improve efficiency and
hence to increase profits. Programmers realise the power of company data in the
application of six sigma methodology to continuous improvement within their
companies. They are writing clever software to handle the data. Statistical software is
becoming more affordable especially for SMEs. Many SMEs now realise the benefits
of having in house six sigma practitioners, trained to use the statistical techniques and
software. Companies are starting to see that the cost to train a six sigma black belt
will save them money in the long term. A company’s own in house statistical
specialist will reduce, or even obviate, the need to hire a consultant. This assumes that
the six sigma black belt will have a sound understanding of statistics. Standardisation
of six sigma curricula was an area of great interest in pro-ENBIS and the discussion
will continue in Enbis.
Statistics is becoming more and more mainstream as businesses see the benefits and
opportunities that it brings. Even the use of basic level statistics, such as charts, to
illustrate performance or down time can have a dramatic effect on the morale of the
workforce and on the efficiency of the process. The democratisation of statistics,
which is enhanced with six sigma and other quality improvement initiatives, is good
in that it raises awareness of the power of numbers and quantitative analysis. On the
other hand, it is dangerous in that the finer detail may be overlooked and this can have
serious consequences in some, perhaps rare, circumstances (Burnham, 2004)
The importance of any improvement in a business is vital in the face of increasing
competition. Any way to sharpen the competitive edge can make the difference
between survival and closure for the company. Statistical techniques can give this
edge to a company. A statistical practitioner who applies the techniques rigorously
can make a process more efficient so that the company will beat its competition on
such things as product quality, production costs and delivery times.
The health sector has great potential for making deeper use of statistical thinking.
In the UK, the National Health Service (NHS) is Britain’s largest employer. It
promotes scientific research at all levels but has been slow to take up modern
statistical methods for quality improvement. There is now a new initiative to promote
statistical thinking. The problems of quantifying quality are particularly relevant in
this sector. Some six sigma projects have shown the benefits of logical problem
solving techniques but also show the knock-on effects of making changes. For
example, projects aimed at reducing the length of stay in hospital have the knock-on
effect of more need for health services outside of the hospital and recently the ‘hotbedding’ resulting from shorter stays in hospital has been accused of increasing cases
of MRSA infection.
Interest in SPC in the NHS is strong at the moment, for example Regional Strategic
Health Authorities are active in providing seminars to review the possibilities for SPC
in the NHS both as part of a six sigma campaign and as a stand-alone skill set. There
are vast and complex opportunities for data analysis and problem solving in the health
sector and as health care is needed in all European countries, there is ample scope for
joint projects across the pro-ENBIS partnership.
pro-ENBIS has been a great support for statistical practitioners throughout Europe.
Contributions are invited to this state of the art report and can be sent to the proENBIS co-ordinators ([email protected] and [email protected])
and via the Enbis discussion forum.
References
Burnham, S. C. (2004) ‘Democratization of statistics’, dissertation for MSc University
of Newcastle, supervisor S.Y.Coleman, passed with Distinction.
Coleman, S.Y., C. Faucherand and A. Robinson (2001) ‘Pipeline inspection with
sparse sampling’, poster presented at RSS Spatial Statistics conference in Glasgow
Van Lottum, C.E. (2003) ‘Opportunities for the application of quantitative
management and industrial statistics’, dissertation for MSc University of Newcastle,
supervisor S.Y.Coleman, passed with Distinction.