ABSTRACTS - June 3-6, 2008

Wednesday, June 4
10:30 AM – 12:15 PM
Invited Plenary Session 2 – Joint Work with George Box

Speaker: George Tiao, University of Chicago
Title: George Box and the Development of Statistics Program at Wisconsin

Abstract: TBD


Speaker: John S (Stu) Hunter, Princeton University
Title: Early exciting days with George Box

Abstract: Shortly after his paradigm shattering paper introducing response surface experimental designs (JRSS,B (1951)) George Box was invited by Gertrude Cox to spend a year as a research professor at the Institute of Statistics at NC State. In his letter of acceptance he wrote that he was interested not only in experimental design but in testing the “robust” nature of the F and t tests, an interest that naturally flowed from some earlier work he published in Biometrika (1941). In January 1953 George Box arrived in Raleigh and I and Sigurd Andersen were assigned to him as graduate students. This brief talk hopes to capture some of the spirit and joy of that wonderful year.

Wednesday, June 4
1:30 – 3:00 PM
(ICS1) George Box’s Contributions to to Quality
Soren Bisgaard, University of Massachusetts Amherst

Abstract: Almost all of George Box’s career has been devoted if not directly certainly indirectly to industrial applications of statistics and quality improvement. Although he has contributed path breaking research to diverse areas of statistical theory such a robustness, design of experiments, nonlinear model building and estimation, time series analysis and process control, Bayesian inference, statistical computing and the philosophy of science and the role of statistics herein, there is behind these eclectic contributions a striking unity, unifying approach and unwavering principle to all his creative output that will be outlined in this presentation. A portion of the presentation will be a video recorded interview with George Box followed by a discussion.

Title: TBD
Speaker: TBD
Abstract:

Title: TBD
Speaker: TBD
Abstract: TBD

Title: TBD
Speaker: TBD
Abstract: TBD

Wednesday, June 4
1:30 – 3:00 PM
(ICS2) Robust Parameter Design After 25 Years
Jeff Wu, Georgia Tech

Title: Simulation on Demand
Speaker: Bruce Ankenman, Northwestern University

Abstract: For years the operations research community has worked on making manufacturing/distribution/service system simulations faster or more statistically efficient. However, in many contexts, simulation time is plentiful, but decision-maker time is scarce. The philosophy of “Simulation on Demand” is to use intensive simulation up front to build a high quality response surface model that will allow for very fast, but very precise, queries throughout the factor space without the need for additional simulation runs. Since the emphasis is on the efficiency of obtaining useful simulation results, rather than on the efficiency of the simulation run itself, this is a design of experiments problem for building the response surface model to a specified precision through simulation experiments. An example of a semiconductor fabrication plant will be used to show how, “Simulation on Demand” represents a bridge between the flexibility and accuracy of a simulation model and the insight provided by an analytical queueing model. (Joint work with Barry Nelson, Feng Yang, Jerry Mackulak and John Fowler.)


Title: Robust Design, Modeling and Analysis of Measurement Systems
Speaker: Tirthankar Dasgupta, Harvard University

Abstract: An integrated approach for estimation and reduction of measurement variation (and its components) through a single parameter design experiment is developed. Systems with a linear signal-response relationship are considered. The noise factors are classified into a few distinct categories based on their impact on the measurement system. A random coefficients model that accounts for the effect of control factors and each category of noise factors on the signal-response relationship is proposed. Two different data analysis strategies - response function modeling (RFM) an performance measure modeling (PMM) are proposed and compared. The effectiveness of the proposed method is demonstrated with a simulation study and Taguchi's drive shaft experiment. (joint with Arden Miller and Jeff Wu)

Title: Bayesian Optimal Single Arrays for Robust Parameter Design
Speaker: V. Roshan Joseph, Georgia Institute of Technology

Abstract: It is critical to estimate control-by-noise interactions in robust parameter design. This can be achieved by using a cross array, which is a cross product of a design for control factors and another design for noise factors. However, the total run size of such arrays can be prohibitively large. To reduce the run size, single arrays are proposed in the literature, where a modified effect hierarchy principle is used for the optimal selection of the arrays. In this article, we argue that effect hierarchy is a property of the system and cannot be altered depending on the objective of an experiment. We propose a Bayesian approach to develop single arrays which incorporate the importance of control-by-noise interactions without altering the effect hierarchy. The approach places no restrictions on the number of runs or levels or type of factors and therefore, is very general. The advantages of the proposed approach are illustrated using several examples. (Joint work with Lulu Kang.)

Wednesday, June 4
1:30 – 3:00 PM
(ICS3) Six Sigma in China
Bill Parr, China Europe International Business School

Abstract: Six Sigma has been applied with great publicity and success at a large variety and number of companies, beginning in the United States and spreading to some extent to Europe and Asia. We investigate the extent, success and nature of penetration of Six Sigma methods in China. The focus of the research reported on in this presentation is on the extent (how much activity), the success (how well have typical Six Sigma financial and process goals been accomplished) and the nature of the penetration (how the methodology has been adapted to deal with the unique nature of business in China, ranging from MNCs with activity in China to SOEs).

The results are based on literature surveys and in-depth interviews with executives and other key personnel (including Master Black Belts and Black Belts) involved in Six Sigma initiatives in a set of companies in China, and with key representatives of consulting firms who offer Six Sigma consulting services in China.

We conclude with observations of unique strengths and weaknesses in Six Sigma as it is being introduced in Chinese firms, and with suggestions for strengthening the applications.

Wednesday, June 4
3:15 – 4:45 PM

Invited Plenary Session 3– Presentation of the 2008 Conference Honoree Award, C.F. Jeff Wu, Georgia Tech

Honoree’s Address: Quality Technology in the High-Tech Age
C.F. Jeff Wu, Georgia Tech

Abstract: Traditional work in quality control and improvement had its main impetus from needs in manufacturing like autos, electronics and chemicals. As the economies in developed countries are increasingly dependent on high value-added products, it is natural to ask what role can quality professionals play in the high-tech age. Without the benefit of crystal gazing, I will use several research examples to search for hints to the future. I would describe the essence of some ongoing research projects, including the robust growth conditions of nano-materials, study of nano-mechanical properties, computer-aided design and management of data centers, and electronics packaging. The following features distinguish these problems from traditional quality technology. The experiments are more elaborate and the responses can be sensitive to input conditions; they often require complex modeling; how to scale up from lab conditions to industrial productions; they often use extensive computer modeling (e.g., finite element analysis) before conducting (or even foregoing) physical experiments for verification. Based on these new features, I will try to address the ultimate question: “will there be a paradigm shift and what to expect”?

Thursday, June 5
8:30 – 10:00 AM
(ICS4) Design and analysis of computer Experiments
Doug Montgomery, Arizona State University


Title: Design and Analysis of Computer Experiments
Speakers: Bradley Jones, JMP

Abstract: In an effort to speed the development of new products and processes, many companies are turning to computer simulations to avoid the expense and lost time of building prototypes. These computer simulations are often very complex and may take hours to complete one run. If there are many variables affecting the results of the simulation, then it makes sense to design an experiment to gain the most information possible from a limited number of computer simulation runs and build a response surface type of model, usually called a surrogate model. The absence of noise is the key difference between computer simulation experiments and experiments in the real world. Since there is no variability in the results of computer experiments, optimal designs based on reducing variance have questionable utility. Replication, usually a "good thing", is clearly undesirable in computer experiments. Thus, a new approach to experimentation is necessary. This talk takes a case study approach using computer models to demonstrate new software that supports both computer simulation design and analysis.

Title: Comparing Designs for Computer Simulation Models
Speakers: Rachel T. Johnson, Arizona State University

Abstract: The use of simulation as a modeling and analysis tool is wide spread. Simulation is a powerful tool, which enables the users the ability to experiment virtually on a validated computer environment. Often the output from computer experiments results in a complex response surface. Determining which experimental design should be used is a non trivial task. An evaluation of several computer experiments is made assuming the modeler is interested in fitting a high order polynomial to the response data. Optimal and space-filling designs are explored. Additionally, a new class of hybrid designs are introduced, which can be used for deterministic or stochastic simulation models.

Discussant: Geoff Vining, Virginia Tech

Thursday, June 5
8:30 AM – 10:00 AM
(ICS5) Reliability
William Q. Meeker, Iowa State University


Title: An Algorithm for Computing Approximate Variances and Covariances of Model Parameters from a Repeated Measures Degradation Model with Applications
Speaker: Brian Weaver, Iowa State University

Abstract: Repeated measures degradation studies can be used to assess product or component reliability when there are few or even no failures during the study. This talk will describe an algorithm that can be used to compute approximate large-sample variances and covariances from a mixed effect linear regression model for repeated measures degradation data. The algorithm has several important applications: assessing the effect of sample size (number of units and number of measurements within the units), evaluation of proposed test plan properties to allow the design of statistically efficient test plans, and the computation of approximate confidence intervals for functions of the parameters such as quantiles of the degradation distribution, failure probabilities, and quantiles of the failure time distribution. These applications will be illustrated with examples. The algorithm has been implemented in R and S-PLUS.

This talk is based on joint work with William Q. Meeker and Luis A. Escobar.


Title: Materials Degradation Experiments
Speaker: Joanne Wendelberger, Statistical Sciences Group, Los Alamos National Laboratory

Abstract: Materials degradation experiments are used to investigate the impact of various experimental factors on the degradation of specified response variables as a function of time. Results obtained from studying degradation in a controlled setting provide information that can be used to develop predictions of the performance of materials as they age. Often, degradation experiments involve acceleration factors, such as temperature, that allow the experimenter to investigate degradation processes over a much shorter time span than simply observing the materials over ordinary time. A number of statistical issues arise in the design of materials degradation experiments, including selection of runs, sampling frequency, and choice of acceleration levels. Analysis may involve a variety of types of responses, such as images or functional data. Examples of materials experiments will be used to illustrate the statistical challenges associated with degradation studies.


Title: Reliability Analysis of Accelerated Degradation Data
Speaker: David C. Trindade, SUN Microsystems

Abstract: In accelerated reliability testing, the occurrence time of failures are used to estimate the life distribution for extrapolation to field conditions via an acceleration model. Often, because of time limitations on testing, many if not most of the units undergoing stress do not fail, resulting in censored observations. In certain situations, however, it is possible to do an analysis even without any actual failures, provide that there is a measurable product parameter that is degrading over time towards a value that is defined to be a failure level. In this talk, we will review key concepts in degradation analysis including physical models, assumptions, analysis methods, and acceleration considerations. We will illustrate, using an example in microelectronics, how field reliability projections can be obtained from accelerated life tests. We also will provide some simulation results to reveal factors that can significantly influence the estimation error.

Thursday, June 5
8:30 – 10:00 AM
(ICS6) Extreme Risk Inference and Management
Regina Liu, Rutgers University

Title: Pitting Corrosion: Analysis of Designed Experiments with Extreme Value Distributed Responses
Speaker: Holger RootzŽn, Chalmers University of Technology, Sweden

Abstract: This talk discusses how Extreme Value Statistics can be used to validate and improve designed experiments with extremal responses, and how to extrapolate and compare results.

A main motivation is corrosion tests: Localized, or "pitting", corrosion can limit the usefulness of magnesium and other new lightweight materials aimed at reducing weight, and thus CO2 emissions from cars. It makes judicious choice of alloys and surface treatments necessary. Standard methods to evaluate corrosion test are based on analyzing weight loss by ANOVA. These methods may be misleading in two ways. Usually it is not weight loss but the risk of perforation, i.e. the depth of the deepest pit which is of interest. Further the standard ANOVA assumptions of normality and homogeneity of variances typically are not satisfied by pit depth measurements, and don't give credible extrapolation into extreme tails.

The talk presents a streamlined approach to analysis of such experiments. It consists of 1.) a preliminary analysis to check that the experiment has worked out as intended, 2.) a separate analysis for each individual experimental condition, and 3.) a strategy for pairwise comparisons. Of particular interest is to try to separate situations with many but shallow pits from potentially more dangerous situations with rare but deep pits.

Further, a class of hierarchical random effects models for extreme value data is discussed. The models are obtained by mixing extreme value distributions over positive stable distributions. They open up exciting new possibilities to develop extreme value analogues of normal random effects models.

The big challenge is to develop a full theory of design and analysis of experiments with extreme value distributed responses – the present results are just an early beginning.

Short biography: Hologer RootzŽn is professor of Mathematical Statistics at Chalmers University of Technology, Sweden, and has earlier worked in Lund, Chapel Hill and Copenhagen. He is Editor of Bernoulli, was founding editor of Extremes, is leader of GMMC, the Gothenburg Mathematical Modelling Centre, and is chairman of the Gothenburg Stochastic Centre. He has led joint research projects with a number of companies, including Volvo Car. His research interests include stochastic theory, extreme values, financial risk, and applications in engineering, medicine, and industry.


Title: Thresholding Events of Extreme in Simultaneous Monitoring of Multiple Aviation Risk Indicators
Speaker: Jun Li, University of California, Riverside

Abstract: Risk assessments involving extreme events often encounter settings with very few or no occurrences in reality. Inferences about such extreme events face the problem of insufficient data. Extreme value theory is particularly well suited for handling this type of problems. In this talk, we uses a multivariate extreme value theory approach to establish thresholds for signaling varying levels of extreme in the context of simultaneous monitoring of multiple measurements. The threshold system is well justified in terms of extreme multivariate quantiles, and its sample estimator is shown to be consistent. The proposed approach is applied to developing a threshold system for monitoring airline performances. This threshold system assigns different risk levels to observed airline performances. In particular, it divides the sample space into regions with increasing levels of risk. Moreover, in the univariate case, such a threshold technique can be used to determine a suitable cut-off point on a runway for holding short of landing aircrafts. This cut-off point is chosen to ensure a certain required level of safety when allowing simultaneous operations on two intersecting runways in order to ease air traffic congestion.
This is joint work with John H. J. Einmahl (Tilburg University) and Regina Y. Liu (Rutgers University). Short biography: Jun Li is Assistant Professor of the Department of Statistics at University of California, Riverside. Her research interests include nonparametric multivariate analysis, applications of data depth, and extreme value theory and quality control. She received her Ph.D. in Statistics from Rutgers University. During her graduate study at Rutgers University, she was supported by fellowships from NSF and FAA. She also received Laha Travel Award from the Institute of Mathematical Statistics in 2005.

Title: Estimating Extreme Quantile Regions for Two Dependent Risks
Speaker: John H.J. Einmahl, Tilburg University, The Netherlands

Abstract: When simultaneously monitoring two dependent, positive risks one is often interested in quantile regions with very small probability p. These extreme regions contain hardly or no data and therefore statistical inference is difficult. In particular when we want to protect ourselves against a calamity that has not yet occurred, we consider the case where p is smaller than 1/n, with n the sample size. In practice, p typically ranges from .01 to .0000001.

We consider quantile regions of the type {(x,y): f(x,y)<= c}, where f is the joint density. Such a region has the property that its complement consists of the most likely points, that is, is as small as possible. Using extreme value theory, we construct a natural, nonparametric estimator of such a quantile region and study its behavior.

This is joint work with Laurens de Haan (Rotterdam and Lisbon).

Short biography: John H.J. Einmahl is professor of Statistics at the Department of Econometrics and research fellow at CentER, both at Tilburg University. He has published in leading journals in Statistics and Probability Theory. His main research area is nonparametric statistics, including statistics of extremes, empirical likelihood, generalized quantiles, and empirical processes. He obtained his Ph.D. from Nijmegen University. He is is an Associate Editor of Bernoulli and Extremes. He is an elected fellow of the Institute of Mathematical Statistics, and since 2005 a member of its council. He was a co-director of the Stochastics of Extremes and Risk Analysis program at the European research institute Eurandom (Eindhoven) from 2001-2004. He was an Associate Editor of The Annals of Probability (1999-2003), and of The Annals of Statistics (2001-2004). In 1998 he visited Florida State University as a Senior Fulbright Scholar.

Thursday, June 5
8:30 – 10:00 AM
(CCS1) Screening Designs with Reasonably-Balanced Projections
Tim Kramer, Global Statistical Sciences, Lilly Research Laboratories

Title: Screening Designs with Reasonably-Balanced Projections
Speaker: Tim Kramer, Global Statistical Sciences, Lilly Research Laboratories
Co-Author: David M. Steinberg (Tel Aviv)

Abstract: Screening designs are useful in product development to quickly identify a few factors having relatively large impact on a response from a much larger set of factors. They are also useful in robustness studies to demonstrate that product performance is insensitive to a host of variables within specified ranges. Regular fractions of 2- and 3-level factorial designs are often used for screening designs and tables and properties of these designs are well-understood. This paper introduces a balance metric which is obtained by examining all possible projections of the design into one, two and three dimensions. This loss function is shown to be related to a projectivity measure and entropy. Standard fractional factorial screening designs are shown to be optimally-balanced relative to this measure.

A general strategy for constructing screening designs with reasonably-balanced projections is presented. Some example designs are given, including a design that was employed by the author in the pharmaceutical industry. Analysis approaches are described. SAS code is available to construct designs of arbitrary levels and runs.


Thursday, June 5
10:30 AM – 12:15 PM
(ICS7) DOE Software Trends – Brad Jones, SAS

Title: What Commercial DOE Software Should Do
Speaker: Chris Nachtsheim, University of Minnesota - Minneapolis

Abstract: This presentation advocates a problem-driven approach to the design of experiments and provides a wish-list of software capabilities needed to support this approach. We make the case that existing books and software take a design-driven approach to the design of experiments, to the detriment of most students and users. With the design-driven approach, practitioners are exposed to the properties of various classes of designs, and are then taught to navigate through these classes to find the design that will best meet the goals of the investigation. Frequently, the goals must be compromised because the match is not exact. This approach expects a great deal of the user in terms of DOE expertise. The problem-driven, it is argued, does not require the same level of expertise and will therefore remove or lower a barrier to use. The approach employs the following steps: 1) describe the nature of the factors and responses; 2) describe the experimental objectives; 3) describe the constraints; 4) produce the design that best meets the objectives; 5) do sensitivity analysis. Designs produced may be classical, tabled designs, or computer-generated optimal designs, depending on the problem. The methodology will be illustrated with a series of examples.


Title: Computing, Theory, and Finding Good Designs
Speaker: John P. Morgan, Virginia Tech
Abstract: It is no accident that discrete design has been relatively late in seeing significant benefit from computational tools. Block designs, orthogonal arrays, and related combinatorial structures, save the very smallest cases, are notoriously difficult to enumerate. Statistical optimality arguments that reduce the number of competing structures are critical: they can help bring design problems within computational reach. The same is true of mathematical work on design isomorphism. This talk discusses the interplay of such arguments with recently developed software in creating design catalogs, focusing on various types of optimal block designs.


Title: Algorithms to Search for Optimal Designs
Speaker: William Li, University of Minnesota – Minneapolis

Abstract: A general algorithmic approach to the construction of optimal designs is introduced. The algorithms use a columnwise exchange procedure, which can be particularly useful when certain designs structures need to be retained during the search for optimal designs. The efficacy and efficiency of the algorithms are demonstrated by their applications in several areas that include optimal balanced designs, mixed-level designs, Latin-hypercubes, and design repair problems. Some comparisons between the columnwise algorithms and other competing algorithms are obtained.

Thursday, June 5
10:30 AM – 12:15 PM
(ICS8) Environmental Aspects of Reliability
Reinaldo Gonzales, General Electric Co.


Title: Implications of Environmental Requirements (Pb-free) on the Reliability and Testing of Flip-Chip Solder Joints in Packaged IC Chips
Speaker: Timothy D. Sullivan, IBM Microelectronics
Authors: Timothy D. Sullivan, Thomas Wassick, Charles Goldsmith, et al., IBM Microelectronics

Abstract: The environmentally mandated industry migration from Pb-based solder to lead-free solder requires reliability evaluation of new solder joint configurations. Other than being lead-free, properties of the preferred substitute solder (Sn) are generally inferior for the solder bump application, so that candidate alloys and configurations intended for this application require extensive testing. Two of the more troublesome properties are solder bump stiffness and electromigration resistance, both of which are detrimental to solder joint reliability. Solder bump stiffness becomes a problem when packaged parts experience temperature changes. Electromigration (EM) is the movement of metal atoms in response to an applied current. This movement produces voids in the solder joint that increase the joint resistance and eventually lead to chip failure. More attention has been paid to EM recently because the industry is moving to denser chips with higher power. At the same time that the joint size is decreasing, the industry is also moving to Sn based solders which exhibit different behavior than does Pb, primarily because of the asymmetry of the tetragonal crystal structure. Because of limitations on the acceleration of such testing, considerable time and expense are necessary to define the reliability of the lead-free joints. Due to this expense, there are practical limitations on the degree to which solder bump reliability can be determined in the near future, and this requires a conservative limit to current and temperature to be applied to operating conditions.


Title: Synergizing Reliability Engineering Tools
Speaker: Bill Denson, Corning Inc.

Abstract: There are many tools and techniques used by reliability engineering practitioners in the execution of reliability programs. Some of these tools include: Quality Function Deployment (QFD), Failure Mode and Effects Analysis (FMEA), Design of Experiments (DOE), Test Planning, Accelerated Testing, and Life modeling. Each of these has a specific purpose, and is often accomplished with different teams within a project. As a result, they are often executed in an autonomous fashion, with little or no leveraging amongst the various tools. While each of these tools accomplishes a specific goal with different approaches, there is a significant amount of synergy that exists if they are treated as a “system” of tools, and are leveraged to the maximum extent. This presentation will discuss the manner in which these tools can be linked, resulting in a more effective and efficient reliability program.


Title: Accelerated Testing: The Catalyst Converging Physics And Reliability Statistics
Speaker: Larry Edson, Hobbs Engineering Corp

Abstract: Accelerated test methods are essential in developing product maturity. The market place is full of highly competitive companies prepared to offer the latest technologies by the time you wake up tomorrow. Consumers fuel this cadence with demands of; “I want it all… and I want it now!” Engineers and statisticians work diligently to fulfill these demands with reliable products, but find it difficult to agree upon test methods that are both effective in proving reliability and can do so quickly using small sample sizes. The “physics of failure” approach to reliability was the first major step toward finding solutions to this problem and has formed a strong synergistic relationship with the world of accelerated testing.

The Calibrated Accelerated Life Test method (CALT) is a product of today’s synergy between the physical sciences and statistics while building upon the legendary work of Wayne Nelson and William Meeker. The CALT method is based upon the concept of a life-stress relationship, which covers a wide range of phenomenon in the world of physics and chemistry. Unique to this method is the ability to select the stress levels that will ideally allow the test process to occur within the time window required by the engineer.

This discussion will provide the rational behind this methodology through examples from the automotive industry. Additionally, the concept of Degradation Analysis applied within the CALT framework will be explained with examples shown.

The CALT methodology decouples time and sample size from the reliability requirement. This simple attribute is exceptionally powerful and useful in today’s world of product development.

Title: Chip-Scale Package Reliability Study for Medical Electronics
Speaker: James E. Simpson, Micro and Nano Structures Technologies, GE Global Research
Authors: James E. Simpson and James W. Rose

Abstract: The intermediate and long-term vision for medical imaging systems is for continuing increases in system-level performance and functionality combined with decreasing cost and size. Improvements in image quality, in particular, increased spatial and temporal resolution to allow clinicians to make better diagnoses, leading to better patient outcomes. These requirements, in turn, flow down to increasing channel densities for the sensitive signal electronics hardware inside these imaging systems. This densification requires novel electronics packaging solutions. Present-day imaging systems may have several hundred thousand channels combined with a system-level 10-year life requirement.

One technique used to increase channel densities is through the widespread use of Chip-Scale Packages (CSPs) for sensitive mixed-signal integrated circuits. In these packages, the die is mounted on an interposer which makes electrical connections to a Printed Circuit Board (PCB) through an array of solder balls. The most common failure mechanism for properly designed CSP assemblies fail in low-cycle fatigue at the solder joint.

We present the results from a CSP reliability study. The primary factors influencing reliability are the solder ball pitch, contact metallurgy and solder ball location. To determine reliability, a series of accelerated life tests under cyclic thermal stress were performed. Two different pitches were tested (0.8 and 1.0 mm), along with two different contact materials (Cu/OSP and Ni/Au). Solder ball location was a key noise parameter. Test assemblies were specially designed to emulate the structure of the final assembly and allow for testing of the critical variables. The assembly consisted of an array of CSPs with a “daisy-chain” interconnect pattern solder attached to a PCB substrate with controlled contact metallurgy and structures to allow easy electrical testing. The accelerated test protocol was air-to-air thermal cycling from 0C to 100C over several thousand cycles. Results from the accelerated life testing indicated that all the configurations investigated should exceed the 10-year system life requirement.

Thursday, June 5
10:30 AM – 12:15 PM
(ICS9) New Methods for Network and Recurrent Event Analysis
Ta-Hsin Li, IBM Research

Title: Optimal Designs for Network Traffic Measurement Schemes
Speaker: George Michailidis, University of Michigan

Abstract: Computer network traffic can be sampled at routers to determine the origin and destination of different packets. This is in turn used to estimate the volume of traffic between all origin-destination pairs. Use of higher sampling rate leads to better estimates but uses more resources. In this work we formulate the problem of determining the network-wide sampling rates as an optimal design problem. This problem is solved using semi-definite programming. The proposed approach is evaluated on a number of real and synthetic data sets.

Title: Analysis of Window-Observation Recurrence Data
Speaker: Huaiqing Wu, Iowa State University

Abstract: Many systems experience recurrent events. Recurrence data are collected to analyze quantities of interest, such as the mean cumulative number of events. Methods of analysis are available for recurrence data with left and/or right censoring. Due to practical constraints, however, recurrence data are sometimes recorded only in windows. Between the windows, there are gaps over which the process cannot be observed. We extend existing statistical methods, both nonparametric and parametric, to window-observation recurrence data. The nonparametric estimator requires minimum assumptions, but will be inconsistent if the size of the risk set is not positive over the entire period of interest. There is no such difficulty when using a parametric model for the recurrence data. For cases in which the size of the risk set is zero for some periods of time, we propose and compare two alternative hybrid estimators. The methods are illustrated with two example applications. (This is joint work with Jianying Zuo and William Q. Meeker).

Title: Maximum Likelihood Estimators of Clock Offset and Skew under Exponential Delays
Speakers: Jun Li and Daniel R. Jeske, University of California at Riverside

Abstract: Accurate clock synchronization is essential for many data network applications. Various algorithms for synchronizing clocks rely on estimators of the offset and skew parameters that describe the relation between times measured by two different clocks. Maximum likelihood estimation (MLE) of these parameters has previously been considered under the assumption of symmetric and exponentially distributed network delays. We derive the MLEs under the more common case of asymmetric exponentially distributed network delays and compare their mean squared error properties to a recently proposed alternative estimator. We investigate the robustness of the derived MLE to the assumption of non-exponential network delays, and demonstrate the effectiveness of a bootstrap bias-correction technique.

Thursday, June 5
1:30 – 3:00 PM
(ICS10) Teaching and Training of Statistics
Stefan Steiner, Dept. of Statistics and Actuarial Science, University of Waterloo

Title: Translating Understanding of a Complex System into a Statistical Model
Speaker: Christine M. Anderson-Cook, Statistical Sciences Group, Los Alamos National Laboratory

Abstract: Developing appropriate statistical models for system reliability of complex systems involve many diverse aspects and expertise from a variety of different experts. Understanding the system structure typically requires the expertise of several engineers, while building a model that can incorporate multiple types of data involves advanced statistical thinking. Through a case study, the talk describes the process and some of the potential pitfalls for eliciting information from collaborators with a variety of backgrounds, synthesizing different types of engineering and statistical knowledge, and verifying that the final statistical model reflects actual system characteristics.


Title: A Simple Experimental Scenario for Teaching DOE
Speaker: Scott Kowalski, Minitab Inc.

Abstract: Classroom discussion is an effective way for students to interact and learn. A class on design of experiments usually covers many related design strategies. It can be useful to have one running example (similar to a case study) that can be used throughout the class to extend earlier topics to later more complex topics. This talk presents one such running example that studies plant growth, mostly in a greenhouse. The greenhouse experiment begins with a simple t-test and ends with response surface methodology. By making small changes to the scenario, most of the standard topics covered in a one-semester design of experiments class are discussed.


Title: Teaching Applied Statistics Using a Virtual Manufacturing Process
Speaker: Stefan Steiner, Dept. of Statistics and Actuarial Science, University of Waterloo

Abstract: This non technical talk describes an innovative and successful use of technology, through a virtual process, to aid in the teaching of statistical concepts and methodology. The virtual process simulates a manufacturing process for automobile camshafts that has a number of processing steps and many inputs.

At the start of an upper year undergraduate course Stat 435/835: Statistical Methods for Process Improvement, each team of students is given a budget and assigned the task of reducing variation in a critical output characteristic of a different version of the virtual process. Throughout the term, the teams plan and analyze a series of process investigations (~1/week) to first learn about how their process works and, by the end of term, how to improve it. The teams interact with the virtual process through a web interface. Each team submits a weekly written report describing their recent progress and twice per term presents to the class at a “management review meeting.” The virtual process is also used as the context for all midterms and exams. Based on anecdotal evidence and survey results, students find interacting with the virtual process fun, stimulating and challenging.

The goals of this talk are to show how the virtual process aids in the teaching of material and concepts in Stat 435/835 and to describe its main pedagogical benefits. With thought and some adaptation something similar should be possible for other applied statistics courses.

Stefan Steiner is an Associate Professor in the Statistics and Actuarial Science department, and director of the Business and Industrial Statistics Research Group (BISRG) at the University of Waterloo.

Thursday, June 5
1:30 – 3:00 PM
(ICS11) Experimental Design Techniques for Optimization
Will Guthrie, National Institute of Standards and Technology

Title: Analysis of Optimization Experiments
Speaker: James Delaney, Carnegie Mellon University

Abstract: The typical practice for analyzing industrial experiments is to identify statistically significant effects with a 5% level of significance and then to optimize the model containing only those effects. In this article, we illustrate the danger in utilizing this approach. We propose methodology using the practical significance level, which is a quantity that a practitioner can easily specify. We also propose utilizing empirical Bayes estimation which accounts for the randomness in the observations. Interestingly, the mechanics of statistical testing can be viewed as an approximation to empirical Bayes estimation, but with a significance level in the range of 15--40%. We also establish the connections of our approach with a less known but intriguing technique proposed by Taguchi. A real example and simulations are used to demonstrate the advantages of the proposed methodology. (Joint work with V. Roshan Joseph)

Title: Adaptive One Factor at a Time Experiments: Extensions of Theory and Practice
Speaker: Dan Frey, Massachusetts Institute of Technology

Abstract: This talk will review a number of recent advancements related to adaptive-One-Factor-at-a-Time (aOFAT) experimentation. Previously, our research group showed that aOFAT provides high expected values of improvements and high probabilities of exploiting two-factor interactions, especially when these interactions are large. More recently, we have investigated the properties of adaptive experimentation from a Bayesian perspective. This enables one to determine the degree to which aOFAT represents an efficient response to step-by-step accrual of sample information and helps one assess aOFAT in comparison to "greedy" algorithms. We also extend the aOFAT approach by incorporating concepts from "ensemble methods." In the approach, the experimenter plans and executes multiple adaptive experiments on the same system with differences in starting points and orders in which factors are varied. Decisions are made by aggregating the multiple outcomes of aOFAT experiments. This approach is compared to alternative methods using both a case study to illustrate the approach and a hierarchical probability model to seek more general conclusions.


Title: Experimental Design in the Scheduling of Two Unrelated Parallel Processors
Speaker: Dennis Leber, Statistical Engineering Division, National Institute of Standards & Technology

Abstract: Given a group of tasks and two non-identical processors with the ability to complete each task, which tasks should be assigned to which processor in order to complete the given group of tasks in as short amount of time as possible? Processors may be airport runways, shipping trucks, manufacturing lines, computer processors, or surgeons. The tasks may be airplane take-offs and landings, packages to be shipped, products to be built, computer codes to be run, or patients to be operated on. This problem has been formalized in the scheduling literature as the minimization of the makespan (time required to complete all tasks) for two unrelated parallel processors.

One possible approach to solving this problem is to create a computer simulation model of the process and apply this model to all possible task assignments (complete enumeration), selecting the assignment that produces the minimum makespan. A typical implementation of the complete enumeration approach is to calculate the makespan for each assignment schedule one by one, noting the minimum makespan and associated assignment schedule observed thus far. This talk discusses the benefit realized by implementing the complete enumeration approach using a 2k full factorial experimental design framework, as illustrated by a printed circuit board assembly example. Also explored is the possibility of employing 2k - p fractional factorial experimental designs in the solution of the two unrelated parallel processors problem.

Thursday, June 5
1:30 – 3:00 PM
(ICS12) The Books of Madison, Wisconsin - Jim Alloway

Madison, Wisconsin, the host city for the 2008 Quality and Productivity Research Conference, is closely linked with several seminal texts on productivity and quality. This session brings three authors together to discuss the enabling factors that this city provided to develop what can be viewed as the fundamental tools for improving quality and productivity, as well as their individual works.

Brain Joiner’s “Fourth Generation Management” provides the organizational framework to realize quality gains. Peter Scholtes’ books, “The Team Handbook” and “The Leader’s Handbook”, provide the tools and philosophy necessary to run and manage effective improvement teams. J. Stuart Hunter’s “Statistics for Experimenters” provides the statistical engine to extract the maximum amount of information from the data collected by the teams. The work of all three authors fit together to drive overall improvement.

Speaker: J. Stuart Hunter, coauthor of “Statistics for Experimenters”.
Abstract: Regarded as the bible for design of experiments, the text presents the theoretical and applied aspects of statistically designed experiments, a fundamental tool for effective data collection and analysis.

Speaker: Brian L. Joiner, author of “Fourth Generation Management”.
Abstract: This book follows the evolution of management and quality, presenting what upper management needs to do to prepare the organization to realize the gains of improved quality and productivity.

Speaker: Peter R. Scholtes, author of “The Team Handbook” and “The Leader’s Handbook”.
Abstract: These books address the “how to” tools needed by quality team members, as well as their managers to achieve.

Thursday, June 5
3:15 – 4:45 PM
(ICS13) Statistical Methods in Reliability
Martha Gardner, General Electric Co.


Title: Analysis of Accelerated Life Tests with Random Effects
Speaker: Ramon Leon, University of Tennessee – Knoxville

Abstract: We show how to use Bayesian methods to make inferences from an accelerated life test where test units come from different groups (such as batches) and the group effect is random and significant both statistically and practically. Our approach can handle multiple random effects and several accelerating factors. We illustrate the method with an application concerning pressure vessels wrapped in Kevlar fibers, where the fiber of each vessel comes from a single spool and the spool effect is random. Bayesian analysis using Markov chain Monte Carlo (MCMC) methods is used to answer questions interest in accelerated life tests with random effects that are not as easily answered with more traditional frequentist methods. For example, we can predict the lifetime of a pressure vessel wound with a Kevlar fiber either from a spool used in the accelerated life test or from another random spool from the population of spools. We comment on the implications that this analysis has on the estimates of reliability (and for the Space Shuttle, which has a system of 22 such pressure vessels. Our approach is implemented freely available WinBUGS software so that readers can apply the method to their own data.


Title: Risk, Reliability, and Repair
Speaker: Brock Osborn, General Electric Co.

Abstract: Reliability analytics are important not only for determining when a part or complex system will fail but also in developing an overall strategy for determining on-going operational costs, quantifying financial risk, and developing a comprehensive maintenance and repair strategy. In this talk we will discuss various reliability and repair models and will provide examples from industry of how these techniques are used to quantify the risk associated with customer service agreements.


Title: Monitoring Reliability Data
Speaker: Emmanuel Yashchin, IBM Research Division

Abstract: We consider the problem of monitoring parameters of a lifetime distribution based on the sequence of life tests with censoring. This problem arises in a number of settings, for example in administration and interpretation of Ongoing Reliability Tests (ORT) data or warranty data. The main technical difficulty of the monitoring problem is related to the fact that information obtained at each point in time affects data for every life test under observation. This feature necessitates use of special methods and special metrics for their evaluation. In this paper we describe a new approach to monitoring such data streams and give examples related to detection of unfavorable changes in warranty data in the context of a large scale system for warranty data analysis.

Thursday, June 5
3:15 – 4:45 PM
(ICS14) Design, Analysis and Utilization of Complex Computer Models
Peter Qian, University of Wisconsin-Madison

Title: Calibrating a Computer Code in the Presence of Systematic Discrepancy
Speaker: Brian Williams, Los Alamos National Laboratory
Authors: Jason Loeppky, (UBC-O), William Welch (UBC), and Brian Williams (LANL)

Abstract: Computer models to simulate physical phenomena are now widely available in engineering and science. Before relying on a computer model, a natural first step is often to compare its output with physical or field data, to assess whether the computer model reliably represents the real world. Field data, when available, can also be used to calibrate unknown parameters in the computer model. Calibration can be particularly problematic in the presence of systematic discrepancies between the computer model and field observations. In this talk we present results on a simulation study that is designed to assess how well the calibration parameter has been estimated, and the conditions under which calibration is possible. By simulating both computer model data, and physical observations from a Gaussian process the uncertainty due to using the incorrect model does not arise. This allows us a more accurate picture of the problems that can arise when attempting to calibrate the model in the presence of systematic discrepancy.


Title: A Level-Set-Based Multistage Metamodeling Approach for Design Optimization
Speaker: Wei Chen, Northwestern University
Authors: Shikui Chen, Wei Chen (Northwestern University)

Abstract: A level-set-based multistage metamodeling approach is proposed for design optimization. Different from the existing objective-oriented sequential sampling methods, where design objective and constraints have to be combined into a single response of interest, our method offers the flexibility of building metamodels for multiple responses (design objective and constraints) simultaneously. Uncertainty quantification (UQ) is introduced for each metamodel to represent the confidence interval due to the lack of sufficient samples in the early stages of metamodeling. Based on the extreme values of the optimal solution identified, the level set method (LSM), together with a series of Boolean operations, are used to identify the region(s) of interests with arbitrary topology. Introducing LSM facilitates region manipulations, detection of disconnected regions of interest, and visualization for design exploration. The proposed method possesses a superior efficiency in design exploration to conventional sequential sampling strategy and allows the use of multiple samples in each sampling stage. The next-stage metamodel is adaptively fit based on the pre-existing and newly introduced sample points. As the metamodeling process moves on, the region of interests is progressively reduced, and the optimal design is asymptotically approached. The proposed approach is demonstrated with mathematical benchmark examples and an engineering design problem. The results are compared with those from one-stage metamodeling using the Optimal Latin Hypercubes (OLH) experiments. Our results show that the proposed method can effectively capture the region of interests with any shape and topology and locate the optimum design with improved efficiency.


Title: Evaluating and Improving the Usefulness of Computational Models
Speaker: Genetha Anne Gray, Sandia National Laboratories
Author: Genetha Anne Gray, Sandia National Laboratories

Abstract: Despite their obvious advantages, the incorporation of computer simulations in the study of complex science and engineering systems introduces many challenges. To justify the inclusion of information obtained from numerical models, some level of predictive capability of the corresponding simulation codes must be guaranteed. In addition, the degree to which a computer model accurately represents the real world phenomena it seeks to represent must be quantified. Moreover, calibration techniques should be applied to both improve the predictive capability of the model and to curtail the loss of information caused by using a numerical model instead of the actual system.

Evaluating and calibrating a computer model is dependent on three main components: experimental data, model parameters, and algorithmic techniques. Data is critical as it defines reality, and inadequate data can render model evaluation procedures useless. Model parameters should be studied to determine which affect the predictive capabilities of model and which do not. Those that do are subject to calibration. Techniques for model evaluation and calibration should both sufficiently sample the design space and limit the computational burden.

In this talk, we will discuss the problems inherent in model calibration and validation processes in terms of data, parameters and algorithms. In particular, we will focus on suitable techniques of optimization and statistics for some specific numerical models from the areas of electrical engineering, medicine, and groundwater control. We will demonstrate some successful and some not so successful approaches.

Thursday, June 5
3:15 – 4:45 PM
(ICS15) Special INFORMS Invited Session: Fault Monitoring and Diagnosis for Complex Systems
Shiyu Zhou, Associate Professor, Department of Industrial and Systems Engineering, University of Wisconsin – Madison

Title: Diagnosis of Process and Sensor Faults in Manufacturing Processes
Speaker: Shan Li and Yong Chen, Department of Mechanical and Industrial Engineering, University of Iowa

Abstract: This paper presents an approach to identify process mean shift faults and sensor mean shift faults simultaneously in manufacturing processes based on Bayesian variable selection techniques. The potential process faults and sensor faults always outnumbers the number of sensor measurements. The proposed approach utilizes the fact that only a few process and sensor faults can occur simultaneously. A guideline on how to select the values of hyperparameters in the Bayesian variable selection is also given. Several simulated examples and a real assembly example are studied to demonstrate the performance of the developed diagnosis method. A sensitivity analysis is conducted for the assembly example, which shows that the proposed approach is a robust procedure.

Title: Multiple Fault Signature Integration and Enhancing for Variation Source Identification in Manufacturing Processes
Speakers: Li Zeng, Nong Jin, and Shiyu Zhou, Department of Industrial and Systems Engineering, University of Wisconsin-Madison

Abstract: Recently, signature matching as a popular method for variation source identification in manufacturing processes has drawn significant attention. In this method, the variation source is identified through matching the variation patterns of specific process faults, also called fault signatures, with the variation patterns in the newly collected quality data. There are situations that a fault occurred multiple times before and consequently multiple signatures exist for the same fault. A technique is proposed in this article which can integrate multiple signatures together to enhance the accuracy of variation source identification. A linearly combined fault signature is constructed to increase the detection power of the fault identification. A numerical study is also presented to validate the effectiveness and robustness of the proposed method.


Title: Classification Methods for Highly Imbalanced Class Sizes in Warranty Data
Speakers: Eunshin Byon, Abhishek K. Shrivastava, and Yu Ding, Department of Industrial and Systems Engineering, Texas A&M University

Abstract: Warranty data can play a critical role in quality control. A quantitative relationship between this product feedback information and manufacturing data can be used to predict the product failures, identify the process parameters that cause them, and ultimately make improvements to the process. The main challenge in establishing such a relationship is the small number of faulty units in a lot. This leads to a highly skewed distribution of the response variable, field return, which prevents the well-known off-the-shelf classification methods from performing well. We propose a statistical procedure in such a class-imbalance problem. The proposed procedure is applied to a real dataset from a cell phone manufacturer and it performs significantly better than well known off-the-shelf methods.

Title: Multiscale Wavelet Analysis for Multiple Embedded Operations Monitoring Using Aggregated Signals
Speakers: Jionghua (Judy) Jin, Jing Li, and Yong Lei, Industrial & Operations Engineering Department, University of Michigan

Abstract: Aggregated signals are referred to the system responses contributed from the multiple operations embedded in a system. While abundant literature exists in analyzing signal profiles of system responses, little has been found on how to identify features of the aggregated signals in order to map these features to the individual operations. For this purpose, this research develops a multiscale feature mapping algorithm using wavelets. It shows that the proposed mutliscale mapping can enhance the separation of the individual operations that contribute to the aggregated signals not only in the localized time segments corresponding to the working ranges of individual operations, but also at the different scales corresponding to the frequency ranges of the operations. A case study is provided to validate the developed method.

Friday, June 6
8:30 – 10:00 AM
(ICS16) Reliability Assurance for Mission-Critical Systems
Narendra Soman, General Electric Co.


Title: Reflections on Storage Systems Reliability
Speaker: KK Rao, Distinguished Engineer and Senior Manager, Advanced Storage Subsystems, IBM Almaden Research Center

Abstract: Storage systems have an essential role in enterprise and mission critical applications - to provide uninterrupted access to data, prevent its loss and safeguard its integrity, and to do all of these while maximizing application performance. In this talk, we will present some of the challenges continued to be faced by storage systems designers that cause data loss or data integrity issues. We will also present some of the trends and modeling techniques employed in storage systems in achieving high reliability.


Title: Application of Reliability Methods in Validation of Medical Devices
Speaker: Dr. David Burns, Chief Statistician, Boston Scientific

Abstract: At Boston Scientific, our mission is “to improve the quality of patient care and the productivity of health care delivery through the development and advocacy of less-invasive medical devices and procedures.” Consequently, our customers - our patients - expect our products to perform their intended function every time, time-after-time. As you can imagine, it may be either impossible or impractical to physically test every component of every product. However, the statistical methods of acceptance sampling and distribution fitting can be very effective at determining if a process is capable of meeting its specifications with a high degree of confidence. In this presentation, we explain the relation between risk and reliability, and we show how to use reliability methods to address some situations when the usual validation approach may no longer be appropriate.


Title: Reliability Testing for Life Support Systems
Speaker: Todd Heydt, Senior Reliability Program Manager, GE Healthcare

Abstract: In a Perfect World all components and subsystems would have stringent reliability requirements that are proven via testing of multiple units over long durations at conditions identical to their use environment. However, we must live in the Real World. In the Real World, we must prove our products meet their reliability requirements within the limitations of budget and schedule. By looking at three examples of real components/subsystems with varying degrees of complexity, significance, and resource constraints, we will discuss the process followed within GE Healthcare-Life Support Solutions for developing reliability requirements and demonstrating compliance with those requirements.

Friday, June 6
8:30 – 10:00 AM
(ICS17) George Box’s Contributions to Q&P
Geoff Vining, Virginia Tech


Title: George Box: A Source of Inspiration for Quality and Productivity
Speaker: Bovas Abraham, University of Waterloo

Abstract: George Box is one of the most respected living statisticians. He has made fundamental contributions to many areas of statistics. In this talk we will look at some of his contributions to industrial statistics and quality. We will touch on several events such as the birth of response surfaces, creation of Technometrics, and the founding of the centre for quality and productivity improvement at the University of Wisconsin. Prof. Box's papers form a collection of contributions that have had a significant impact on the research and practice of statistics and quality around the world during the last half a century.


Title: George's Contributions to Process Control
Speaker: John MacGregor, McMaster University, Canada

Abstract: TBD


Title: George's Contributions to Industrial Experimentation
Speaker: Geoff Vining, Virginia Tech

Abstract: George Box is truly one of the towering figures in the history of industrial experimentation, going back to his seminal work on response surface methodology, Box and Wilson (1951). This talk outlines George pivotal works in industrial experimentation through out his career. Of especial interest are his contributions to sequential experimentation, designs to support second-order response surfaces, the impact of bias due to an underspecified model, evolutionary operations, and the importance of balancing many often conflicting criteria when selecting appropriate experimental designs.


Friday, June 6
8:30 – 10:00 AM
(ICS18) Industrial Split-lot Experimentation
Di Michelson, Sematech


Title: Designing and Running Super-Efficient Experiments: Optimum Blocking for Fractional Factorials or for More than One Hard-to-Change Factor
Speaker: James M. Lucas, J.M. Lucas & Associates
Co-Authors: Frank T. Anbari, The George Washington University and Derek Webb, Bemidji State University

Abstract: In the paper “Designing and Running Super-Efficient Experiments: Optimum Blocking with One Hard-to-Change Factor,” published in the January 2008 issue of JQT, Anbari and Lucas showed that for 2k experiments with one hard-to-change (H-T-C) factor, there is always a super-efficient blocking procedure that dominates a completely randomized design (CRD). The optimally blocked split-plot experiment is less expensive to run (because it requires fewer changes of the H-T-C factor) and more precise (because it has a smaller variance of prediction). Because a 2k CRD is D- and G-Optimal it is 100% efficient; it achieves the minimum variance of prediction. (However, the limitation that this optimality is only across designs with a single variance component was seldom if ever stated in the vast literature on optimum design.) Therefore, the more precise optimally blocked split-plot experiment is “super-efficient.” When our criterion is to minimize the cost of information, all numbers of blocks up to the optimum block size are candidates. Experiments using more blocks are dominated because they are more expensive to run and have a larger variance of prediction. The optimum number of blocks depends on the variance ratio (swhole-plot2/ssplit-plot2) and on the ratio of the costs of changing the Easy-to-Change and H-T-C factors. When costs of changing the H-T-C factor are sufficiently high, the 2-block experiment that requires only one reset of the H-T-C factor minimizes the cost of information. This fact justifies the frequent use of this procedure. A “Randomized-not-Reset” experiment that uses a random run order but does not reset each factor on every run is also dominated by an optimally blocked split-plot experiment. Here we extend these results to experiments with more than one H-T-C factor and to fractional factorial experiments. Our results extend naturally in both of these situations. For two H-T-C factors super-efficient blocking is found for four and more factors. When there are more than two H-T-C factors super-efficient blocking can be achieved only in experiments having larger numbers of factors. For fractional factorial experiments our results are compared with minimum aberration split-plot (MASP) designs. Most of the MASP literature examines a very limited set of split plot designs and visits each level of the H-T-C factor only once. Visiting each level of the H-T-C factor only once gives a large variance of prediction so we usually recommend against it. Even when this restriction is removed the MASP approach can give a larger variance of prediction than the optimum blocking approach. We recommend consideration of minimum-aberration only after optimum blocking. Our recommendation makes explicit the approach implicitly used in some of the historical split-plot literature. Our work shows that optimally blocked split-plot experiments should be used more frequently and that CRD and MASP designs should be used less frequently.


Title: Fractional Factorial Designs for Multi-Step Processes
Speaker: Jose Ramirez, W.L. Gore & Associates, Inc.

Abstract: In a recent article Kulahci et al. (2006) showed that minimum aberration is not designed to distinguish among the different types of two factor interactions when designing fractional factorial split-plots. Here we expand this idea to fractional factorial designs in situations where more than one split is involved; e.g., multi-step processes. We give design and model constraints to generate fractional factorial designs for multi-step situations, and examples of a 3-step process (split-split-plot), a 4-step (split-split-split-plot), and a 5-step (split-split- split-split-plot).


Title: A New Type of Design for Experiments With More Than One Hard-To-Change Factor
Speaker: Heidi Arnouts, Universiteit Antwerpen
Co-author: Peter Goos, Universiteit Antwerpen

Abstract: In many industrial experiments, some of the factors are not independently reset for each run. This is due to time and/or cost constraints and to the hard-to-change nature of these factors. A lot of research has been done on the design of experiments with hard-to-change factors. Most of this work restricts the attention to split-plot designs in which all the hard-to-change factors are reset at the same points in time. This constraint is to some extent relaxed in split-split-plot designs because these require the least hard-to-change factors to be reset more often than the most hard-to-change factors. A key feature of the split-split-plot designs, however, is that the least hard-to-change factors are reset whenever the most hard-to-change factors are reset. In this presentation, we relax this constraint and present a new type of design which allows the hard-to-change factor levels to be reset at entirely different points in time. We show that the new designs are cost-efficient and that they outperform split-plot and split-split-plot designs in terms of statistical efficiency.

Friday, June 6
10:30 AM – 12:00 PM
(ICS19) Applications in Reliability
Nalini Ravishanker, University of Connecticut

Title: Inference for the Step-Stress Model with Lagged Effects
Speaker: Nandini Kannan, University of Texas at San Antonio

Abstract: We consider models for experiments in which the stress levels are altered at intermediate stages during the observational period. These experiments, referred to as step-stress tests, belong to the class of accelerated models that are extensively used in reliability and life-testing applications. Models for step-stress tests have largely relied on the cumulative exposure model (CEM) discussed by Nelson. Unfortunately, the assumptions of the model are fairly restrictive and quite unreasonable for applications. We introduce a new step-stress model based on the Weibull distribution where the hazard function is continuous. We consider a simple experiment with only two stress level and develop a model that allows for a lag period before the effects of the change in stress are observed. Using this formulation in terms of the hazard function, we obtain the maximum likelihood estimators of the unknown parameters. A Monte Carlo simulation study is performed to study the behavior of the estimators obtained for different choices of sample sizes and parameter values. We analyze a real dataset obtained from an airforce study on altitude decompression and show that the model provides an excellent fit. This is joint work with D. Kundu, Department of Mathematics and Statistics, Indian Institute of Technology Kanpur, India.
.
Title: Bayesian Sample Size Determination for Reliability Studies of a Binary Trait
Speaker: Cyr E. M’Lan, University of Connecticut

Abstract: Reliability methods are often used to assess the consistency of different psychiatrists in rating the functional and mental status of their patients. To this end, the intra-class correlation coefficient (ICC) is used as a sensible measure of the inter-rater reliability. The focus here is on an experiment with a binary trait such as diseased/healthy. To conduct such a study, one needs to decide on the number of subjects, n, and the number of raters, k. Although sample size calculation for ICC has been widely studied in the frequentist literature (see Shoukri, Asyali, and Donner, 2004 for a review), there is virtually no literature in the Bayesian paradigm. This work is meant to fill this void and show how sample size calculations for ICC can be addressed from a Bayesian standpoint. The goal is primarily to know how many subjects for a fixed number of raters and how many raters for a fixed number of subjects are required to guarantee an expected width of the credible interval for ICC no larger than a pre-specified value. Optimal allocation of (n,k) for N=nk fixed to conduct a reliability study is also investigated, keeping cost considerations in mind.

Title: Methodologies for Recurrent Event Data under Competing Risks
Speaker: Ananda Sen, University of Michigan

Abstract: The focus of the talk is recurrent events for which the relevant data comprise successive event times for a recurrent phenomenon along with a event-count indicator. Situations in which individuals or systems in some population experience recurrent events are quite common in areas such as manufacturing, software debugging, risk analysis, and clinical trials. While the different application areas dealing with recurrent data enjoy certain commonalities, often the issues specific to an area warrant a non-standard modification of the basic methodologies. In this talk, I shall discuss competing risks analysis in the context of recurrent failures that are subjected to multiple failure modes. Models and methodologies for analyzing a single as well as multiple clusters of recurrent events will be discussed in detail. The situation dealing with missing or masked cause of failure will further be addressed.

Friday, June 6
10:30 AM – 12:00 Noon
(ICS20) Statistical Process Control
Dan Jeske, University of California - Riverside

Title: Using Statistical Process Control for Change Detection in Network Systems
Speaker: Mazda Marvasti, Chief Technology Officer, Integrien Corporation

Abstract: The occurrence of change in data centers hosting servers and applications is a common phenomenon. The detection of when change has occurred often eludes datacenter operators due to a lack of process or agility requirement of enterprises. The current work proposes an on-line mechanism where real-time network data is compared with historical behavior to determine if a change has occurred in the environment. The detection scheme is performed in terms of quality control parameters namely number of outliers and distance measure between the data and it's dynamically calculated threshold. Dynamic threshold (DT) defines the historical upper and lower bounds of the data which the current method uses to determine data-to-DT interdependence.


Title: Statistical Process Control for Non-stationary Processes: An Example Based Tour
Authors: Bart De Ketelaere, Bert Ostyn, Kristof Mertens, Josse De Baerdemaeker and Paul Darius, Division of Mechatronics, Biostatistics and Sensors (MeBioS)
Speaker:

Abstract: Classical statistical process control (SPC) is focused on applications where a certain target value for the quality characteristic is available, or where the process is stationary. However, those assumptions are often not met and -often customized- adaptations to the classical concept of SPC are needed. In this contribution, several practical examples from industry will be given where such adaptations are required. Starting from a univariate case of monitoring livestock, the talk will focus on monitoring very high dimensional data (e.g. power spectral densities, near infrared spectra, .) that possibly exhibit a non-stationary behavior. Special attention will be paid to the use of data reduction techniques (supervised and non-supervised) such as principal components –including their recursive versions- for multivariate SPC.


Title: Using the Repeated Two-Sample Rank Procedure for Detecting Anomalies in Space and Time
Speaker: Ron Fricker, Naval Postgraduate School

Abstract: The Repeated Two-Sample Rank (RTR) procedure is a nonparametric statistical process control methodology that applies to both univariate and multivariate data. The method transforms a sample of data into univariate statistics; changes in the distribution of the data are then detected using nonparametric rank tests. In this discussion we explore its use as a spatio-temporal event detection and monitoring methodology for use in applications such as biosurveillance, crime mapping, or IED incidence change detection. The methodology is designed to sequentially incorporate information from individual observations as they arrive into an automated systems and thus can operate on data in real-time. Upon a signal of a possible distributional change, the methodology suggests a way to graphically indicate the likely location of the distributional change.

Title: A Nonparametric Cusum Algorithm for Timeslot Sequences with Applications to Network Surveillance
Speakers: Veronica Montes de Occa and Daniel R. Jeske, Department of Statistics, University of California - Riverside

Abstract: We adapt the classic cusum change-point detection algorithm for Applications to data network monitoring where various and numerous performance and reliability metrics are available to aid with early identification of realized or impending failures. Three significant challenges that must be overcome are:

1) the need for a nonparametric technique so that a wide variety of metrics (including discrete metrics) can be included in the monitoring process,

2) the need to handle time varying distributions for the metrics that reflect natural cycles in work load and traffic patterns, and

3) the need to be computationally efficient with data processing of the massive number of metrics that are available from IT environments.

Friday, June 6
10:30 AM – 12:00 PM
(ICS21) Response Surface Models with Random Effects
Professor Andre' I. Khuri, Professor Emeritus, Department of Statistics, University of Florida

Title: A Discussion of Various Ways to Analyze Data From Response Surface Split-Plot Experiments
Speaker: Professor Peter Goos, Department of Mathematics, Statistics, and Actuarial Sciences,
University of Antwerp, Belgium

Abstract: Split-plot and other multi-stratum structures are widely used in factorial and response surface experiments and residual maximum likelihood (REML) and generalized least squares (GLS) estimation is seen as the state-of-the-art method of data analysis for nonorthogonal designs. We analyze data from an experiment run to study the effects of five process factors on the drying rate for freeze dried coffee and find that the main-plot variance component is estimated to be zero. We show that this is a typical property of REML-GLS estimation which is highly undesirable and can give misleading conclusions. We also investigate approaches that require replication for estimation the variance components. However, we recommend a Bayesian analysis, using an informative prior distribution for the main-plot variance component and implemented using Markov chain Monte Carlo sampling. Paradoxically, the Bayesian analysis is less dependent on prior assumptions than the REML-GLS analysis. Bayesian analyses of the coffee freeze drying data give more realistic conclusions than REML-GLS analysis, providing support for our recommendation.


Title: Robust Parameter Design Using Generalized Linear Mixed Models
Speaker: Professor Shaun Wulff, Department of Statistics, The University of Wyoming

Abstract: This talk will address a problem from semiconductor manufacturing where the response is non-normal and there is a random effect representing an extraneous source of variation. Levels of the control variables are to be chosen that minimize the variation arising from the random effect when the mean of the response is constrained to a target value. A generalized linear mixed model is used to develop a response surface approach to address this problem.


Title: Designs for Response Surface Models with Random Block Effects
Speaker: Sourish Saha, GlaxoSmithKline

Abstract: One traditional criterion for comparing response surface designs is the size of the prediction variance. Another criterion is based on utilizing the power of an F-test concerning the fixed unknown parameters in the associated model. The purpose of this talk is to compare response surface designs on the basis of the prediction variance and power criteria when the model contains a random block effect, in addition to the fixed polynomial portion of the model. The proposed approach uses quantiles of these design criteria on concentric surfaces, either inside the experimental region or within the so-called alternative space associated with the unknown parameters. The dependence of these quantiles on the unknown value of the ratio of two variance components, namely, the ones for the block effect and the experimental error, is depicted by plotting the so-called quantile dispersion graphs. (This is a joint work with André I. Khuri)

Friday, June 6
10:30 AM – 12:00 PM
(ICS22) Demand Management in Large-Scale Computing
Bill Heavlin, Google, Inc.

Title: User Preference and Search Engine Latency
Speaker: Jake Brutlag, Google, Inc.
Authors: Jake Brutlag, Hilary Hutchinson, Maria Stone, Google, Inc.

Abstract: Industry research advocates a 4-second rule for web pages to load. Usability engineers note that a response time over 1 second may interrupt a user's flow of thought. There is a general belief that all other factors equal, users will abandon a slow search engine in favor of a faster alternative.

This study compares two mock search engines that differ only in branding (generic color scheme) and latency (fast vs. slow). The fast latency was fixed at 250 ms, while 4 different slow latencies were evaluated: 2s, 3s, 4s, and 5s.

When the slower search engine latency is 5 seconds, users state they perceive the faster engine as faster. When the slower search engine latency is 4 or 5 seconds, users choose to use the faster engine more often. Based on pooling data for 2 s and 3 s, once slow latency exceeds 3 seconds, users are 1.5 times as likely to choose the faster engine.


Title: A Statistical Approach to Optimal Consolidation of Computer Workload
Speaker: Ta-Hsin Li, IBM TJ Watson Research Center

Abstract: In on-demand computing services, a customer pays for what they actually use and the service provider is free to reallocate unused capacity to other customers. An important issue in such a shared environment is optimal workload consolidation and capacity planning. In this talk, a quantile-based statistical approach is proposed for analyzing the benefits and risks of workload consolidation. An optimization problem is formulated for workload consolidation with the aim of maximizing the benefit while keeping the risk at an acceptable level.


Title: Forecasting Telecommunications Traffic and Internet Activity
Speaker: Gordon Reikard, Sprint-Nextel

Abstract: Forecasting and simulation of data on internet activity, and communications networks in general, are complicated by nonlinear variability. At high frequencies, time series from the world-wide web and communications networks exhibit heavy tails and irregular bursts of activity characterized by intermittent outliers. At lower resolutions, they also exhibit underlying signals.

Empirical tests were run on call series at Sprint-Nextel and information downloads from company websites. At high frequencies, evidence of heavy tails is confirmed, while the fractal dimension is non-integer, indicating a high probability of extreme events. At lower frequencies, the nonlinear variability dissipates, leaving underlying signals, specifically a stochastic trend and 24-hour and 7-day cycles.

Forecasts using both ARIMA and transfer function models were considered. In spite of nonlinear variability, forecasting at lower resolutions is straightfotward, and becomes the basis for important cost savings.