A Data Science Central Community

I'm conducting a simulation analysis of how different system characteristics of a casualty evacuation aircraft (speed, capacity, range, number of aircraft) affect system performance (delays, timing through different levels of care etc). The purpose of the study is not to determine the optimal or near optimal set of characteristics, but to determine the marginal benefit or penalty of each aircraft system characteristic to the patient compared to the current system, so that if and when the system is funded, designed and implemented, constraints such as cost, weight, volume can be evaluated within the context of patient outcomes, and the best *affordable* and *practical* design will be chosen based on benefit to patient outcome.

My question is whether anyone has recommendations for an experimental design to handle the following factor/levels:

A-2

B-2

C-4

D-5

E-5

All of these factors are significant and are the result of a sensitivity analysis starting with 12 factors. Factors C through E have so many levels because I'm not interested in just the extreme values, I'm interested the response each marginal change in factor levels has against a baseline. From previous experience with similar output, the response is highly nonlinear in factors C through E. Two factor interactions between C through E are significant. In places the response surface is not smooth or well behaved. The above experiment needs to be repeated 7 times under different sets of assumptions. To examine every permutation in all 7 cases requires 2800 runs each replicated 40 times, which translates to about 11 days of continuous computing on a single machine. I have a lab at my disposal and can run several machines, and I have the time to do it.

BUT. Do I need to? Given the nonlinear, poorly behaved nature of the response surface, does anyone have any recommendations as to how I can sample the decision space to adequately estimate the difference between any hypothetical system within the bounds of the experiment and the current capability (which is also simulated in a separate experiment)?

Thanks for any insights,

Jon

My question is whether anyone has recommendations for an experimental design to handle the following factor/levels:

A-2

B-2

C-4

D-5

E-5

All of these factors are significant and are the result of a sensitivity analysis starting with 12 factors. Factors C through E have so many levels because I'm not interested in just the extreme values, I'm interested the response each marginal change in factor levels has against a baseline. From previous experience with similar output, the response is highly nonlinear in factors C through E. Two factor interactions between C through E are significant. In places the response surface is not smooth or well behaved. The above experiment needs to be repeated 7 times under different sets of assumptions. To examine every permutation in all 7 cases requires 2800 runs each replicated 40 times, which translates to about 11 days of continuous computing on a single machine. I have a lab at my disposal and can run several machines, and I have the time to do it.

BUT. Do I need to? Given the nonlinear, poorly behaved nature of the response surface, does anyone have any recommendations as to how I can sample the decision space to adequately estimate the difference between any hypothetical system within the bounds of the experiment and the current capability (which is also simulated in a separate experiment)?

Thanks for any insights,

Jon

Tags: Design, Experiments, Response, Rurface, Simulation, of

© 2020 AnalyticBridge.com is a subsidiary and dedicated channel of Data Science Central LLC Powered by

Badges | Report an Issue | Privacy Policy | Terms of Service

**Most Popular Content on DSC**

To not miss this type of content in the future, subscribe to our newsletter.

- Book: Statistics -- New Foundations, Toolbox, and Machine Learning Recipes
- Book: Classification and Regression In a Weekend - With Python
- Book: Applied Stochastic Processes
- Long-range Correlations in Time Series: Modeling, Testing, Case Study
- How to Automatically Determine the Number of Clusters in your Data
- New Machine Learning Cheat Sheet | Old one
- Confidence Intervals Without Pain - With Resampling
- Advanced Machine Learning with Basic Excel
- New Perspectives on Statistical Distributions and Deep Learning
- Fascinating New Results in the Theory of Randomness
- Fast Combinatorial Feature Selection

**Other popular resources**

- Comprehensive Repository of Data Science and ML Resources
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- 100 Data Science Interview Questions and Answers
- Cheat Sheets | Curated Articles | Search | Jobs | Courses
- Post a Blog | Forum Questions | Books | Salaries | News

**Archives:** 2008-2014 |
2015-2016 |
2017-2019 |
Book 1 |
Book 2 |
More

**Most popular articles**

- Free Book and Resources for DSC Members
- New Perspectives on Statistical Distributions and Deep Learning
- Time series, Growth Modeling and Data Science Wizardy
- Statistical Concepts Explained in Simple English
- Machine Learning Concepts Explained in One Picture
- Comprehensive Repository of Data Science and ML Resources
- Advanced Machine Learning with Basic Excel
- Difference between ML, Data Science, AI, Deep Learning, and Statistics
- Selected Business Analytics, Data Science and ML articles
- How to Automatically Determine the Number of Clusters in your Data
- Fascinating New Results in the Theory of Randomness
- Hire a Data Scientist | Search DSC | Find a Job
- Post a Blog | Forum Questions