Back to Courses

Model Validation

Up to 3 hours
Validation methodology Sensitivity analysis Data quality assessment

Course Overview

This session explores validation techniques for ensuring that simulation models accurately represent the real-world system for their intended purpose. Validation confirms that the model’s output matches real-world behavior within an acceptable range — a critical step before using any model for decision-making.


Objectives

  • Understand the importance and scope of model validation
  • Learn various techniques for validating simulation models
  • Apply validation approaches in AnyLogic to confirm model accuracy
  • Identify common sources of validation errors and discrepancies

Prerequisites

  • Basic understanding of simulation modeling concepts
  • Familiarity with AnyLogic

Topics Covered

1. Validation Concepts

  • Definition: confirming that a model provides outputs consistent with the real system for its intended purpose
  • Validation is goal-specific — a model valid for one purpose may not be valid for another
  • Defining key variables, indicators, and acceptable accuracy ranges before model construction

2. Validation Techniques

  • Animation: Visually compare model behavior with real-world activities
  • Comparison with other models: Validate against other validated models or analytical results
  • Degenerate and extreme condition tests: Test model behavior under extreme parameter values
  • Event validity: Compare simulated events to real-world event data
  • Face validity: Expert review of model logic and outputs
  • Internal validity: Multiple replications to assess stochastic variability
  • Operational graphics: Visualize performance measures over time
  • Sensitivity analysis: Vary inputs to assess output sensitivity patterns
  • Predictive validation: Compare model predictions with real-world outcomes
  • Turing tests: Blind comparison — can experts distinguish model from real data?

3. Common Sources of Validation Errors

  • Non-representative data: Data from systems other than the one being modeled
  • Incorrect data type or format: Ambiguity in whether data includes downtime or off-shift periods
  • Measurement and rounding errors: Excessive rounding transforming continuous distributions into discrete ones
  • Biased data: Self-interested reporting leading to inaccurate inputs
  • Unknown collection conditions: Uncertainty about when and how data was gathered

Acquired Knowledge

By the end of this session, participants will be able to:

  • Apply multiple validation techniques to simulation models
  • Identify and resolve validation errors and discrepancies
  • Ensure simulation models are valid and reliable for their intended decision-support purposes

Interested in this course?

Get in Touch