April 18, 2026
•
AI & Machine Learning
Test Data Management for Engineering Teams: How to Connect Simulation and Physical Test
Simulation and test data live separately. Unifying them with manual reconciliation, point integrations, and purpose-built infrastructure
In most engineering organisations, simulation data and physical test data are treated as two separate worlds.
Different teams own them. Different systems store them. Different conventions govern them. And almost nobody connects them, not because the value is unclear, but because no tool in the stack was designed to.
Then we wonder why validation takes three times longer than it should. Why AI models trained on simulation data fail against physical test results. Why validation knowledge evaporates when a senior engineer retires. Industry bodies like SAE International have published validation and verification guidance for decades, but almost all of it assumes simulation and test are handled as separate disciplines, which is exactly the pattern that creates the integration gap we see every day.
This article covers why the disconnection exists, the cost you are probably paying for it right now, and the three approaches organisations take to fix it, with the trade-offs of each.
Why simulation and test data are disconnected by default
The disconnection is structural. Simulation tools like Ansys and Siemens Simcenter were built by and for simulation engineers. They produce outputs in formats meaningful to that community. Test tools, DAQ systems, dynamometers, acoustic chambers, correlation environments, were built by and for test engineers. Neither was built to talk to the other.
The outputs are different. The variable names are different. The metadata conventions are different. The teams managing them report to different VPs in most organisations.
What this means in practice: when a validation engineer needs to compare a simulation prediction against a physical test result, they do it by hand. Extract from simulation, extract from test, align variables, reconcile units, spend half a day setting up what should be a query. On every program.
The cost you’re already paying
The cost of this disconnection shows up in three places, and most organisations pay all three simultaneously without realising they are the same cost.
Longer validation cycles
The 3× multiplier is not an exaggeration. Every comparison between simulation and test requires manual reconciliation, and that reconciliation time compounds across every variant, every load case, every operating condition. What should be a half-day validation pass becomes a two-week exercise.
Failed AI deployment
Engineering AI initiatives that train on simulation data break when they encounter test data. The model has learned a representation that is internally consistent with simulation but has not been calibrated against the physical reality test data captures. This is the single largest cause of AI models failing to generalise in engineering contexts. McKinsey’s 2025 State of AI research confirms this at scale: organisations that successfully deploy AI in engineering have redesigned their data architecture first, and organisations that fail cite data and workflow issues, not model issues, as the reason.
Knowledge loss
The people who know which simulation corresponds to which test, which boundary conditions were used, which instrumentation was active, that knowledge lives in engineers’ heads, in email threads, in notebooks. When they leave, it leaves. NAFEMS guidance has documented this failure mode repeatedly: without structured management, simulation knowledge reliably disappears with the engineers who held it.
The 3 approaches to connecting simulation and test data
Three approaches are actually in use across engineering organisations today. Each one fits a specific context. Here are the honest trade-offs.
Approach 1 - Manual reconciliation (the status quo)
A validation engineer manually pulls simulation data and test data, aligns them by hand in Excel or a notebook, and generates the comparison. Repeated every validation pass, every program.
Pros
- No tooling cost.
- Full flexibility, any comparison you can think of, you can do.
- Appropriate for very small teams or one-off validations.
Cons
- 3× longer validation cycles than necessary.
- Knowledge lives in the engineer’s head and leaves with them.
- Impossible to retroactively query "all simulations correlated to tests in the last 18 months."
- AI and surrogate modelling cannot be built on top of this.
- The engineer doing the reconciliation is not doing engineering, opportunity cost compounds.
Verdict: The default. Works for small validation scope. Fails the moment you want cross-program insight or AI on simulation data.
Approach 2 - Point integrations between specific tools
Custom scripts or vendor-built integrations connect specific simulation tools to specific test environments (e.g., Ansys to a specific DAQ system). Usually commissioned program-by-program.
Pros
- Solves the specific workflow it was built for.
- Lower cost than full platform deployment.
- Owned internally - no vendor dependency for the core logic.
Cons
- Each integration is bespoke, no reuse across tool combinations.
- Adding a new solver or test system requires a new integration project.
- Metadata, lineage, and queryability are usually afterthoughts.
- Maintenance burden grows with every new integration added.
- Does not solve the cross-program or cross-team knowledge problem.
Verdict: Useful for bridging one high-value workflow. Does not scale as an organisational strategy.
Approach 3 - Purpose-built engineering data infrastructure
A unified layer that ingests from simulation environments and test environments into the same structured, queryable data model. Engineering data infrastructure platforms like Key Ward are designed to do this natively, without requiring tool changes on either side.
Pros
- One data layer, both data types, queryable together.
- New simulation tools and new test systems are added via connectors, not rebuilds.
- Knowledge stays institutional, workflows survive team changes.
- AI and surrogate modelling become viable because training data is unified.
- Validation cycles compress dramatically, 60%+ is typical.
- Existing solvers, test benches, and PLM stay exactly where they are.
Cons
- Requires initial setup to connect environments and define canonical schemas.
- Newer category, fewer public case studies than PLM or point integrations.
- Paid platform, though deployment is weeks, not months.
Verdict: The approach engineering organisations serious about cross-team insight and AI deployment are investing in.
The 7-environment pattern
A leading automotive engineering team we worked with was generating data across seven simulation and test environments: high-fidelity CFD, lower-fidelity aero simulations, thermal, NVH, two physical test rigs, and one correlation environment.
Every quarterly review, leadership asked the same question: "how are we tracking against the previous program?" The standard answer was "give me two days to pull the data together." Because there was no unified layer, the data existed in seven places, in seven formats, with seven naming conventions.
When we structured the ingestion, one connector per environment, a shared schema on landing, a single queryable layer underneath, the same question took 30 seconds to answer. Dashboards updated automatically as new runs completed. Leadership went from quarterly reviews to live visibility.
The quantifiable outcome: 60% faster insight generation, €230,000 in engineering time saved in the first year. The unquantifiable outcome: engineers stopped being the pipeline.
How to start without disrupting existing workflows
The mistake most organisations make is trying to boil the ocean. They decide to unify all simulation and test data at once, which means changing how every engineer works. This is consistent with NAFEMS deployment guidance, the failed deployments in their 20-year case history are almost all attempts to do everything at once.
The approach that actually works is smaller. Pick the one simulation-test pair that costs your organisation the most in manual reconciliation. Usually this is the primary validation workflow for your most critical program. Connect those two sources into a unified structured layer. Leave every other workflow untouched.
Measure the impact on validation cycle time for that one workflow. In almost every case, it is dramatic enough that the business case to expand to the next pair writes itself.

Stop treating simulation and test as separate worlds
Two ways to move forward, pick what fits your stage.



