JBris/model-calibration-evaluation

Evaluating model calibration methods for sensitivity analysis, uncertainty analysis, optimisation, and Bayesian inference

32
/ 100
Emerging

This project helps researchers and engineers assess and improve the accuracy of their complex computer models. It takes your model's outputs and applies various calibration methods to evaluate how well the model parameters match real-world observations or experimental data. The output provides insights into the best calibration techniques for your specific modeling needs, benefiting anyone performing sensitivity analysis, uncertainty analysis, optimization, or Bayesian inference.

No commits in the last 6 months.

Use this if you need to determine the most effective way to calibrate your simulation models to ensure they accurately reflect real-world phenomena.

Not ideal if you are looking for a tool to build or run simulations, as this project focuses solely on evaluating calibration methods.

scientific-modeling engineering-simulations uncertainty-quantification parameter-estimation model-validation
Stale 6m No Package No Dependents
Maintenance 0 / 25
Adoption 6 / 25
Maturity 16 / 25
Community 10 / 25

How are scores calculated?

Stars

15

Forks

2

Language

Python

License

MIT

Last pushed

Mar 18, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/JBris/model-calibration-evaluation"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.