ModelSentry monitors your production ML models for data drift — locally, with no data ever leaving your machine.
Six months after you deployed the churn model, something changed. New marketing channels brought in a different demographic — younger users, different income brackets, different behavior patterns. The model kept predicting. Confidently. Silently. On data it had never seen.
You found out when a product manager pinged you asking why churn had jumped 12% in Q3.
You pulled the feature distributions. The age column had shifted by nearly 15 years. Income looked nothing like training. The model had been extrapolating for months — and you had no early warning system in place.
"There was no alarm. There should have been."
Runs anywhere Python runs. No Docker. No cloud account. No infra to manage.
One decorator. No changes to your model logic. No SDK calls inside predict().
Opens localhost:8080. PSI and KS drift scores update as predictions roll in. You get an email the moment drift crosses your threshold.
Most monitoring tools tell you "accuracy dropped." ModelSentry tells you which features drifted, by how much, and whether it's worth waking up for.
| Feature | PSI | KS p-val | Severity |
|---|---|---|---|
| age | 0.31 | 0.001 | ■ critical |
| income | 0.18 | 0.023 | ▲ warning |
| country | 0.04 | 0.412 | ● stable |
ModelSentry computes statistical profiles locally — histograms, means, PSI scores, KS statistics. Raw feature values and raw predictions are never written to disk, never transmitted over a network, never seen by anyone.
The dashboard runs at
localhost:8080.
Email alerts are sent from your own Gmail account via your own app password.
We have no servers that touch your data. There is nothing to breach.
We're opening the beta to a small group of data scientists. No credit card. No enterprise sales process. No SaaS subscription to cancel.
Install, monitor, and tell us what breaks.
Free for beta users, forever.