great-expectations/great_expectations
Always know what to expect from your data.
This project helps data professionals ensure the quality and reliability of their datasets before using them for analysis or operations. You define "Expectations" about your data, like a column always containing numbers, and the tool then checks if your data meets these expectations, providing reports on any deviations. Data analysts, data scientists, and data engineers use this to maintain high data quality.
11,270 stars. Used by 3 other packages. Actively maintained with 41 commits in the last 30 days. Available on PyPI.
Use this if you need a systematic way to validate the quality of your incoming data, catch issues early, and document your data's expected characteristics.
Not ideal if you're looking for a general-purpose data cleaning tool, as its primary focus is on validation and documentation rather than automated data transformation.
Stars
11,270
Forks
1,697
Language
Python
License
Apache-2.0
Category
Last pushed
Mar 18, 2026
Commits (30d)
41
Dependencies
18
Reverse dependents
3
Get this data via API
curl "https://pt-edge.onrender.com/api/v1/quality/mlops/great-expectations/great_expectations"
Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.
Recent Releases
Related tools
ray-project/ray
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI...
kedro-org/kedro-viz
Visualise your Kedro data and machine-learning pipelines and track your experiments.
polyaxon/traceml
Engine for ML/Data tracking, visualization, explainability, drift detection, and dashboards for Polyaxon.
wandb/wandb
The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models...
flyteorg/flytekit
Extensible Python SDK for developing Flyte tasks and workflows. Simple to get started and learn...