ablate turns deep learning experiments into structured, human-readable reports. It is built around five principles:
- composability: sources, queries, blocks, and exporters can be freely combined
- immutability: query operations never mutate runs in-place, enabling safe reuse and functional-style chaining
- extensibility: sources, blocks, and exporters are designed to be easily extended with custom implementations
- readability: reports are generated with humans in mind: shareable, inspectable, and format-agnostic
- minimal friction: no servers, no databases, no heavy integrations: just Python and your existing logs
Currently, ablate supports the following sources and exporters:
Install ablate using pip:
pip install ablate
The following optional dependencies can be installed to enable additional features:
ablate[clearml]
to use ClearML as an experiment sourceablate[mlflow]
to use MLflow as an experiment sourceablate[tensorboard]
to use TensorBoard as an experiment sourceablate[wandb]
to use WandB as an experiment sourceablate[jupyter]
to use ablate in a Jupyter notebook
ablate is built around five composable modules:
- ablate.sources: load experiment runs from various sources
- ablate.queries: apply queries and transformations to the runs
- ablate.blocks: structure content as tables, text, figures, and other blocks
- ablate.Report: create a report from the runs and blocks
- ablate.exporters: export a report to various formats
To create your first Report, define one or more experiment sources. For example, the built in Mock can be used to simulate runs:
import ablate
source = ablate.sources.Mock(
grid={"model": ["vgg", "resnet"], "lr": [0.01, 0.001]},
num_seeds=2,
)
Each run in the mock source has accuracy, f1, and loss metrics, along with a seed parameter as well as the manually defined parameters model and lr. Next, the runs can be loaded and processed using functional-style queries to e.g., sort by accuracy, group by seed, aggregate the results by mean, and finally collect all results into a single list:
runs = (
ablate.queries.Query(source.load())
.sort(ablate.queries.Metric("accuracy", direction="max"))
.groupdiff(ablate.queries.Param("seed"))
.aggregate("mean")
.all()
)
Now that the runs are loaded and processed, a Report comprising multiple blocks can be created to structure the content:
report = ablate.Report(runs)
report.add(ablate.blocks.H1("Model Performance"))
report.add(
ablate.blocks.Table(
columns=[
ablate.queries.Param("model", label="Model"),
ablate.queries.Param("lr", label="Learning Rate"),
ablate.queries.Metric("accuracy", direction="max", label="Accuracy"),
ablate.queries.Metric("f1", direction="max", label="F1 Score"),
ablate.queries.Metric("loss", direction="min", label="Loss"),
]
)
)
Finally, the report can be exported to a desired format such as Markdown:
ablate.exporters.Markdown().export(report)
This will produce a report.md
file with the following content:
# Model Performance
| Model | Learning Rate | Accuracy | F1 Score | Loss |
| :----- | ------------: | -------: | -------: | ------: |
| resnet | 0.01 | 0.94285 | 0.90655 | 0.0847 |
| vgg | 0.01 | 0.92435 | 0.8813 | 0.0895 |
| resnet | 0.001 | 0.9262 | 0.8849 | 0.0743 |
| vgg | 0.001 | 0.92745 | 0.90875 | 0.08115 |
To compose multiple sources, they can be added together using the +
operator
as they represent lists of Run objects:
runs1 = ablate.sources.Mock(...).load()
runs2 = ablate.sources.Mock(...).load()
all_runs = runs1 + runs2 # combines both sources into a single list of runs
ablate queries are functionally pure such that intermediate results are not modified and can be reused:
runs = ablate.sources.Mock(...).load()
sorted_runs = Query(runs).sort(ablate.queries.Metric("accuracy", direction="max"))
filtered_runs = sorted_runs.filter(
ablate.queries.Metric("accuracy", direction="max") > 0.9
)
sorted_runs.all() # still contains all runs sorted by accuracy
filtered_runs.all() # only contains runs with accuracy > 0.9
By default, ablate reports populate blocks based on the global list of runs passed to the report during initialization. To create more complex reports, blocks can be populated with a custom list of runs using the runs parameter:
report = ablate.Report(sorted_runs.all())
report.add(ablate.blocks.H1("Report with Sorted Runs and Filtered Runs"))
report.add(ablate.blocks.H2("Sorted Runs"))
report.add(
ablate.blocks.Table(
columns=[
ablate.queries.Param("model", label="Model"),
ablate.queries.Param("lr", label="Learning Rate"),
ablate.queries.Metric("accuracy", direction="max", label="Accuracy"),
]
)
)
report.add(ablate.blocks.H2("Filtered Runs"))
report.add(
ablate.blocks.Table(
runs = filtered_runs.all(), # use filtered runs only for this block
columns=[
ablate.queries.Param("model", label="Model"),
ablate.queries.Param("lr", label="Learning Rate"),
ablate.queries.Metric("accuracy", direction="max", label="Accuracy"),
]
)
)
ablate is designed to be extensible, allowing you to create custom sources, blocks, and exporters by implementing their respective abstract classes.
To contribute to ablate, please refer to the contribution guide.