The notebooks in this repository were used to execute the experimental evaluation of our paper "Towards Realistic Error Models for Tabular Data". Specifically,
- the notebook
dataset_generation.ipynbcontains the procedure we followed to generate datasets corresponding to the error scenarios we describe in our paper. - The notebook
dataset_analysis.ipynbcontains our analysis of theHOSPdataset. - The notebook
plots.ipynbcontains the procedure we use to generate the figures in our publication. It reads experiment's results from theerror_paper/measurements/directory -- check the notebook's code for details.
We use poetry to manage dependencies. Simply run poetry install to install all dependencies.
In our experiments, we examine data cleaning and downstream machine learning task impact using tab_err.
- In the first part of the data cleaning experiments, we generate various erroneous versions of the
HOSPdataset and clean them withHoloClean(benchmarks/hosp-impact). - We then proceed to generate various erroneous versions of datasets
bridges,beers,restaurantandcarsand correct them with algorithms baran&raha, holoclean and renuver (benchmarks/cleaning-impact). - In the downstream machine learning task impact, we look at how ML models behave given data with various errors (
benchmarks/ml_downstream_experiments).
Check the documentation in benchmarks/README.md for instructions on how to replicate our measurements.
We also looked at the memory and runtime of tab_err using various error models and dataset sizes. See the directory benchmarks/profiling for examples.