Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
guillaume-millot authored May 1, 2024
1 parent 08a5ab5 commit 9fc59c1
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions eval/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@

## Setup

To run the evaluation scripts, we need some additional requirements that are not listed in the project dependencies.
To run the evaluation scripts, you need some additional requirements that are not listed in the project dependencies.

```
apt-get install wkhtmltopdf
Expand All @@ -12,7 +12,7 @@ apt-get install wkhtmltopdf

First, you need to generate evaluation data with the `eval_table_extraction.py` script. This script will iterate through several reports and apply the set of table extraction algorithms you provided in your yaml configuration.

As an example, you might select the pages in the report from their filename and then apply several table extraction algorithms. Check out `configs/eval_table_extraction.yaml` for a suitable evaluation script.
Check out `configs/eval_table_extraction.yaml` for a suitable yaml configuration.

You can then call the script as :

Expand All @@ -37,6 +37,6 @@ To run the application, it is as simple as :
streamlit run eval/eval_app.py eval/data/data_step2_before-currency-unit_eval.csv
```

`data_step2_before-currency-unit_eval.csv` is a cleaned up version of the `data_step2_before-currency-unit.csv` file which contains reference data extracted and manually cleaned up by the team.
`data_step2_before-currency-unit_eval.csv` is a cleaned up version of the `data_step2_before-currency-unit.csv` file which contains reference data extracted and manually cleaned up by the TaxObservatory team.

At launch, you will be requested to provide a pickle file with extracted data. You might select `eval_20240408_200249.plk` in the `eval/data/` directory. It contains extracted tables for multiple reports and extractions and is a great way to get started.
At launch, you will be requested to provide a pickle file with extracted data. You might select `eval_20240408_200249.plk` from the `eval/data/` directory. It contains extracted tables for multiple reports and extractions and is a great way to get started.

0 comments on commit 9fc59c1

Please sign in to comment.