-
Notifications
You must be signed in to change notification settings - Fork 1
Simplify scripts #45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Simplify scripts #45
Conversation
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
Ensure the fp, fn, tp gdf have the necessary columns so we don't have to check when calculating the metrics
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
I found out that the saved global metrics where always computed with the macro-average method and I changed the code to make it dependant on the chosen method. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Everything works well !
Just spotted to places that need your review.
scripts/assess_detections.py
Outdated
@@ -336,13 +336,13 @@ def main(cfg_file_path): | |||
# we select the threshold which maximizes the f1-score on the val dataset | |||
if 'val' in metrics_cl_df_dict.keys() and CONFIDENCE_THRESHOLD: | |||
logger.error('The confidence score was determined over the val dataset, but a confidence score is given in the config file.') |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
"confidence threshold" and not "confidence score" ?
|
||
if tested_seed == max_seed-1: | ||
logger.warning(f'No satisfying seed found between 0 and {max_seed}.') | ||
logger.info(f'The best seed was {seed} with ~{best_split} class subsets containing the correct proportion (trn~0.7, val~0.15, tst~0.15).') |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
rather test_seed
?
I tried to simplify the tileset generation by dividing it into several scripts. Everything is called by the script
generate_tilesets.py
, so the commands are still the same.I tested the result on the roads and border points. I did not check the functionality for FP and empty tiles and I would have to find some files to do so.