Skip to content

Commit 7b2724e

Browse files
version 1.0.3 (#18)
1 parent bc8b2b3 commit 7b2724e

13 files changed

+35
-112
lines changed

baselines/inference/MegaDescriptor-B-224.ipynb

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
"from timm import create_model\n",
2222
"\n",
2323
"from wildlife_tools.features import DeepFeatures\n",
24-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
24+
"from wildlife_tools.data import WildlifeDataset\n",
2525
"from wildlife_tools.similarity import CosineSimilarity\n",
2626
"from wildlife_tools.inference import KnnClassifier\n",
2727
"\n",
@@ -83,17 +83,15 @@
8383
" ])\n",
8484
"\n",
8585
" database = WildlifeDataset(\n",
86-
" metadata=metadata,\n",
86+
" metadata=metadata.query('split == \"train\"'),\n",
8787
" root=f'{root_images}/{name}/',\n",
8888
" transform=transform,\n",
89-
" split=SplitMetadata('split', 'train'),\n",
9089
" )\n",
9190
"\n",
9291
" query = WildlifeDataset(\n",
93-
" metadata=metadata,\n",
92+
" metadata=metadata.query('split == \"test\"'),\n",
9493
" root=f'{root_images}/{name}/',\n",
9594
" transform=transform,\n",
96-
" split=SplitMetadata('split', 'test'),\n",
9795
" )\n",
9896
"\n",
9997
" matcher = CosineSimilarity()\n",

baselines/inference/MegaDescriptor-L-224.ipynb

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
"from timm import create_model\n",
2222
"\n",
2323
"from wildlife_tools.features import DeepFeatures\n",
24-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
24+
"from wildlife_tools.data import WildlifeDataset\n",
2525
"from wildlife_tools.similarity import CosineSimilarity\n",
2626
"from wildlife_tools.inference import KnnClassifier\n",
2727
"\n",
@@ -67,7 +67,7 @@
6767
},
6868
{
6969
"cell_type": "code",
70-
"execution_count": 2,
70+
"execution_count": null,
7171
"id": "99691108",
7272
"metadata": {
7373
"scrolled": true
@@ -521,19 +521,18 @@
521521
" ])\n",
522522
"\n",
523523
" database = WildlifeDataset(\n",
524-
" metadata=metadata,\n",
524+
" metadata=metadata.query('split == \"train\"'),\n",
525525
" root=f'{root_images}/{name}/',\n",
526526
" transform=transform,\n",
527-
" split=SplitMetadata('split', 'train'),\n",
528527
" )\n",
529528
"\n",
530529
" query = WildlifeDataset(\n",
531-
" metadata=metadata,\n",
530+
" metadata=metadata.query('split == \"test\"'),\n",
532531
" root=f'{root_images}/{name}/',\n",
533532
" transform=transform,\n",
534-
" split=SplitMetadata('split', 'test'),\n",
535533
" )\n",
536534
"\n",
535+
"\n",
537536
" matcher = CosineSimilarity()\n",
538537
" similarity = matcher(query=extractor(query), database=extractor(database))\n",
539538
" preds = KnnClassifier(k=1, database_labels=database.labels_string)(similarity['cosine'])\n",

baselines/inference/MegaDescriptor-L-384.ipynb

Lines changed: 5 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
},
1212
{
1313
"cell_type": "code",
14-
"execution_count": 1,
14+
"execution_count": null,
1515
"id": "c7f400b1",
1616
"metadata": {},
1717
"outputs": [
@@ -30,7 +30,7 @@
3030
"from timm import create_model\n",
3131
"\n",
3232
"from wildlife_tools.features import DeepFeatures\n",
33-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
33+
"from wildlife_tools.data import WildlifeDataset\n",
3434
"from wildlife_tools.similarity import CosineSimilarity\n",
3535
"from wildlife_tools.inference import KnnClassifier\n",
3636
"\n",
@@ -76,7 +76,7 @@
7676
},
7777
{
7878
"cell_type": "code",
79-
"execution_count": 2,
79+
"execution_count": null,
8080
"id": "99691108",
8181
"metadata": {},
8282
"outputs": [
@@ -528,17 +528,15 @@
528528
" ])\n",
529529
"\n",
530530
" database = WildlifeDataset(\n",
531-
" metadata=metadata,\n",
531+
" metadata=metadata.query('split == \"train\"'),\n",
532532
" root=f'{root_images}/{name}/',\n",
533533
" transform=transform,\n",
534-
" split=SplitMetadata('split', 'train'),\n",
535534
" )\n",
536535
"\n",
537536
" query = WildlifeDataset(\n",
538-
" metadata=metadata,\n",
537+
" metadata=metadata.query('split == \"test\"'),\n",
539538
" root=f'{root_images}/{name}/',\n",
540539
" transform=transform,\n",
541-
" split=SplitMetadata('split', 'test'),\n",
542540
" )\n",
543541
"\n",
544542
" matcher = CosineSimilarity()\n",

baselines/inference/MegaDescriptor-S-224.ipynb

Lines changed: 6 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@
1111
},
1212
{
1313
"cell_type": "code",
14-
"execution_count": 3,
14+
"execution_count": null,
1515
"id": "904f6e6c",
1616
"metadata": {},
1717
"outputs": [],
@@ -21,7 +21,7 @@
2121
"from timm import create_model\n",
2222
"\n",
2323
"from wildlife_tools.features import DeepFeatures\n",
24-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
24+
"from wildlife_tools.data import WildlifeDataset\n",
2525
"from wildlife_tools.similarity import CosineSimilarity\n",
2626
"from wildlife_tools.inference import KnnClassifier\n",
2727
"\n",
@@ -67,7 +67,7 @@
6767
},
6868
{
6969
"cell_type": "code",
70-
"execution_count": 2,
70+
"execution_count": null,
7171
"id": "e4381543",
7272
"metadata": {},
7373
"outputs": [
@@ -519,19 +519,18 @@
519519
" ])\n",
520520
"\n",
521521
" database = WildlifeDataset(\n",
522-
" metadata=metadata,\n",
522+
" metadata=metadata.query('split == \"train\"'),\n",
523523
" root=f'{root_images}/{name}/',\n",
524524
" transform=transform,\n",
525-
" split=SplitMetadata('split', 'train'),\n",
526525
" )\n",
527526
"\n",
528527
" query = WildlifeDataset(\n",
529-
" metadata=metadata,\n",
528+
" metadata=metadata.query('split == \"test\"'),\n",
530529
" root=f'{root_images}/{name}/',\n",
531530
" transform=transform,\n",
532-
" split=SplitMetadata('split', 'test'),\n",
533531
" )\n",
534532
"\n",
533+
"\n",
535534
" matcher = CosineSimilarity()\n",
536535
" similarity = matcher(query=extractor(query), database=extractor(database))\n",
537536
" preds = KnnClassifier(k=1, database_labels=database.labels_string)(similarity['cosine'])\n",

baselines/inference/MegaDescriptor-T-224.ipynb

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
"from timm import create_model\n",
2222
"\n",
2323
"from wildlife_tools.features import DeepFeatures\n",
24-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
24+
"from wildlife_tools.data import WildlifeDataset\n",
2525
"from wildlife_tools.similarity import CosineSimilarity\n",
2626
"from wildlife_tools.inference import KnnClassifier\n",
2727
"\n",
@@ -67,7 +67,7 @@
6767
},
6868
{
6969
"cell_type": "code",
70-
"execution_count": 2,
70+
"execution_count": null,
7171
"id": "8d8543ef",
7272
"metadata": {},
7373
"outputs": [
@@ -519,17 +519,15 @@
519519
" ])\n",
520520
"\n",
521521
" database = WildlifeDataset(\n",
522-
" metadata=metadata,\n",
522+
" metadata=metadata.query('split == \"train\"'),\n",
523523
" root=f'{root_images}/{name}/',\n",
524524
" transform=transform,\n",
525-
" split=SplitMetadata('split', 'train'),\n",
526525
" )\n",
527526
"\n",
528527
" query = WildlifeDataset(\n",
529-
" metadata=metadata,\n",
528+
" metadata=metadata.query('split == \"test\"'),\n",
530529
" root=f'{root_images}/{name}/',\n",
531530
" transform=transform,\n",
532-
" split=SplitMetadata('split', 'test'),\n",
533531
" )\n",
534532
"\n",
535533
" matcher = CosineSimilarity()\n",

baselines/training/MegaDescriptor-B-224.ipynb

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
"import pandas as pd\n",
2323
"import torchvision.transforms as T\n",
2424
"from torch.optim import SGD\n",
25-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
25+
"from wildlife_tools.data import WildlifeDataset\n",
2626
"from wildlife_tools.train import ArcFaceLoss, BasicTrainer\n",
2727
"\n",
2828
"\n",
@@ -36,9 +36,8 @@
3636
" T.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)),\n",
3737
"])\n",
3838
"dataset = WildlifeDataset(\n",
39-
" metadata = metadata, \n",
39+
" metadata = metadata.query('split == \"train\"'), \n",
4040
" root = image_root,\n",
41-
" split = SplitMetadata('split', 'train'),\n",
4241
" transform=transform\n",
4342
")\n",
4443
"\n",

baselines/training/MegaDescriptor-L-224.ipynb

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
"import pandas as pd\n",
2323
"import torchvision.transforms as T\n",
2424
"from torch.optim import SGD\n",
25-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
25+
"from wildlife_tools.data import WildlifeDataset\n",
2626
"from wildlife_tools.train import ArcFaceLoss, BasicTrainer\n",
2727
"\n",
2828
"\n",
@@ -36,13 +36,11 @@
3636
" T.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)),\n",
3737
"])\n",
3838
"dataset = WildlifeDataset(\n",
39-
" metadata = metadata, \n",
39+
" metadata = metadata.query('split == \"train\"'), \n",
4040
" root = image_root,\n",
41-
" split = SplitMetadata('split', 'train'),\n",
4241
" transform=transform\n",
4342
")\n",
4443
"\n",
45-
"\n",
4644
"# Backbone and loss configuration\n",
4745
"backbone = timm.create_model('swin_large_patch4_window7_224', num_classes=0, pretrained=True)\n",
4846
"with torch.no_grad():\n",

baselines/training/MegaDescriptor-L-384.ipynb

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
"import pandas as pd\n",
2323
"import torchvision.transforms as T\n",
2424
"from torch.optim import SGD\n",
25-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
25+
"from wildlife_tools.data import WildlifeDataset\n",
2626
"from wildlife_tools.train import ArcFaceLoss, BasicTrainer\n",
2727
"\n",
2828
"\n",
@@ -36,9 +36,8 @@
3636
" T.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)),\n",
3737
"])\n",
3838
"dataset = WildlifeDataset(\n",
39-
" metadata = metadata, \n",
39+
" metadata = metadata.query('split == \"train\"'), \n",
4040
" root = image_root,\n",
41-
" split = SplitMetadata('split', 'train'),\n",
4241
" transform=transform\n",
4342
")\n",
4443
"\n",

baselines/training/MegaDescriptor-S-224.ipynb

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
"import pandas as pd\n",
2323
"import torchvision.transforms as T\n",
2424
"from torch.optim import SGD\n",
25-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
25+
"from wildlife_tools.data import WildlifeDataset\n",
2626
"from wildlife_tools.train import ArcFaceLoss, BasicTrainer\n",
2727
"\n",
2828
"\n",
@@ -36,9 +36,8 @@
3636
" T.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)),\n",
3737
"])\n",
3838
"dataset = WildlifeDataset(\n",
39-
" metadata = metadata, \n",
39+
" metadata = metadata.query('split == \"train\"'), \n",
4040
" root = image_root,\n",
41-
" split = SplitMetadata('split', 'train'),\n",
4241
" transform=transform\n",
4342
")\n",
4443
"\n",

baselines/training/MegaDescriptor-T-224.ipynb

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@
2222
"import pandas as pd\n",
2323
"import torchvision.transforms as T\n",
2424
"from torch.optim import SGD\n",
25-
"from wildlife_tools.data import WildlifeDataset, SplitMetadata\n",
25+
"from wildlife_tools.data import WildlifeDataset\n",
2626
"from wildlife_tools.train import ArcFaceLoss, BasicTrainer\n",
2727
"\n",
2828
"\n",
@@ -36,9 +36,8 @@
3636
" T.Normalize(mean=(0.485, 0.456, 0.406), std=(0.229, 0.224, 0.225)),\n",
3737
"])\n",
3838
"dataset = WildlifeDataset(\n",
39-
" metadata = metadata, \n",
39+
" metadata = metadata.query('split == \"train\"'), \n",
4040
" root = image_root,\n",
41-
" split = SplitMetadata('split', 'train'),\n",
4241
" transform=transform\n",
4342
")\n",
4443
"\n",

0 commit comments

Comments
 (0)