Skip to content

Commit 7d3d8ae

Browse files
committed
finale?
1 parent 2efaadd commit 7d3d8ae

File tree

2 files changed

+6
-8
lines changed

2 files changed

+6
-8
lines changed

README.md

Lines changed: 6 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -407,7 +407,7 @@ Tensorboard: Sep03_11-23-24_bert_smart
407407
| BERT-SMART | 51.6 % | 88.8 % | 43.8 % |
408408

409409
The bad sts correlation is because SMART uses MSE loss for its calculation of adverserial loss.
410-
We did not change it yet.
410+
We did not change it **yet**.
411411

412412
### Tuning SMART
413413
We did another Optuna SMART run for base BERT.
@@ -429,7 +429,7 @@ python -u optuna_smart.py --use_gpu --batch_size 50 --objective para --one_embed
429429
| BERT-STS | 27.29 | 6.65e-6 | 0.0002 | 7.84e-6 | L1 |
430430

431431
The bad sts correlation is because SMART uses MSE loss for its calculation of adverserial loss.
432-
We did not change it yet.
432+
We did not change it **yet**.
433433

434434
### Final model
435435
We combined some of our results in the final model.
@@ -516,22 +516,20 @@ This could be achieved be generating more (true) data from the datasets sst and
516516
- give other losses different weights.
517517
- with or without combined losses.
518518
- maybe based in dev_acc performance in previous epoch.
519-
520-
## Contributing
521-
522-
>📋 Pick a licence and describe how to contribute to your code repository.
519+
- implement SMART for BERT-STS
520+
- Dropout and weight decay tuning for BERT (AdamW and Sophia)
523521

524522
## Member Contributions
525523
Dawor, Moataz: Generalisations on Custom Attention, Splitted and reordererd batches, analysis_dataset
526524

527525
Lübbers, Christopher L.: Part 1 complete; Part 2: sBERT, Tensorboard (metrics + profiler), sBERT-Baseline, SOPHIA, SMART, Optuna, sBERT-Optuna for Optimizer, Optuna for sBERT and BERT-SMART, Optuna for sBERT-regularization, sBERT with combinded losses, sBERT with gradient surgery, README for those tasks
528526

529-
Niegsch, Luaks*: Generalisations on Custom Attention, Splitted and reordererd batches,
527+
Niegsch, Lukas*: Generalisations on Custom Attention, Splitted and reordererd batches,
530528

531529
Schmidt, Finn Paul:
532530

533531

534-
##Submit commands
532+
## Submit commands
535533

536534
Für sophia base mit optimierten parametern zu trainieren:
537535
```

0 commit comments

Comments
 (0)