-
Notifications
You must be signed in to change notification settings - Fork 19
feat: Review docs #749
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
feat: Review docs #749
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Completed tech pubs review of files, providing copyedits, formatting recommendations, and style guide feedback.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Completed tech pubs review of files, providing copyedits, formatting recommendations, and style guide feedback.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Resubmitting review comments on selected files
| @@ -1,16 +1,16 @@ | |||
| # LLM Pre-Training with NeMo AutoModel | |||
|
|
|||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LLM Pre-Training with NeMo Automodel
| # LLM Pre-Training with NeMo AutoModel | ||
|
|
||
| This guide covers **FineWeb** data preparation, **defining** a [NanoGPT‑style](https://github.com/KellerJordan/modded-nanogpt) model, and **launching and monitoring** a NeMo AutoModel pre‑training run. | ||
|
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
delete this section
docs/guides/llm/pretraining.md
Outdated
| In particular, it will show you how to: | ||
| In particular, it will show you how to: | ||
| 1. [Install NeMo AutoModel from git](#1-environment-setup). | ||
| 2. [Pre-process and tokenize the FineWeb dataset](#2-pre-process-the-fineweb-dataset). | ||
| 3. [Introduction to the NeMo AutoModel training workflow](#3-introduction-to-the-nemo-automodel-training-workflow). | ||
| 4. [Define your own model architecture](#4-define-your-own-model-architecture). | ||
| 5. [Inspect and adjust the YAML configuration](#5-inspect-and-adjust-the-yaml-configuration). | ||
| 6. [Launch training](#6-launch-training). | ||
| 7. [Monitoring and evaluation](#7-monitoring-and-evaluation). |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
delete this section
docs/guides/llm/pretraining.md
Outdated
|
|
||
| --- | ||
|
|
||
| ## 1. Environment setup |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| ## Set Up Your Environment |
docs/model-coverage/overview.md
Outdated
| - Add custom implementations to extend support to models not yet covered upstream. | ||
| - Provide optimized, extended or faster implementations for specific models while retaining the same AutoModel interface. | ||
|
|
||
| ## Supported Hugging Face Auto classes |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| ## Supported Hugging Face Auto Classes |
docs/model-coverage/overview.md
Outdated
| | `AutoModelForSequenceClassification`| Sequence Classification | WIP | Early support; interfaces may change. | | ||
|
|
||
|
|
||
| ## When a model listed on Hugging Face Hub may not be supported |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| ## Troubleshooting Unsupported Models |
docs/model-coverage/overview.md
Outdated
|
|
||
| ## When a model listed on Hugging Face Hub may not be supported | ||
|
|
||
| Sometimes a model listed on the Hugging Face Hub may not support finetuning in NeMo AutoModel. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| Sometimes a model listed on the Hugging Face hub may not support fine-tuning in NeMo Automodel. |
docs/model-coverage/overview.md
Outdated
| | Model upper-bounds transformer version, requiring older version | — | Open a new GitHub issue. | | ||
| | Unsupported checkpoint format | OSError: `meta-llama/Llama-2-70b` does not appear to have a file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt or flax_model.msgpack. | Open a new GitHub issue or request from the model publisher to share a safetensors checkpoint. | | ||
|
|
||
| These cases typically stem from upstream packaging or dependency constraints and would surface similarly when using `transformers` directly. AutoModel mirrors the familiar load/finetune semantics. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| These cases typically stem from upstream packaging or dependency constraints. You would encounter the same issues when using `transformers` directly, as Automodel mirrors the familiar load and fine-tune semantics. |
docs/model-coverage/overview.md
Outdated
|
|
||
| If you encounter any issue, you can try: | ||
|
|
||
| - Upgrade to a NeMo AutoModel release that supports the required `transformers` version. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
| - Upgrade to a NeMo Automodel release that supports the required `transformers` version. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Resubmitting review comments on selected files. Continue to get this error: Unexpected token '<', "<!DOCTYPE "... is not valid JSON
Co-authored-by: jgerh <163925524+jgerh@users.noreply.github.com> Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
Co-authored-by: jgerh <163925524+jgerh@users.noreply.github.com> Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
2d86254 to
9d21f68
Compare
|
/ok to test 9d21f68 |
|
/ok to test d79426d |
|
/ok to test f874220 |
Docs review for: