Skip to content

Conversation

@akoumpa
Copy link
Contributor

@akoumpa akoumpa commented Oct 31, 2025

@copy-pr-bot
Copy link

copy-pr-bot bot commented Oct 31, 2025

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

@akoumpa akoumpa requested a review from jgerh October 31, 2025 19:54
@akoumpa akoumpa changed the title feat: Update docs feat: Review docs Oct 31, 2025
Copy link
Contributor

@jgerh jgerh left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Completed tech pubs review of files, providing copyedits, formatting recommendations, and style guide feedback.

Copy link
Contributor

@jgerh jgerh left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Completed tech pubs review of files, providing copyedits, formatting recommendations, and style guide feedback.

Copy link
Contributor

@jgerh jgerh left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Resubmitting review comments on selected files

@@ -1,16 +1,16 @@
# LLM Pre-Training with NeMo AutoModel

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LLM Pre-Training with NeMo Automodel

# LLM Pre-Training with NeMo AutoModel

This guide covers **FineWeb** data preparation, **defining** a [NanoGPT‑style](https://github.com/KellerJordan/modded-nanogpt) model, and **launching and monitoring** a NeMo AutoModel pre‑training run.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete this section

Comment on lines 5 to 12
In particular, it will show you how to:
In particular, it will show you how to:
1. [Install NeMo AutoModel from git](#1-environment-setup).
2. [Pre-process and tokenize the FineWeb dataset](#2-pre-process-the-fineweb-dataset).
3. [Introduction to the NeMo AutoModel training workflow](#3-introduction-to-the-nemo-automodel-training-workflow).
4. [Define your own model architecture](#4-define-your-own-model-architecture).
5. [Inspect and adjust the YAML configuration](#5-inspect-and-adjust-the-yaml-configuration).
6. [Launch training](#6-launch-training).
7. [Monitoring and evaluation](#7-monitoring-and-evaluation).
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

delete this section


---

## 1. Environment setup
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
## Set Up Your Environment

- Add custom implementations to extend support to models not yet covered upstream.
- Provide optimized, extended or faster implementations for specific models while retaining the same AutoModel interface.

## Supported Hugging Face Auto classes
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
## Supported Hugging Face Auto Classes

| `AutoModelForSequenceClassification`| Sequence Classification | WIP | Early support; interfaces may change. |


## When a model listed on Hugging Face Hub may not be supported
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
## Troubleshooting Unsupported Models


## When a model listed on Hugging Face Hub may not be supported

Sometimes a model listed on the Hugging Face Hub may not support finetuning in NeMo AutoModel.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Sometimes a model listed on the Hugging Face hub may not support fine-tuning in NeMo Automodel.

| Model upper-bounds transformer version, requiring older version || Open a new GitHub issue. |
| Unsupported checkpoint format | OSError: `meta-llama/Llama-2-70b` does not appear to have a file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt or flax_model.msgpack. | Open a new GitHub issue or request from the model publisher to share a safetensors checkpoint. |

These cases typically stem from upstream packaging or dependency constraints and would surface similarly when using `transformers` directly. AutoModel mirrors the familiar load/finetune semantics.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
These cases typically stem from upstream packaging or dependency constraints. You would encounter the same issues when using `transformers` directly, as Automodel mirrors the familiar load and fine-tune semantics.


If you encounter any issue, you can try:

- Upgrade to a NeMo AutoModel release that supports the required `transformers` version.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
- Upgrade to a NeMo Automodel release that supports the required `transformers` version.

Copy link
Contributor

@jgerh jgerh left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Resubmitting review comments on selected files. Continue to get this error: Unexpected token '<', "<!DOCTYPE "... is not valid JSON

@akoumpa akoumpa added the docs-only With great power comes great responsibility. label Nov 6, 2025
@akoumpa akoumpa requested a review from jgerh November 6, 2025 18:26
akoumpa and others added 4 commits November 6, 2025 10:26
Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
Co-authored-by: jgerh <163925524+jgerh@users.noreply.github.com>
Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
Co-authored-by: jgerh <163925524+jgerh@users.noreply.github.com>
Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
@akoumpa akoumpa force-pushed the akoumparouli/update_docs2 branch from 2d86254 to 9d21f68 Compare November 6, 2025 18:26
@akoumpa
Copy link
Contributor Author

akoumpa commented Nov 6, 2025

/ok to test 9d21f68

Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
@akoumpa
Copy link
Contributor Author

akoumpa commented Nov 6, 2025

/ok to test d79426d

Signed-off-by: Alexandros Koumparoulis <akoumparouli@nvidia.com>
@akoumpa
Copy link
Contributor Author

akoumpa commented Nov 6, 2025

/ok to test f874220

@akoumpa akoumpa marked this pull request as ready for review November 6, 2025 20:53
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

docs-only With great power comes great responsibility.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants