Skip to content

Commit 371f7e4

Browse files
[Doc] Fix broken links and unlinked docs, add shortcuts to home sidebar (#18627)
Signed-off-by: DarkLight1337 <tlleungac@connect.ust.hk>
1 parent 15b45ff commit 371f7e4

File tree

6 files changed

+16
-11
lines changed

6 files changed

+16
-11
lines changed

docs/.nav.yml

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,13 @@ nav:
99
- getting_started/examples/offline_inference
1010
- getting_started/examples/online_serving
1111
- getting_started/examples/other
12-
- Roadmap: https://roadmap.vllm.ai
13-
- Releases: https://github.com/vllm-project/vllm/releases
12+
- Quick Links:
13+
- User Guide: serving/offline_inference.md
14+
- Developer Guide: contributing/overview.md
15+
- API Reference: api/README.md
16+
- Timeline:
17+
- Roadmap: https://roadmap.vllm.ai
18+
- Releases: https://github.com/vllm-project/vllm/releases
1419
- User Guide:
1520
- Inference and Serving:
1621
- serving/offline_inference.md
@@ -38,7 +43,7 @@ nav:
3843
- contributing/overview.md
3944
- glob: contributing/*
4045
flatten_single_child_sections: true
41-
- contributing/model
46+
- Model Implementation: contributing/model
4247
- Design Documents:
4348
- V0: design
4449
- V1: design/v1

docs/contributing/model/tests.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -33,14 +33,14 @@ These tests compare the model outputs of vLLM against [HF Transformers](https://
3333

3434
#### Generative models
3535

36-
For [generative models][generative-models], there are two levels of correctness tests, as defined in <gh-file:tests/models/utils.py>:
36+
For [generative models](../../models/generative_models.md), there are two levels of correctness tests, as defined in <gh-file:tests/models/utils.py>:
3737

3838
- Exact correctness (`check_outputs_equal`): The text outputted by vLLM should exactly match the text outputted by HF.
3939
- Logprobs similarity (`check_logprobs_close`): The logprobs outputted by vLLM should be in the top-k logprobs outputted by HF, and vice versa.
4040

4141
#### Pooling models
4242

43-
For [pooling models][pooling-models], we simply check the cosine similarity, as defined in <gh-file:tests/models/embedding/utils.py>.
43+
For [pooling models](../../models/pooling_models.md), we simply check the cosine similarity, as defined in <gh-file:tests/models/utils.py>.
4444

4545
[](){ #mm-processing-tests }
4646

docs/features/spec_decode.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -170,7 +170,7 @@ A variety of speculative models of this type are available on HF hub:
170170
## Speculating using EAGLE based draft models
171171

172172
The following code configures vLLM to use speculative decoding where proposals are generated by
173-
an [EAGLE (Extrapolation Algorithm for Greater Language-model Efficiency)](https://arxiv.org/pdf/2401.15077) based draft model. A more detailed example for offline mode, including how to extract request level acceptance rate, can be found [here](<gh-file:examples/offline_inference/eagle.py>).
173+
an [EAGLE (Extrapolation Algorithm for Greater Language-model Efficiency)](https://arxiv.org/pdf/2401.15077) based draft model. A more detailed example for offline mode, including how to extract request level acceptance rate, can be found [here](gh-file:examples/offline_inference/eagle.py).
174174

175175
```python
176176
from vllm import LLM, SamplingParams

docs/models/supported_models.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ title: Supported Models
33
---
44
[](){ #supported-models }
55

6-
vLLM supports [generative](generative-models) and [pooling](pooling-models) models across various tasks.
6+
vLLM supports [generative](./generative_models.md) and [pooling](./pooling_models.md) models across various tasks.
77
If a model supports more than one task, you can set the task via the `--task` argument.
88

99
For each task, we list the model architectures that have been implemented in vLLM.
@@ -376,7 +376,7 @@ Specified using `--task generate`.
376376

377377
### Pooling Models
378378

379-
See [this page](pooling-models) for more information on how to use pooling models.
379+
See [this page](./pooling_models.md) for more information on how to use pooling models.
380380

381381
!!! warning
382382
Since some model architectures support both generative and pooling tasks,
@@ -628,7 +628,7 @@ Specified using `--task generate`.
628628

629629
### Pooling Models
630630

631-
See [this page](pooling-models) for more information on how to use pooling models.
631+
See [this page](./pooling_models.md) for more information on how to use pooling models.
632632

633633
!!! warning
634634
Since some model architectures support both generative and pooling tasks,

docs/serving/openai_compatible_server.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ title: OpenAI-Compatible Server
55

66
vLLM provides an HTTP server that implements OpenAI's [Completions API](https://platform.openai.com/docs/api-reference/completions), [Chat API](https://platform.openai.com/docs/api-reference/chat), and more! This functionality lets you serve models and interact with them using an HTTP client.
77

8-
In your terminal, you can [install](../getting_started/installation.md) vLLM, then start the server with the [`vllm serve`][serve-args] command. (You can also use our [Docker][deployment-docker] image.)
8+
In your terminal, you can [install](../getting_started/installation/README.md) vLLM, then start the server with the [`vllm serve`][serve-args] command. (You can also use our [Docker][deployment-docker] image.)
99

1010
```bash
1111
vllm serve NousResearch/Meta-Llama-3-8B-Instruct --dtype auto --api-key token-abc123

docs/seed_parameter_behavior.md renamed to docs/serving/seed_parameter_behavior.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# Seed Parameter Behavior in vLLM
1+
# Seed Parameter Behavior
22

33
## Overview
44

0 commit comments

Comments
 (0)