@@ -57,7 +57,7 @@ Major features include:
57
57
- Provides [ Backend API] ( https://github.com/triton-inference-server/backend ) that
58
58
allows adding custom backends and pre/post processing operations
59
59
- Supports writing custom backends in python, a.k.a.
60
- [ Python-based backends.] ( https://github.com/triton-inference-server/backend/blob/main /docs/python_based_backends.md#python-based-backends )
60
+ [ Python-based backends.] ( https://github.com/triton-inference-server/backend/blob/r24.10 /docs/python_based_backends.md#python-based-backends )
61
61
- Model pipelines using
62
62
[ Ensembling] ( docs/user_guide/architecture.md#ensemble-models ) or [ Business
63
63
Logic Scripting
@@ -170,10 +170,10 @@ configuration](docs/user_guide/model_configuration.md) for the model.
170
170
[ Python] ( https://github.com/triton-inference-server/python_backend ) , and more
171
171
- Not all the above backends are supported on every platform supported by Triton.
172
172
Look at the
173
- [ Backend-Platform Support Matrix] ( https://github.com/triton-inference-server/backend/blob/main /docs/backend_platform_support_matrix.md )
173
+ [ Backend-Platform Support Matrix] ( https://github.com/triton-inference-server/backend/blob/r24.10 /docs/backend_platform_support_matrix.md )
174
174
to learn which backends are supported on your target platform.
175
175
- Learn how to [ optimize performance] ( docs/user_guide/optimization.md ) using the
176
- [ Performance Analyzer] ( https://github.com/triton-inference-server/perf_analyzer/blob/main /README.md )
176
+ [ Performance Analyzer] ( https://github.com/triton-inference-server/perf_analyzer/blob/r24.10 /README.md )
177
177
and
178
178
[ Model Analyzer] ( https://github.com/triton-inference-server/model_analyzer )
179
179
- Learn how to [ manage loading and unloading models] ( docs/user_guide/model_management.md ) in
@@ -187,14 +187,14 @@ A Triton *client* application sends inference and other requests to Triton. The
187
187
[ Python and C++ client libraries] ( https://github.com/triton-inference-server/client )
188
188
provide APIs to simplify this communication.
189
189
190
- - Review client examples for [ C++] ( https://github.com/triton-inference-server/client/blob/main /src/c%2B%2B/examples ) ,
191
- [ Python] ( https://github.com/triton-inference-server/client/blob/main /src/python/examples ) ,
192
- and [ Java] ( https://github.com/triton-inference-server/client/blob/main /src/java/src/main/java/triton/client/examples )
190
+ - Review client examples for [ C++] ( https://github.com/triton-inference-server/client/blob/r24.10 /src/c%2B%2B/examples ) ,
191
+ [ Python] ( https://github.com/triton-inference-server/client/blob/r24.10 /src/python/examples ) ,
192
+ and [ Java] ( https://github.com/triton-inference-server/client/blob/r24.10 /src/java/src/main/java/triton/client/examples )
193
193
- Configure [ HTTP] ( https://github.com/triton-inference-server/client#http-options )
194
194
and [ gRPC] ( https://github.com/triton-inference-server/client#grpc-options )
195
195
client options
196
196
- Send input data (e.g. a jpeg image) directly to Triton in the [ body of an HTTP
197
- request without any additional metadata] ( https://github.com/triton-inference-server/server/blob/main /docs/protocol/extension_binary_data.md#raw-binary-request )
197
+ request without any additional metadata] ( https://github.com/triton-inference-server/server/blob/r24.10 /docs/protocol/extension_binary_data.md#raw-binary-request )
198
198
199
199
### Extend Triton
200
200
@@ -203,7 +203,7 @@ designed for modularity and flexibility
203
203
204
204
- [ Customize Triton Inference Server container] ( docs/customization_guide/compose.md ) for your use case
205
205
- [ Create custom backends] ( https://github.com/triton-inference-server/backend )
206
- in either [ C/C++] ( https://github.com/triton-inference-server/backend/blob/main /README.md#triton-backend-api )
206
+ in either [ C/C++] ( https://github.com/triton-inference-server/backend/blob/r24.10 /README.md#triton-backend-api )
207
207
or [ Python] ( https://github.com/triton-inference-server/python_backend )
208
208
- Create [ decoupled backends and models] ( docs/user_guide/decoupled_models.md ) that can send
209
209
multiple responses for a request or not send any responses for a request
@@ -212,7 +212,7 @@ designed for modularity and flexibility
212
212
decryption, or conversion
213
213
- Deploy Triton on [ Jetson and JetPack] ( docs/user_guide/jetson.md )
214
214
- [ Use Triton on AWS
215
- Inferentia] ( https://github.com/triton-inference-server/python_backend/tree/main /inferentia )
215
+ Inferentia] ( https://github.com/triton-inference-server/python_backend/tree/r24.10 /inferentia )
216
216
217
217
### Additional Documentation
218
218
0 commit comments