@@ -55,7 +55,7 @@ Major features include:
55
55
- Provides [ Backend API] ( https://github.com/triton-inference-server/backend ) that
56
56
allows adding custom backends and pre/post processing operations
57
57
- Supports writing custom backends in python, a.k.a.
58
- [ Python-based backends.] ( https://github.com/triton-inference-server/backend/blob/main /docs/python_based_backends.md#python-based-backends )
58
+ [ Python-based backends.] ( https://github.com/triton-inference-server/backend/blob/r25.04 /docs/python_based_backends.md#python-based-backends )
59
59
- Model pipelines using
60
60
[ Ensembling] ( docs/user_guide/architecture.md#ensemble-models ) or [ Business
61
61
Logic Scripting
@@ -167,10 +167,10 @@ configuration](docs/user_guide/model_configuration.md) for the model.
167
167
[ Python] ( https://github.com/triton-inference-server/python_backend ) , and more
168
168
- Not all the above backends are supported on every platform supported by Triton.
169
169
Look at the
170
- [ Backend-Platform Support Matrix] ( https://github.com/triton-inference-server/backend/blob/main /docs/backend_platform_support_matrix.md )
170
+ [ Backend-Platform Support Matrix] ( https://github.com/triton-inference-server/backend/blob/r25.04 /docs/backend_platform_support_matrix.md )
171
171
to learn which backends are supported on your target platform.
172
172
- Learn how to [ optimize performance] ( docs/user_guide/optimization.md ) using the
173
- [ Performance Analyzer] ( https://github.com/triton-inference-server/perf_analyzer/blob/main /README.md )
173
+ [ Performance Analyzer] ( https://github.com/triton-inference-server/perf_analyzer/blob/r25.04 /README.md )
174
174
and
175
175
[ Model Analyzer] ( https://github.com/triton-inference-server/model_analyzer )
176
176
- Learn how to [ manage loading and unloading models] ( docs/user_guide/model_management.md ) in
@@ -184,14 +184,14 @@ A Triton *client* application sends inference and other requests to Triton. The
184
184
[ Python and C++ client libraries] ( https://github.com/triton-inference-server/client )
185
185
provide APIs to simplify this communication.
186
186
187
- - Review client examples for [ C++] ( https://github.com/triton-inference-server/client/blob/main /src/c%2B%2B/examples ) ,
188
- [ Python] ( https://github.com/triton-inference-server/client/blob/main /src/python/examples ) ,
189
- and [ Java] ( https://github.com/triton-inference-server/client/blob/main /src/java/src/main/java/triton/client/examples )
187
+ - Review client examples for [ C++] ( https://github.com/triton-inference-server/client/blob/r25.04 /src/c%2B%2B/examples ) ,
188
+ [ Python] ( https://github.com/triton-inference-server/client/blob/r25.04 /src/python/examples ) ,
189
+ and [ Java] ( https://github.com/triton-inference-server/client/blob/r25.04 /src/java/src/main/java/triton/client/examples )
190
190
- Configure [ HTTP] ( https://github.com/triton-inference-server/client#http-options )
191
191
and [ gRPC] ( https://github.com/triton-inference-server/client#grpc-options )
192
192
client options
193
193
- Send input data (e.g. a jpeg image) directly to Triton in the [ body of an HTTP
194
- request without any additional metadata] ( https://github.com/triton-inference-server/server/blob/main /docs/protocol/extension_binary_data.md#raw-binary-request )
194
+ request without any additional metadata] ( https://github.com/triton-inference-server/server/blob/r25.04 /docs/protocol/extension_binary_data.md#raw-binary-request )
195
195
196
196
### Extend Triton
197
197
@@ -200,7 +200,7 @@ designed for modularity and flexibility
200
200
201
201
- [ Customize Triton Inference Server container] ( docs/customization_guide/compose.md ) for your use case
202
202
- [ Create custom backends] ( https://github.com/triton-inference-server/backend )
203
- in either [ C/C++] ( https://github.com/triton-inference-server/backend/blob/main /README.md#triton-backend-api )
203
+ in either [ C/C++] ( https://github.com/triton-inference-server/backend/blob/r25.04 /README.md#triton-backend-api )
204
204
or [ Python] ( https://github.com/triton-inference-server/python_backend )
205
205
- Create [ decoupled backends and models] ( docs/user_guide/decoupled_models.md ) that can send
206
206
multiple responses for a request or not send any responses for a request
@@ -209,7 +209,7 @@ designed for modularity and flexibility
209
209
decryption, or conversion
210
210
- Deploy Triton on [ Jetson and JetPack] ( docs/user_guide/jetson.md )
211
211
- [ Use Triton on AWS
212
- Inferentia] ( https://github.com/triton-inference-server/python_backend/tree/main /inferentia )
212
+ Inferentia] ( https://github.com/triton-inference-server/python_backend/tree/r25.04 /inferentia )
213
213
214
214
### Additional Documentation
215
215
0 commit comments