[FLINK-38857][Model] Add docs for Triton inference model #27490
+1,010
−0
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What is the purpose of the change
This PR adds comprehensive documentation for the Triton inference model in Flink SQL, enabling users to perform real-time model inference by calling NVIDIA Triton Inference Server from Flink.
The documentation introduces how to define, configure, and use Triton-backed models in Flink SQL, covering common inference scenarios as well as advanced production use cases.
What is the change
docs/connectors/modelsCREATE MODELsyntaxML_PREDICTusage for model inferenceWhy is the change needed
Triton is a widely used, high-performance inference serving system supporting multiple ML frameworks.
Providing first-class documentation for Triton model integration helps users:
How was this change tested
Does this PR introduce any user-facing change
Yes.
This PR introduces new user-facing documentation describing how to use Triton inference models in Flink SQL.
Checklist