-
Notifications
You must be signed in to change notification settings - Fork 0
Comparing changes
Open a pull request
base repository: ufxpri/python_backend
base: main
head repository: triton-inference-server/python_backend
compare: main
- 13 commits
- 34 files changed
- 9 contributors
Commits on Jan 25, 2025
-
feat: Add parameters support to InferResponse (triton-inference-serve…
…r#394) * Add parameters support to InferResponse * Infer response to track parameters * Add parameters to binding infer response * Rank parameters argument up among InferResponse constructor arguments * Add setting parameters to Triton response * Send response parameters only on non-error * Fix double declaration * Unify py dictionary parameters to json str * Add documentation * Mark response parameters accessor const and JSON serializable * [Docs] Note BLS response parameters are not populated currently * [comment] Clarify why PbTensor::LoadFromSharedMemory() requires holding GIL
Configuration menu - View commit details
-
Copy full SHA for 1ea48a6 - Browse repository at this point
Copy the full SHA 1ea48a6View commit details
Commits on Feb 6, 2025
-
Configuration menu - View commit details
-
Copy full SHA for f61d423 - Browse repository at this point
Copy the full SHA f61d423View commit details
Commits on Feb 20, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 1b797d6 - Browse repository at this point
Copy the full SHA 1b797d6View commit details
Commits on Apr 8, 2025
-
feat: add BLS decoupled response iterator's cancel() method for the r…
…equest cancellation (triton-inference-server#398) Adding a cancel() method to the BLS decoupled response iterator, so that it will be able to cancel the Triton server inference request corresponding to the response iterator if the stub process gets the enough response from the response iterator. Due to each stub InferenceRequest object can create multiple BLS Triton Server inference requests, so adding cancel() to the response iterator would be more feasible to manage cancelling individual request rather than cancelling all requests generated with the stub InferenceRequest object. More details can be found in the change of the README.md
Configuration menu - View commit details
-
Copy full SHA for 7f21b67 - Browse repository at this point
Copy the full SHA 7f21b67View commit details
Commits on May 7, 2025
-
Configuration menu - View commit details
-
Copy full SHA for bb82100 - Browse repository at this point
Copy the full SHA bb82100View commit details
Commits on Jun 30, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 87f6f2a - Browse repository at this point
Copy the full SHA 87f6f2aView commit details
Commits on Jul 7, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 8cfdffe - Browse repository at this point
Copy the full SHA 8cfdffeView commit details -
Configuration menu - View commit details
-
Copy full SHA for 7d1333e - Browse repository at this point
Copy the full SHA 7d1333eView commit details
Commits on Jul 15, 2025
-
build: TPRD-1622 - Update version to CMake 4.0 or 3.31 ( 3.31.8 ) whi…
…ch has support for CUDA new layout (triton-inference-server#407)
Configuration menu - View commit details
-
Copy full SHA for 1ab97b7 - Browse repository at this point
Copy the full SHA 1ab97b7View commit details
Commits on Aug 7, 2025
-
fix: Improve error handling for Python backend model initialization f…
…ailures (triton-inference-server#408)
Configuration menu - View commit details
-
Copy full SHA for 8f2df55 - Browse repository at this point
Copy the full SHA 8f2df55View commit details
Commits on Aug 13, 2025
-
fix: Add input validation to model load (triton-inference-server#404)
Validate input parameters used within python_backend model load
Configuration menu - View commit details
-
Copy full SHA for 8b5a055 - Browse repository at this point
Copy the full SHA 8b5a055View commit details
Commits on Oct 8, 2025
-
perf: optimize string tensor deserialization with high performance c+…
…+ implementation (triton-inference-server#416) * perf: optimize string tensor deserialization with high performance c++ implementation * Address PR comments --------- Co-authored-by: Wei Chen <[email protected]>
Configuration menu - View commit details
-
Copy full SHA for 389c770 - Browse repository at this point
Copy the full SHA 389c770View commit details
Commits on Nov 4, 2025
-
Configuration menu - View commit details
-
Copy full SHA for 0f65dd6 - Browse repository at this point
Copy the full SHA 0f65dd6View commit details
This comparison is taking too long to generate.
Unfortunately it looks like we can’t render this comparison for you right now. It might be too big, or there might be something weird with your repository.
You can try running this command locally to see the comparison on your machine:
git diff main...main