Skip to content
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: ufxpri/python_backend
Failed to load repositories. Confirm that selected base ref is valid, then try again.
Loading
base: main
Choose a base ref
...
head repository: triton-inference-server/python_backend
Failed to load repositories. Confirm that selected head ref is valid, then try again.
Loading
compare: main
Choose a head ref
Checking mergeability… Don’t worry, you can still create the pull request.
  • 13 commits
  • 34 files changed
  • 9 contributors

Commits on Jan 25, 2025

  1. feat: Add parameters support to InferResponse (triton-inference-serve…

    …r#394)
    
    * Add parameters support to InferResponse
    
    * Infer response to track parameters
    
    * Add parameters to binding infer response
    
    * Rank parameters argument up among InferResponse constructor arguments
    
    * Add setting parameters to Triton response
    
    * Send response parameters only on non-error
    
    * Fix double declaration
    
    * Unify py dictionary parameters to json str
    
    * Add documentation
    
    * Mark response parameters accessor const and JSON serializable
    
    * [Docs] Note BLS response parameters are not populated currently
    
    * [comment] Clarify why PbTensor::LoadFromSharedMemory() requires holding GIL
    kthui authored Jan 25, 2025
    Configuration menu
    Copy the full SHA
    1ea48a6 View commit details
    Browse the repository at this point in the history

Commits on Feb 6, 2025

  1. Configuration menu
    Copy the full SHA
    f61d423 View commit details
    Browse the repository at this point in the history

Commits on Feb 20, 2025

  1. Configuration menu
    Copy the full SHA
    1b797d6 View commit details
    Browse the repository at this point in the history

Commits on Apr 8, 2025

  1. feat: add BLS decoupled response iterator's cancel() method for the r…

    …equest cancellation (triton-inference-server#398)
    
    Adding a cancel() method to the BLS decoupled response iterator, so that it will be able to cancel the Triton server inference request corresponding to the response iterator if the stub process gets the enough response from the response iterator.
    
    Due to each stub InferenceRequest object can create multiple BLS Triton Server inference requests, so adding cancel() to the response iterator would be more feasible to manage cancelling individual request rather than cancelling all requests generated with the stub InferenceRequest object.
    
    More details can be found in the change of the README.md
    richardhuo-nv authored Apr 8, 2025
    Configuration menu
    Copy the full SHA
    7f21b67 View commit details
    Browse the repository at this point in the history

Commits on May 7, 2025

  1. Configuration menu
    Copy the full SHA
    bb82100 View commit details
    Browse the repository at this point in the history

Commits on Jun 30, 2025

  1. Configuration menu
    Copy the full SHA
    87f6f2a View commit details
    Browse the repository at this point in the history

Commits on Jul 7, 2025

  1. Configuration menu
    Copy the full SHA
    8cfdffe View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    7d1333e View commit details
    Browse the repository at this point in the history

Commits on Jul 15, 2025

  1. Configuration menu
    Copy the full SHA
    1ab97b7 View commit details
    Browse the repository at this point in the history

Commits on Aug 7, 2025

  1. Configuration menu
    Copy the full SHA
    8f2df55 View commit details
    Browse the repository at this point in the history

Commits on Aug 13, 2025

  1. fix: Add input validation to model load (triton-inference-server#404)

    Validate input parameters used within python_backend model load
    mattwittwer authored Aug 13, 2025
    Configuration menu
    Copy the full SHA
    8b5a055 View commit details
    Browse the repository at this point in the history

Commits on Oct 8, 2025

  1. perf: optimize string tensor deserialization with high performance c+…

    …+ implementation (triton-inference-server#416)
    
    * perf: optimize string tensor deserialization with high performance c++ implementation
    
    * Address PR comments
    
    ---------
    
    Co-authored-by: Wei Chen <[email protected]>
    wweic and Wei Chen authored Oct 8, 2025
    Configuration menu
    Copy the full SHA
    389c770 View commit details
    Browse the repository at this point in the history

Commits on Nov 4, 2025

  1. Configuration menu
    Copy the full SHA
    0f65dd6 View commit details
    Browse the repository at this point in the history
Loading