forked from triton-inference-server/server
-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Babak/upgrade triton to v2.43.0 #3
Open
babakbehzad
wants to merge
630
commits into
main
Choose a base branch
from
babak/upgrade-triton-to-v2.43.0
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
…ver#5911) * Add test for detecting S3 http2 upgrade request * Enhance testing * Copyright year update
…5922) * Add HPCX dependencies to search path * Copy hpcx to CPU-only container * Add ucc path to CPU-only image * Fixed if statement * Fix df variable * Combine hpcx LD_LIBRARY_PATH
…riton-inference-server#5915) * Add test case for metric lifetime error handling * Address comment * Use different MetricFamily name
…erver#5810) * Add testing for Pytorch instance group kind MODEL * Remove unused item * Update testing to verify the infer result * Add copyright * Remove unused import * Update pip install * Update the model to use the same add sub logic * Add torch multi-gpu and multi-device models to L0_io * Fix up model version
…rence-server#5937) * Add test for passing config via load api * Add more docs on instance update behavior * Update to suggested docs Co-authored-by: Ryan McCormick <[email protected]> * Use dictionary for json config * Modify the config fetched from Triton instead --------- Co-authored-by: Ryan McCormick <[email protected]>
…ence-server#5945) * Add redis config and use local logfile for redis server * Move redis log config to CLI * Have separate redis logs for unit tests and CLI tests
…ce-server#5885) * Add test on rate limiter max resource decrease update * Add test with explicit resource * Check server log for decreased resource limit
…ntrypoint updates (triton-inference-server#5910) * Allow changing ping behavior based on env variable in SageMaker * Add option for additional args * Make ping further configurable * Allow further configuration of grpc and http ports * Update docker/sagemaker/serve * Update docker/sagemaker/serve --------- Co-authored-by: GuanLuo <[email protected]>
…ce-server#5967) * Be more specific with MPI removal * Delete all libmpi libs
…rver#5963) * Add print statements for debugging * Add debugging print statements * Test using grpc client with stream to fix race * Use streaming client in all non-batch tests * Switch all clients to streaming GRPC * Remove unused imports, vars * Address comments * Remove random comment * Set inputs as separate function * Split set inputs based on test type
* Auto-format * Change to clang-format-15 in CONTRIBTUING
…-server#5976) * Add test for >1000 files * Capitalization for consistency * Add bucket cleaning at end * Move test pass/fail to end * Check number of files in model dir at load time
* Add testing for GPU tensor error handling * Fix up * Remove exit 0 * Fix jetson * Fix up
* Add test for Python BLS model loading API * Fix up
* Update trace_summery script * Remove GRPC_WAITREAD and Overhead
* Add gsutil cp retry helper function * Add max retry to GCS upload * Use simple sequential upload
…ver#6833) * Handle empty output * Add test case for 0 dimension output * Fix up number of tests
* tensorrt-llm benchmarking test
* Update README and versions for 2.42.0 / 24.01 (triton-inference-server#6789) * Update versions * Update README and versions for 2.42.0 / 24.01 * Fix documentaation genarion (triton-inference-server#6801) * Ser version of sphix to 5.0 * Set verions 5.0.0 * Update README.md and versions post 24.01
…und (triton-inference-server#6834) * Update miniconda version * Install pytest for different py version * Install pytest
* Add test for shutdown while loading * Fix intermittent failure on test_model_config_overwrite
Adding OpenTelemetry Batch Span Processor --------- Co-authored-by: Theo Clark <[email protected]> Co-authored-by: Ryan McCormick <[email protected]>
…erence-server#6755) * Support Double-Type Infer/Response Parameters
* Base Python Backend Support for Windows
…r#6886) * Update README and versions for 2.43.0 / 24.02 * Update Dockefile to reduce image size. * Update path in patch file for model generation
…ference-server#6873) (triton-inference-server#6881) * Add unit test reports to L0_dlpack_multi_gpu * Add unit test reports to L0_warmup
* Elimitated usage of onnx models in tests/ disabled some tests * Verified green: batch 1 * Verified tests: batch 2 * Verified tests: batch 3 * Verified tests: batch 4 * Verified tests: batch 5 * Verified tests: batch 6 * Verified tests: batch 7 * Verified tests: batch 8 * Verified tests: batch 9 * Verified tests: batch 10 * Verified tests: batch 11 * Verified tests: batch 12 * Verified tests: batch 12_follow up * Verified tests: batch 13 * Verified tests: batch 14 * Verified tests: batch 15 * Verified tests: batch 16 * Removed exits
* Revert to previous way of CMake installation * Win10: set python version back to 3.8.10
* Update README.md for 24.02 * Update version to 24.02
# }, | ||
"use_edit_page_button": False, | ||
"use_issues_button": True, | ||
"use_repository_button": True, |
Check warning
Code scanning / CodeQL
Duplicate key in dict literal Warning documentation
Dictionary key 'use_repository_button' is subsequently .
overwritten
Error loading related location
Loading
|
||
class Config(dict): |
Check warning
Code scanning / CodeQL
`__eq__` not overridden when adding attributes Warning
The class 'Config' does not override , but adds the new attribute .
'__eq__'
Error loading related location
Loading s3_regex
Error loading related location
Loading
babakbehzad
force-pushed
the
babak/upgrade-triton-to-v2.43.0
branch
from
March 23, 2024 02:04
4c6b7af
to
93973e7
Compare
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
No description provided.