Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Babak/upgrade triton to v2.43.0 #3

Open
wants to merge 630 commits into
base: main
Choose a base branch
from

Conversation

babakbehzad
Copy link

No description provided.

kthui and others added 30 commits June 8, 2023 14:17
…ver#5911)

* Add test for detecting S3 http2 upgrade request

* Enhance testing

* Copyright year update
…5922)

* Add HPCX dependencies to search path

* Copy hpcx to CPU-only container

* Add ucc path to CPU-only image

* Fixed if statement

* Fix df variable

* Combine hpcx LD_LIBRARY_PATH
…riton-inference-server#5915)

* Add test case for metric lifetime error handling

* Address comment

* Use different MetricFamily name
…erver#5810)

* Add testing for Pytorch instance group kind MODEL

* Remove unused item

* Update testing to verify the infer result

* Add copyright

* Remove unused import

* Update pip install

* Update the model to use the same add sub logic

* Add torch multi-gpu and multi-device models to L0_io

* Fix up model version
…rence-server#5937)

* Add test for passing config via load api

* Add more docs on instance update behavior

* Update to suggested docs

Co-authored-by: Ryan McCormick <[email protected]>

* Use dictionary for json config

* Modify the config fetched from Triton instead

---------

Co-authored-by: Ryan McCormick <[email protected]>
…ence-server#5945)

* Add redis config and use local logfile for redis server

* Move redis log config to CLI

* Have separate redis logs for unit tests and CLI tests
…ce-server#5885)

* Add test on rate limiter max resource decrease update

* Add test with explicit resource

* Check server log for decreased resource limit
…ntrypoint updates (triton-inference-server#5910)

* Allow changing ping behavior based on env variable in SageMaker

* Add option for additional args

* Make ping further configurable

* Allow further configuration of grpc and http ports

* Update docker/sagemaker/serve

* Update docker/sagemaker/serve

---------

Co-authored-by: GuanLuo <[email protected]>
…ce-server#5967)

* Be more specific with MPI removal

* Delete all libmpi libs
…rver#5963)

* Add print statements for debugging

* Add debugging print statements

* Test using grpc client with stream to fix race

* Use streaming client in all non-batch tests

* Switch all clients to streaming GRPC

* Remove unused imports, vars

* Address comments

* Remove random comment

* Set inputs as separate function

* Split set inputs based on test type
* Auto-format

* Change to clang-format-15 in CONTRIBTUING
…-server#5976)

* Add test for >1000 files

* Capitalization for consistency

* Add bucket cleaning at end

* Move test pass/fail to end

* Check number of files in model dir at load time
* Add testing for GPU tensor error handling

* Fix up

* Remove exit 0

* Fix jetson

* Fix up
* Add test for Python BLS model loading API

* Fix up
rmccorm4 and others added 21 commits January 24, 2024 19:21
* Update trace_summery script

* Remove GRPC_WAITREAD and Overhead
* Add gsutil cp retry helper function

* Add max retry to GCS upload

* Use simple sequential upload
…ver#6833)

* Handle empty output

* Add test case for 0 dimension output

* Fix up number of tests
* Update README and versions for 2.42.0 / 24.01 (triton-inference-server#6789)

* Update versions

* Update README and versions for 2.42.0 / 24.01

* Fix documentaation genarion (triton-inference-server#6801)

* Ser version of sphix to 5.0

* Set verions 5.0.0

* Update README.md and versions post 24.01
…und (triton-inference-server#6834)

* Update miniconda version

* Install pytest for different py version

* Install pytest
* Add test for shutdown while loading

* Fix intermittent failure on test_model_config_overwrite
Adding OpenTelemetry Batch Span Processor
---------

Co-authored-by: Theo Clark <[email protected]>
Co-authored-by: Ryan McCormick <[email protected]>
* Base Python Backend Support for Windows
…r#6886)

* Update README and versions for 2.43.0 / 24.02

* Update Dockefile to reduce image size.

* Update path in patch file for model generation
…ference-server#6873) (triton-inference-server#6881)

* Add unit test reports to L0_dlpack_multi_gpu

* Add unit test reports to L0_warmup
* Elimitated usage of onnx models in tests/ disabled some tests

* Verified green: batch 1

* Verified tests: batch 2

* Verified tests: batch 3

* Verified tests: batch 4

* Verified tests: batch 5

* Verified tests: batch 6

* Verified tests: batch 7

* Verified tests: batch 8

* Verified tests: batch 9

* Verified tests: batch 10

* Verified tests: batch 11

* Verified tests: batch 12

* Verified tests: batch 12_follow up

* Verified tests: batch 13

* Verified tests: batch 14

* Verified tests: batch 15

* Verified tests: batch 16

* Removed exits
* Revert to previous way of CMake installation

* Win10: set python version back to 3.8.10
* Update README.md for 24.02

* Update version to 24.02
# },
"use_edit_page_button": False,
"use_issues_button": True,
"use_repository_button": True,

Check warning

Code scanning / CodeQL

Duplicate key in dict literal Warning documentation

Dictionary key 'use_repository_button' is subsequently
overwritten
.

class Config(dict):

Check warning

Code scanning / CodeQL

`__eq__` not overridden when adding attributes Warning

The class 'Config' does not override
'__eq__'
, but adds the new attribute
s3_regex
.
@babakbehzad babakbehzad force-pushed the babak/upgrade-triton-to-v2.43.0 branch from 4c6b7af to 93973e7 Compare March 23, 2024 02:04
@babakbehzad babakbehzad closed this Apr 5, 2024
@babakbehzad babakbehzad reopened this Apr 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.