Skip to content
This repository has been archived by the owner on Oct 11, 2024. It is now read-only.

Commit

Permalink
updates for nm-magic-wand, nightly or release
Browse files Browse the repository at this point in the history
  • Loading branch information
andy-neuma committed May 16, 2024
1 parent 2fcfced commit 1c97968
Show file tree
Hide file tree
Showing 4 changed files with 4 additions and 56 deletions.
1 change: 0 additions & 1 deletion .github/actions/nm-build-vllm/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,6 @@ runs:
VENV="${{ inputs.venv }}-${COMMIT:0:7}"
source $(pyenv root)/versions/${{ inputs.python }}/envs/${VENV}/bin/activate
# TODO: adjust when we need a proper release. use nightly now.
pip3 install nm-magic-wand-nightly
pip3 install -r requirements-cuda.txt -r requirements-build.txt
# build
SUCCESS=0
Expand Down
3 changes: 3 additions & 0 deletions .github/actions/nm-install-test-whl/action.yml
Original file line number Diff line number Diff line change
Expand Up @@ -52,6 +52,9 @@ runs:
pip3 install ${WHL}[sparse]
# report magic_wand version
MAGIC_WAND=$(pip3 show nm-magic-wand-nightly | grep "Version" | cut -d' ' -f2)
if [ -z "${MAGIC_WAND}" ]; then
MAGIC_WAND=$(pip3 show nm-magic-wand | grep "Version" | cut -d' ' -f2)
fi
echo "magic_wand=${MAGIC_WAND}" >> "$GITHUB_OUTPUT"
# test and collect code coverage
SUCCESS=0
Expand Down
43 changes: 0 additions & 43 deletions .github/actions/nm-test-vllm/action.yml

This file was deleted.

13 changes: 1 addition & 12 deletions .github/scripts/build
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,6 @@ usage() {
echo
echo "usage: ${0} <options>"
echo
echo " -a - pypi server address"
echo " -p - python version"
echo " -v - name for virtualenv"
echo " -h - this list of options"
Expand All @@ -17,15 +16,12 @@ PYPI_IP=
PYTHON=
VENV=

while getopts "ha:p:v:" OPT; do
while getopts "hp:v:" OPT; do
case "${OPT}" in
h)
usage
exit 1
;;
a)
PYPI_IP="${OPTARG}"
;;
p)
PYTHON="${OPTARG}"
;;
Expand All @@ -36,12 +32,6 @@ while getopts "ha:p:v:" OPT; do
done

# check if variables are valid
if [ -z "${PYPI_IP}" ]; then
echo "please provide 'pypi' server address"
usage
exit 1
fi

if [ -z "${PYTHON}" ]; then
echo "please provide python version, e.g. 3.10.12"
usage
Expand All @@ -55,6 +45,5 @@ if [ -z "${VENV}" ]; then
fi

source $(pyenv root)/versions/${PYTHON}/envs/${VENV}/bin/activate
pip3 install --index-url http://${PYPI_IP}:8080/ --trusted-host ${PYPI_IP} nm-magic-wand-nightly
pip3 install -r requirements-cuda.txt -r requirements-build.txt
pip3 install -e .

1 comment on commit 1c97968

@github-actions
Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

bigger_is_better

Benchmark suite Current: 1c97968 Previous: 6334dd3 Ratio
{"name": "request_throughput", "description": "VLLM Engine throughput - synthetic\nmodel - NousResearch/Llama-2-7b-chat-hf\nmax_model_len - 4096\nbenchmark_throughput {\n \"use-all-available-gpus_\": \"\",\n \"input-len\": 256,\n \"output-len\": 128,\n \"num-prompts\": 1000\n}", "gpu_description": "NVIDIA A10G x 1", "vllm_version": "0.3.0", "python_version": "3.10.12 (main, May 10 2024, 13:42:25) [GCC 9.4.0]", "torch_version": "2.3.0+cu121"} 3.8374945566455483 prompts/s 3.8418198063652103 prompts/s 1.00
{"name": "token_throughput", "description": "VLLM Engine throughput - synthetic\nmodel - NousResearch/Llama-2-7b-chat-hf\nmax_model_len - 4096\nbenchmark_throughput {\n \"use-all-available-gpus_\": \"\",\n \"input-len\": 256,\n \"output-len\": 128,\n \"num-prompts\": 1000\n}", "gpu_description": "NVIDIA A10G x 1", "vllm_version": "0.3.0", "python_version": "3.10.12 (main, May 10 2024, 13:42:25) [GCC 9.4.0]", "torch_version": "2.3.0+cu121"} 1473.5979097518905 tokens/s 1475.2588056442407 tokens/s 1.00

This comment was automatically generated by workflow using github-action-benchmark.

Please sign in to comment.