Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add an option to use dummy weights #33

Merged
merged 1 commit into from
Apr 9, 2023
Merged

Add an option to use dummy weights #33

merged 1 commit into from
Apr 9, 2023

Conversation

WoosukKwon
Copy link
Collaborator

No description provided.

@WoosukKwon WoosukKwon merged commit ee88a7e into main Apr 9, 2023
@WoosukKwon WoosukKwon deleted the dummy branch April 9, 2023 06:36
hongxiayang pushed a commit to hongxiayang/vllm that referenced this pull request Feb 13, 2024
tianyil1 pushed a commit to tianyil1/vllm that referenced this pull request Jun 5, 2024
* Bucketing/Warmup WIP

* Cleanup

* Revert "Fix model_output_idx on HPU (vllm-project#27)"

This reverts commit 90dfa92.

* Rework selected_token_indices fix to also work with block_size padding

* Simple prompt attention POC

* Remove cumsum

* MQA/GQA support for simple prompt_attention

* Cleanup

* Fix typo

* Restore profiling runs
dllehr-amd pushed a commit to dllehr-amd/vllm that referenced this pull request Jul 22, 2024
…ernel tuning script for rocm.

Merge pull request vllm-project#33  - tuned moe configs v2
bigPYJ1151 pushed a commit to bigPYJ1151/vllm that referenced this pull request Jul 31, 2024
@alixiaodi alixiaodi mentioned this pull request Aug 2, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant