Skip to content

Commit

Permalink
Migrate AquilaForCausalLM to LlamaForCausalLM (vllm-project#2867)
Browse files Browse the repository at this point in the history
  • Loading branch information
esmeetu authored and jimpang committed Feb 22, 2024
1 parent 3d35de8 commit bbda543
Show file tree
Hide file tree
Showing 5 changed files with 2 additions and 417 deletions.
5 changes: 2 additions & 3 deletions vllm/model_executor/models/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -10,8 +10,8 @@

# Architecture -> (module, class).
_MODELS = {
"AquilaModel": ("aquila", "AquilaForCausalLM"),
"AquilaForCausalLM": ("aquila", "AquilaForCausalLM"), # AquilaChat2
"AquilaModel": ("llama", "LlamaForCausalLM"),
"AquilaForCausalLM": ("llama", "LlamaForCausalLM"), # AquilaChat2
"BaiChuanForCausalLM": ("baichuan", "BaiChuanForCausalLM"), # baichuan-7b
"BaichuanForCausalLM": ("baichuan", "BaichuanForCausalLM"), # baichuan-13b
"BloomForCausalLM": ("bloom", "BloomForCausalLM"),
Expand Down Expand Up @@ -41,7 +41,6 @@
"Qwen2ForCausalLM": ("qwen2", "Qwen2ForCausalLM"),
"RWForCausalLM": ("falcon", "FalconForCausalLM"),
"StableLMEpochForCausalLM": ("stablelm", "StablelmForCausalLM"),
"YiForCausalLM": ("yi", "YiForCausalLM")
}

# Models not supported by ROCm.
Expand Down
342 changes: 0 additions & 342 deletions vllm/model_executor/models/aquila.py

This file was deleted.

1 change: 0 additions & 1 deletion vllm/transformers_utils/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@
from vllm.transformers_utils.configs import *

_CONFIG_REGISTRY = {
"aquila": AquilaConfig,
"baichuan": BaiChuanConfig,
"chatglm": ChatGLMConfig,
"mpt": MPTConfig,
Expand Down
2 changes: 0 additions & 2 deletions vllm/transformers_utils/configs/__init__.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@
from vllm.transformers_utils.configs.aquila import AquilaConfig
from vllm.transformers_utils.configs.baichuan import BaiChuanConfig
from vllm.transformers_utils.configs.chatglm import ChatGLMConfig
from vllm.transformers_utils.configs.mpt import MPTConfig
Expand All @@ -9,7 +8,6 @@
from vllm.transformers_utils.configs.falcon import RWConfig

__all__ = [
"AquilaConfig",
"BaiChuanConfig",
"ChatGLMConfig",
"MPTConfig",
Expand Down
Loading

0 comments on commit bbda543

Please sign in to comment.