Skip to content

Commit

Permalink
remove vllm patch
Browse files Browse the repository at this point in the history
  • Loading branch information
hiyouga committed Jan 23, 2024
1 parent cced589 commit 95a30f3
Show file tree
Hide file tree
Showing 4 changed files with 2 additions and 76 deletions.
4 changes: 2 additions & 2 deletions requirements.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
numpy
sse-starlette
transformers
vllm==0.2.6
transformers>=4.34.0
vllm>=0.2.6
2 changes: 0 additions & 2 deletions src/imitater/config/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -14,5 +14,3 @@ class Config:
embed_model_path: str
embed_model_device: List[int]
embed_batch_size: int

enable_attn_bias: bool
4 changes: 0 additions & 4 deletions src/imitater/model/chat_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,6 @@
from vllm import AsyncEngineArgs, AsyncLLMEngine, SamplingParams

from ..agent import get_agent
from ..utils.vllm_monkey_patch import llama_attn_bias_monkey_patch


if TYPE_CHECKING:
Expand All @@ -22,9 +21,6 @@ def __init__(self, config: "Config") -> None:
self._load_generation_config()

def _init_vllm_engine(self) -> None:
if self._config.enable_attn_bias:
llama_attn_bias_monkey_patch()

engine_args = AsyncEngineArgs(model=self._config.chat_model_path, trust_remote_code=True)
engine_args.tensor_parallel_size = len(self._config.chat_model_device)
self._engine = AsyncLLMEngine.from_engine_args(engine_args)
Expand Down
68 changes: 0 additions & 68 deletions src/imitater/utils/vllm_monkey_patch.py

This file was deleted.

0 comments on commit 95a30f3

Please sign in to comment.