Skip to content

Commit

Permalink
Removed vllm.hpu directory and changed relevant imports
Browse files Browse the repository at this point in the history
  • Loading branch information
tzielinski-habana committed Sep 17, 2024
1 parent 4c1ca3a commit d9ff46a
Show file tree
Hide file tree
Showing 16 changed files with 12 additions and 603 deletions.
1 change: 0 additions & 1 deletion .github/workflows/mypy.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -50,6 +50,5 @@ jobs:
mypy vllm/transformers_utils --config-file pyproject.toml
mypy vllm/usage --config-file pyproject.toml
mypy vllm/worker --config-file pyproject.toml
mypy vllm/hpu --config-file pyproject.toml
1 change: 0 additions & 1 deletion format.sh
Original file line number Diff line number Diff line change
Expand Up @@ -113,7 +113,6 @@ mypy vllm/spec_decode --config-file pyproject.toml
mypy vllm/transformers_utils --config-file pyproject.toml
mypy vllm/usage --config-file pyproject.toml
mypy vllm/worker --config-file pyproject.toml
mypy vllm/hpu --config-file pyproject.toml


# If git diff returns a file that is in the skip list, the file may be checked anyway:
Expand Down
1 change: 1 addition & 0 deletions requirements-hpu.txt
Original file line number Diff line number Diff line change
Expand Up @@ -6,3 +6,4 @@ ray == 2.32.0
triton
pandas
tabulate
vllm-hpu-extension @ git+https://github.com/HabanaAI/vllm-hpu-extension.git@30ee2d1
2 changes: 1 addition & 1 deletion tests/lora/test_lora_hpu.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import pytest
import torch
from vllm_hpu_extension.ops import LoraMask

from vllm.hpu.ops import LoraMask
from vllm.lora.layers import _apply_lora, _apply_lora_packed_nslice

from .utils import DummyLoRAManager
Expand Down
6 changes: 3 additions & 3 deletions vllm/attention/backends/habana_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -7,14 +7,14 @@
from typing import Any, Dict, List, Optional, Tuple, Type

import torch
import vllm_hpu_extension.ops as ops
from vllm_hpu_extension import cache_ops
from vllm_hpu_extension.utils import Matmul, Softmax, VLLMKVCache

import vllm.hpu.ops as ops
from vllm.attention.backends.abstract import (AttentionBackend, AttentionImpl,
AttentionMetadata, AttentionType)
from vllm.attention.ops.habana_paged_attn import (HabanaPagedAttention,
HabanaPagedAttentionMetadata)
from vllm.hpu import cache_ops
from vllm.hpu.utils import Matmul, Softmax, VLLMKVCache
from vllm.logger import init_logger

logger = init_logger(__name__)
Expand Down
3 changes: 1 addition & 2 deletions vllm/attention/ops/habana_paged_attn.py
Original file line number Diff line number Diff line change
Expand Up @@ -6,8 +6,7 @@
from typing import Dict, List, Optional, Tuple

import torch

from vllm.hpu import cache_ops, ops
from vllm_hpu_extension import cache_ops, ops

# Should be the same as PARTITION_SIZE in `paged_attention_v2_launcher`.
_PARTITION_SIZE = 512
Expand Down
6 changes: 0 additions & 6 deletions vllm/hpu/__init__.py

This file was deleted.

107 changes: 0 additions & 107 deletions vllm/hpu/cache_ops.py

This file was deleted.

Loading

0 comments on commit d9ff46a

Please sign in to comment.