Skip to content

Commit

Permalink
Use contiguous pa by default (#519)
Browse files Browse the repository at this point in the history
  • Loading branch information
madamczykhabana authored Nov 18, 2024
1 parent dac5d80 commit a4e689a
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion vllm/worker/hpu_model_runner.py
Original file line number Diff line number Diff line change
Expand Up @@ -642,7 +642,7 @@ def __init__(
self._setup_buckets()
self._set_gc_threshold()
self.use_contiguous_pa = os.environ.get('VLLM_CONTIGUOUS_PA',
'false').lower() == 'true'
'true').lower() == 'true'
# For multi-step scheduling
self.cached_step_outputs: List[torch.Tensor] = []

Expand Down

0 comments on commit a4e689a

Please sign in to comment.