Skip to content
This repository has been archived by the owner on Aug 10, 2024. It is now read-only.

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
sasha0552 authored Jul 6, 2024
1 parent a9204f9 commit eae5e30
Showing 1 changed file with 2 additions and 0 deletions.
2 changes: 2 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,8 @@ CI scripts designed to build a Pascal-compatible version of vLLM and Triton.

*Note: the `vllm` command is an alias for the `python3 -m vllm.entrypoints.openai.api_server` command.*

*Note: kernels for all GPUs except Pascal have been excluded to reduce build time and wheel size. You can still use the new GPUs using tensor parallelism with Ray (and using two instances of `vLLM`, one of which will use upstream `vLLM`). Complain in [issues](https://github.com/sasha0552/vllm-ci/issues) if it disrupts your workflow.*

To install the patched `vLLM` (the patched `triton` will be installed automatically):
```sh
# Create virtual environment
Expand Down

0 comments on commit eae5e30

Please sign in to comment.