Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Seems that all models are loaded to the first GPU? #11

Open
DDDOH opened this issue Nov 25, 2024 · 2 comments
Open

Seems that all models are loaded to the first GPU? #11

DDDOH opened this issue Nov 25, 2024 · 2 comments

Comments

@DDDOH
Copy link

DDDOH commented Nov 25, 2024

I use four A100 run python -m prover.launch --config=configs/RMaxTS.py --log_dir=logs/RMaxTS_results
But always get cuda out of memory error.

I added some print message:
image

Then i get
image

All models are tried to load in first GPU. Is this expected? Or maybe pytorch behavior is different for new version (I am using '2.5.1+cu121', and the version in requirement.txt is 2.2.1)

@Luobots
Copy link

Luobots commented Dec 15, 2024

Same error. VLLM can't set the cuda index.

@zzhisthebest
Copy link

@Luobots How to solve this problem?I meet the same error.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants