Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

!CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python==0.2.45 --force-reinstall --no-cache-dir -q #4998

Open
123bksri opened this issue Dec 8, 2024 · 1 comment

Comments

@123bksri
Copy link

123bksri commented Dec 8, 2024

Describe the current behavior
A clear and concise explanation of what is currently happening.

Describe the expected behavior
A clear and concise explanation of what you expected to happen.

What web browser you are using
(Chrome, Firefox, Safari, etc.)

Additional context
Link to a minimal, public, self-contained notebook that reproduces this issue.

  • Share the file using your GitHub account using File > Save a copy as a GitHub Gist.
  • or Share Drive notebooks using the Share button then 'Get Shareable Link'.
@123bksri 123bksri added the bug label Dec 8, 2024
@colaboratory-team
Copy link
Contributor

Can you provide a repro notebook that shows the error you're trying to resolve?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants