-
Notifications
You must be signed in to change notification settings - Fork 62
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ROCm build broken #605
ROCm build broken #605
Conversation
Reviewer's Guide by SourceryThe build flags for llama.cpp were updated to enable ROCm acceleration. Additionally, the script was updated to remove more ROCm libraries. Flow diagram for updated ROCm build configurationgraph TD
A[Start Build] --> B{Container Type?}
B -->|ROCm| C[Enable ROCm/HIP]
B -->|CUDA| D[Enable CUDA]
B -->|CPU| E[CPU Only Build]
C --> F[Set -DGGML_HIP=ON]
D --> G[Set -DGGML_CUDA=ON]
F --> H[Clean Up ROCm Libraries]
G --> I[Build Complete]
E --> I
H --> I
style C fill:#f9f,stroke:#333
style F fill:#f9f,stroke:#333
File-Level Changes
Tips and commandsInteracting with Sourcery
Customizing Your ExperienceAccess your dashboard to:
Getting Help
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hey @ericcurtin - I've reviewed your changes and they look great!
Here's what I looked at during the review
- 🟢 General issues: all looks good
- 🟢 Security: all looks good
- 🟢 Testing: all looks good
- 🟢 Complexity: all looks good
- 🟢 Documentation: all looks good
Help me be more useful! Please click 👍 or 👎 on each comment and I'll use the feedback to improve your reviews.
@maxamillion @bmahabirbu upstream llama.cpp changed their build flags for ROCm, as a result we've been building without ROCm here for a little while as the code would still build, just without ROCm. |
This was happening basically:
|
@rhatdan we probably need new container images, once we get the green PRs in here. ROCm is broken |
Really good catch! I wonder if there is a way to catch these sort of errors? I can look into editing the make file to look for mismatched arguments |
If you find out great, it would be appreciated, I tried briefly to make cmake treat this as an error, but didn't figure it out. |
69af1f4
to
027b813
Compare
I thought of something to check for cmake warnings @bmahabirbu , I'd prefer something that was built-into cmake, as it would be more elegant than this, but this should work. |
027b813
to
d830b51
Compare
This build may actually fail, it might find other hidden warnings around the place |
d830b51
to
a29bbd9
Compare
The build flags changed in upstream llama.cpp so we were no longer building with ROCm acceleration. Signed-off-by: Eric Curtin <[email protected]>
a29bbd9
to
15a3648
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
This is great, thank you for catching this! This might explain why I wasn't seeing any GPU utilization on my Ryzen APU when watching nvtop
... I had tossed it up to nvtop
oddities.
Could have been! If you want to test this before it gets pushed, there is a flag to manually pass container images to RamaLama (you could build this locally). |
The build flags changed in upstream llama.cpp so we were no longer building with ROCm acceleration.
Summary by Sourcery
Fix ROCm builds.
Build:
-DGGML_HIP=ON
build flag instead of-DGGML_HIPBLAS=1
.