-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Update the base image of inference-cpp image to oneapi 2025.0.2 #12802
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
please include legacy1 packages to keep support for Intel pre-Gen12 GPU devices
https://github.com/intel/compute-runtime/blob/master/LEGACY_PLATFORMS.md
related issue: intel/compute-runtime#797
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
I built the image https://github.com/intel-analytics/ipex-llm-workflow/actions/runs/13276223098/job/37066211184 and started a container on ws-001 to verify llama.cpp and Ollama. Both are working fine. |
Description
This PR upgrades oneAPI in
inference-cpp
to 2025.0. Additionally, the Dockerfile has been refactored for better readability and organization.Following @JablonskiMateusz ’s suggestion, this PR also installs both the regular and legacy1 compute runtime packages to ensure compatibility with devices before and after Gen12.
1. Why the change?
2. User API changes
3. Summary of the change
4. How to test?
1234
). And paste your action link here once it has been successfully finished.5. New dependencies
- Dependency1
- Dependency2
- ...
- Dependency1 and license1
- Dependency2 and license2
- ...