8000 chore: bump llama_cpp_python to 0.3.6 by Ian321 · Pull Request #2368 · instructlab/instructlab · GitHub
[go: up one dir, main page]

Skip to content

chore: bump llama_cpp_python to 0.3.6 #2368

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jan 29, 2025
Merged

Conversation

Ian321
Copy link
Contributor
@Ian321 Ian321 commented Oct 2, 2024

Instructlab had been using a outdated version of llama_cpp_python that did not support models such as Mixtral NeMo. This PR simply bumps the version of that dependency to the latest one and updates the pipelines and documentation to use the new building flags.

Checklist:

  • Commit Message Formatting: Commit titles and messages follow guidelines in the
    conventional commits.
  • Changelog updated with breaking and/or notable changes for the next minor release.
  • Documentation has been updated, if necessary.
  • Unit tests have been added, if necessary.
  • Functional tests have been added, if necessary.
  • E2E Workflow tests have been added, if necessary.

@mergify mergify bot added ci-failure PR has at least one CI failure dependencies Relates to dependencies labels Oct 2, 2024
@mergify mergify bot added ci-failure PR has at least one CI failure and removed ci-failure PR has at least one CI failure labels Oct 2, 2024
@tiran
Copy link
Contributor
tiran commented Oct 2, 2024

Latest llama-cpp-python has different CMAKE_ARGS for CUDA, ROCm, and MPS support. You have to update the README, other docs, and tests, too. Look for -DLLAMA.

@mergify mergify bot added CI/CD Affects CI/CD configuration container Affects containization aspects documentation Improvements or additions to documentation testing Relates to testing and removed ci-failure PR has at least one CI failure labels Oct 2, 2024
@Ian321 Ian321 changed the title chore: bump llama_cpp_python to 0.3.1 chore!: bump llama_cpp_python to 0.3.1 Oct 2, 2024
@mergify mergify bot added the ci-failure PR has at least one CI failure label Oct 2, 2024
@mergify mergify bot removed the ci-failure PR has at least one CI failure label Oct 2, 2024
@mergify mergify bot added the ci-failure PR has at least one CI failure label Oct 3, 2024
@Ian321
Copy link
Contributor Author
Ian321 commented Oct 3, 2024

The tests with reduced max_ctx_size fail because of this: abetlen/llama-cpp-python#1759

@alimaredia
Copy link
Contributor

@Ian321 could you make changes to the tests that fail due to reduced max_ctx_size as part of this PR?

@Ian321
Copy link
Contributor Author
Ian321 commented Oct 12, 2024

@alimaredia what kind of changes do you have in mind?

This worked as expected with previous versions of llama-cpp-python and now there seems to have been a regression.

I tested llama.cpp directly with a reduced ctx and it did not crash, so now I'm planning on fixing it in llama-cpp-python and then bump this pr (to hopefully 0.3.2).

@nathan-weinberg
Copy link
Member

@alimaredia what kind of changes do you have in mind?

This worked as expected with previous versions of llama-cpp-python and now there seems to have been a regression.

I tested llama.cpp directly with a reduced ctx and it did not crash, so now I'm planning on fixing it in llama-cpp-python and then bump this pr (to hopefully 0.3.2).

We just need CI to be passing here to ensure there are no regressions, since this is a non-trivial bump

@alimaredia
Copy link
Contributor

If you look at the failure here: https://github.com/instructlab/instructlab/actions/runs/11153487993/job/31003244761?pr=2368#step:15:338, some of the tests here (https://github.com/instructlab/instructlab/blob/main/scripts/functional-tests.sh) are failing.

Those tests get run in order for every PR to merge, so we'd expect changes to the tests in this PR in order to do the version bump. Removal of certain functional tests is on the table if properly justified.

@tiran
Copy link
Contributor
tiran commented Oct 16, 2024

Our downstream build pipeline is now configured to handle llama_cpp_python 0.2.75 and 0.3.1.

@Ian321
Copy link
Contributor Author
Ian321 commented Oct 16, 2024

@alimaredia the tests fail because there is a bug in abetlen/llama-cpp-python#1759 for which I have provided a fix abetlen/llama-cpp-python#1796 . We just have to wait for the next release of it (where it's hopefully merged of fixed some other way).

@tiran if you mean the changes to the pipeline directly, I only see some general cleanup and nothing that would affect this PR directly. The test that caught this should not be modified or removed as it's what caught the above mentioned bug and helped me submit a PR for it. The only thing I could recommend is to add a timeout to .github/workflows/test.yml, so that it won't hang for 6 hours.

I will still rebase it, just in case I missed something.

@mergify mergify bot removed the ci-failure PR has at least one CI failure label Oct 16, 2024
@tiran
Copy link
Contributor
tiran commented Oct 16, 2024

My update regarding the downstream pipeline was for @nathan-weinberg and @alimaredia . We have an internal build pipeline that rebuilds all Python wheel from sources. Some packages like llama-cpp-python need extra configuration to build correctly. llama-cpp-python 0.3 has deprecated some options and introduced new build flags. Our internal builds are now able to handle >=0.2.75 and 0.3.x.

@Ian321 Ian321 changed the title chore!: bump llama_cpp_python to 0.3.5 chore!: bump llama_cpp_python to 0.3.6 Jan 10, 2025
@Ian321 Ian321 changed the title chore!: bump llama_cpp_python to 0.3.6 chore: bump llama_cpp_python to 0.3.6 Jan 10, 2025
@nathan-weinberg nathan-weinberg added this to the 0.24.0 milestone Jan 24, 2025
Copy link
Contributor
mergify bot commented Jan 24, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. @Ian321 please rebase it. https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase This Pull Request needs to be rebased label Jan 24, 2025
@nathan-weinberg
Copy link
Member

@Ian321 I've set this for our 0.24.0 milestone - since we were able to bump to 0.3.2 in ilab 0.23.0 we're hoping to follow up shortly after the release with this! cc @fabiendupont

@mergify mergify bot removed the needs-rebase This Pull Request needs to be rebased label Jan 24, 2025
Copy link
Contributor
mergify bot commented Jan 29, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. @Ian321 please rebase it. https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase This Pull Request needs to be rebased label Jan 29, 2025
Signed-off-by: Ignaz Kraft <ignaz.k@live.de>
@mergify mergify bot removed the needs-rebase This Pull Request needs to be rebased label Jan 29, 2025
@nathan-weinberg nathan-weinberg removed the hold In-progress PR. Tag should be removed before merge. label Jan 29, 2025
@nathan-weinberg nathan-weinberg requested review from a team and removed request for tiran, alimaredia and < 8000 a data-hovercard-type="user" data-hovercard-url="/users/cdoern/hovercard" data-octo-click="hovercard-link-click" data-octo-dimensions="link_type:self" href="/cdoern">cdoern January 29, 2025 14:56
@mergify mergify bot added the one-approval PR has one approval from a maintainer label Jan 29, 2025
@nathan-weinberg nathan-weinberg removed this from the 0.24.0 milestone Jan 29, 2025
@mergify mergify bot added the ci-failure PR has at least one CI failure label Jan 29, 2025
@fabiendupont
Copy link
Contributor

Nice solution @Ian321. Much shorter than I thought.

@mergify mergify bot removed the ci-failure PR has at least one CI failure label Jan 29, 2025
@mergify mergify bot removed the one-approval PR has one approval from a maintainer label Jan 29, 2025
@mergify mergify bot merged commit c3ec416 into instructlab:main Jan 29, 2025
29 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CI/CD Affects CI/CD configuration container Affects containization aspects dependencies Relates to dependencies documentation Improvements or additions to documentation testing Relates to testing
Projects
None yet
Development

Successfully merging this pull request may close these issues.

llama_cpp_python 0.3.2 breaks building from source on aarch64 with CUDA
8 participants
0