Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for gguf #8

Closed
phronmophobic opened this issue Sep 24, 2023 · 3 comments
Closed

Add support for gguf #8

phronmophobic opened this issue Sep 24, 2023 · 3 comments

Comments

@phronmophobic
Copy link
Owner

The latest llama.cpp development has deprecated the ggml format in favor of a new gguf format.

llama.cpp has chosen to break their API and make ggml models useless. The goal for llama.clj is to upgrade without breaking backwards compatibility. More research is required, but the initial plan is something like:

  • treat the new raw API for llama.cpp as a separate library
  • create a protocol for any shared functionality and implement it for both the ggml version and the latest version
  • create independent builds that can be included independently or together
  • Add better support for including your own llama.cpp build
@phronmophobic
Copy link
Owner Author

I updated to the latest version of llama.cpp locally and was able to get a gguf model to run without too many changes. However, there are a still a few updates in progress for llama.cpp that I'll probably wait on before making a new release:

@phronmophobic
Copy link
Owner Author

Fixed in v0.8.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant