Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for Gemma 2? #115

Open
sdmorrey opened this issue Aug 4, 2024 · 2 comments
Open

Support for Gemma 2? #115

sdmorrey opened this issue Aug 4, 2024 · 2 comments

Comments

@sdmorrey
Copy link

sdmorrey commented Aug 4, 2024

What would be required to support Gemma 2?
I'd be happy to chip in and help with the code, I just need to have a bit of insight into what would need to be changed?

@b4rtaz
Copy link
Owner

b4rtaz commented Aug 5, 2024

Hello @sdmorrey,

you should check llama2-tasks.cpp and grok1-tasks.cpp files. For different architectures DL builds a different task list. Tasks are reused of course (in grok1-tasks.cpp you can see the implementation of different tasks than Llama model uses).

I see Gemma 2 has more norm layers. Rope layer it seems it's already implemented (FalconRopeCommand). Probably the tokenizer is something that may need more work (converter), but I'm not sure.

@unclemusclez
Copy link

+1 for Gemma 2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants