Skip to content
This repository has been archived by the owner on Oct 16, 2023. It is now read-only.

66B model load checkpoint #137

Merged
merged 3 commits into from
Aug 30, 2022
Merged

66B model load checkpoint #137

merged 3 commits into from
Aug 30, 2022

Conversation

dujiangsu
Copy link
Contributor

No description provided.

tokenizer_path = "facebook/opt-350m"
# server_host = "127.0.0.1"
# tokenizer_path = "facebook/opt-125m"
tokenizer_path = "facebook/opt-30b"
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

why use the 30b tokenizer? The 66b tokenizer has bug?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Here the config is for 30B model by default.
I add the 66B tokenizer as the candidate.

@dujiangsu dujiangsu merged commit 0000835 into hpcaitech:main Aug 30, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants