You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Pty72
changed the title
[FEATURE]: "Can gradient accumulation be used in the pretraining of llama2-70b?"
[FEATURE]: Can gradient accumulation be used in the pretraining of llama2-70b?
Sep 23, 2023
Hello, we have just supported gradient accumulation on Gemini plugin.
Usage can be referred to docs/source/en/features/gradient_accumulation_with_booster.md, our online tutorials will also be updated within a few days.
Describe the feature
Can gradient accumulation be used in the pretraining of llama2-70b?
And if so, how can it be enabled?
The text was updated successfully, but these errors were encountered: