You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
mtairum
changed the title
[Llama3.1-8B] Support max prefill lengths on demos (within L1 capacity)
[Llama3] Debug accuracy drop on sequence lengths > 64k tokens
Dec 2, 2024
Status:
Prefill seqlen up to 128k supported in Llama3 codebase (1b, 3B, 8B, 11B, 70B), but bad accuracy in some corner cases:
Things we've tried so far:
The text was updated successfully, but these errors were encountered: