Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Backward compatibility with saved llama 2 compiled artifacts #78

Open
dacorvo opened this issue Jan 18, 2024 · 1 comment
Open

Backward compatibility with saved llama 2 compiled artifacts #78

dacorvo opened this issue Jan 18, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@dacorvo
Copy link

dacorvo commented Jan 18, 2024

After upgrading to transformers-neuronx == 0.9.474, I am not able to reload the compiled artifacts for my llama model I saved using transformers-neuronx == 0.8.268.

FileNotFoundError: Could not find a matching NEFF for your HLO in this directory. Ensure that the model you are trying to load is the same type and has the same parameters as the one you saved or call "save" on this model to reserialize it.

The model is identical and so are my parameters. Could you confirm that this is expected (I assume the LLamaForSampling hlo has been modified between the two versions) ?

Will future releases be backward compatible ? This is important to know when defining a strategy for saving neuron models.

@awsilya
Copy link

awsilya commented Jan 19, 2024

@dacorvo - unfortunately we do not support this. The model needs to be recompiled. We are aware of this limitation but at the moment I cannot give you an ETA for the fix.

@awsilya awsilya added the enhancement New feature or request label Jan 22, 2024
@awsilya awsilya closed this as completed Jan 22, 2024
@awsilya awsilya reopened this Jan 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants