-
Notifications
You must be signed in to change notification settings - Fork 148
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FEA] Support feeding pre-trained embeddings to TF4Rec model with high-level api #475
Comments
Is this related to or part of NVIDIA-Merlin/Merlin#211? |
@karlhigley more related to NVIDIA-Merlin/Merlin#471. Not sure about the link to |
When the embedding table are not huge and fit GPU memory, the new |
Is there an example notebook of usage of PretrainedEmbeddingsInitializer to initialize the embedding matrix |
We dont have an example for this feature, but you can refer to the unit test, and try to implement it. |
Thanks, I guess what I am looking for is how to use this along with the input block defined by a model schema, TabularSequenceFeatures (with a series of categorical and continuous features) and tr.NextItemPredictionTask and Electra config. Here's my pseudo code without using the embeddings
|
following up on this ^ |
Any progress? @karunaahuja |
🚀 Feature request
Currently we do not have out of the box support for adding pre-trained embeddings to embedding layer, and ability to freeze them, and train a TF4Rec model. We have embedding_initializer but we never tested if it works accurately and as expected. May be we can create in PyTorch a class like TensorInitializer (TF) as we did in Merlin Models and expose the
embedding initializer
andtrainable
args to the user.We need to
Motivation
This is a FEA coming from our customers and users.
The text was updated successfully, but these errors were encountered: