A basic training example using GGML #707
bssrdf
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
Nice job! Thank you for sharing If you have some feedback of what could be improved - please let us know. The training capabilities in |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, I just want to share what I have been working on recently. This is an example of training a MNIST VAE. The goal is to use only
ggml
pipeline and its implementation of ADAM optimizer.There aren't many training examples using
ggml
.The only one I found is baby-llama. But I think its way of doing opmization is not quite right.Found another training example inllama.cpp
which shows a proper way of using Adam.Some of the mods I have to add
Below are some samples from the VAE trained on MNIST after each epoch (total 10 epochs).
|
|
|
|
|
Beta Was this translation helpful? Give feedback.
All reactions