A basic training example using GGML #707
bssrdf
started this conversation in
Show and tell
Replies: 1 comment 1 reply
-
Nice job! Thank you for sharing If you have some feedback of what could be improved - please let us know. The training capabilities in |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I just want to share what I have been working on recently. This is an example of training a MNIST VAE. The goal is to use only
ggml
pipeline and its implementation of ADAM optimizer.There aren't many training examples using
ggml
.The only one I found is baby-llama. But I think its way of doing opmization is not quite right.Found another training example inllama.cpp
which shows a proper way of using Adam.Some of the mods I have to add
Below are some samples from the VAE trained on MNIST after each epoch (total 10 epochs).
| 
| 
| 
| 
| 
Beta Was this translation helpful? Give feedback.
All reactions