Question

#2
by Enderchef - opened

I really like your models! I have an idea to improve it a lot. It's just a theory and I don't want to waste your time with it, so:
I have a GPU; How can I run a post-train on this model? How long did your post-training take(was it a long time? How long on your RTX 5090)?
If it turns out well, I'll show you the results πŸ™ƒ

Glad your interested!
Currently im working on a way for anyone to contribute (related to poll on website) but it seems like its going to take a while πŸ˜“
This one took around a day to train fully on a 5090.
Pre training was 18 hours, post training was 3.
Currently I dont have a open source platform for fine-tuning these models but some platforms may support it.
If you want to contribute to CompactAI we can discuss on discord (if your open to that)

I'm making a training pipeline to do my idea right now, I'll tell you if anything interesting comes out of it!
(I'm experimenting with RL and CoT, and looking into Effort like Opus has, I'm running the RL pipeline right now)

Just wait for next Haiku, from testing it knows what you are talking about sometimes.
(Sonnet & Opus are waiting for the contributor hardware pooling app)

Can I join the pool for a hour or two?

Sadly we dont actually have any infra built for it. (only a sick frontend :P)
If you know how to code complex apps we would be happy to let you contribute

Sign up or log in to comment