Trinity Mini: 26B parameter MoE (3B active), fully post-trained reasoning model
They did pretraining on their own and are still training the large version on 2048 B300 GPUs
Trinity Mini: 26B parameter MoE (3B active), fully post-trained reasoning model
They did pretraining on their own and are still training the large version on 2048 B300 GPUs