view post Post 4097 I am very sad to say that the budget in creating of SnowflakeCore-G1 1b and 7b MoE models ran out and I can't pre-train them anymore. See translation 7 replies · 😔 8 8 👀 3 3 + Reply
view post Post 543 the training for SnowflakeCore-G1-1B and 7B would be retaken because now I implemented DeepSpeed and management to use two gpus. See translation 👍 1 1 + Reply
view post Post 273 The development of SnowflakeCore-G1-7B-MoE it getting delay. In the mean time I am working on SnowflakeCore-G1-1B-MoE witch would be a pre-train chatbot. See translation 1 reply · 👀 3 3 + Reply