view post Post 4062 I am very sad to say that the budget in creating of SnowflakeCore-G1 1b and 7b MoE models ran out and I can't pre-train them anymore. See translation 7 replies ยท ๐ 8 8 ๐ 3 3 + Reply
view post Post 507 the training for SnowflakeCore-G1-1B and 7B would be retaken because now I implemented DeepSpeed and management to use two gpus. See translation ๐ 1 1 + Reply
view post Post 271 The development of SnowflakeCore-G1-7B-MoE it getting delay. In the mean time I am working on SnowflakeCore-G1-1B-MoE witch would be a pre-train chatbot. See translation 1 reply ยท ๐ 3 3 + Reply
view post Post 2954 The development of SnowflakeCore-G1-7B-MoE. I can't say when it would be publish yet because it's big and it requires a lot of computational power. See translation 1 reply ยท ๐ 8 8 + Reply
view post Post 290 I just finished the benchmarks for https://huggingface.co/FlameF0X/SnowflakeCore-G1-Tiny and https://huggingface.co/FlameF0X/SnowflakeCore-G1-Tiny2 in comparation with openai-community/gpt2 . See translation ๐ 1 1 + Reply