ParScale/ParScale-1.8B-P8
Text Generation
•
2B
•
Updated
•
22
•
6
Base models trained on 1T high-quality tokens, demonstrating strong competitiveness among existing SOTA small models (<2B).