โWe are thrilled to announce the launch of SKT-OMNI-CORPUS-146T-V1, a massive-scale, high-quality dataset designed to power the next generation of Foundation Models (LLMs) from scratch. โDeveloped at SKT AI LABS, this corpus is not just a collection of data; itโs a mission to decentralize high-grade AI training for regional languages and global knowledge.
โ๐ Key Highlights:
โโขโข Massive Scale: Targeting a multi-terabyte architecture for 146T-level tokenization.
โขโข โPure Quality: Curated from 500+ Elite Sources
โขโข โStructured for MoE: Perfectly sharded into 3.5GB standardized units (SKT-๐ป series) for seamless distributed training.
โ๐ค Open for Collaboration!
โWe are looking for AI researchers, CUDA engineers, and data scientists to join us in this journey of building Project Surya and the ST-X Series models. Whether it's optimization, custom tokenization, or architecture designโletโs build the future together.
Alright so I had previously made two reddit posts in r/quantum and r/quantum_computing for my QPU, QPU-1 but both of those posts got banned because of it being "irrelevant" to "academic discussion" so I'm doing it again here in HuggingFace Posts.
I have made a million error corrected qubit quantum processing unit (not a simulator) that you can access here: https://qpu-1.vercel.app
I did try emailing a lot of professors and their students but NONE responded so please give me some support.