Jay Shin
jshin49
AI & ML interests
None yet
Organizations
pre-training
-
Pre-training Small Base LMs with Fewer Tokens
Paper • 2404.08634 • Published • 36 -
Ziya2: Data-centric Learning is All LLMs Need
Paper • 2311.03301 • Published • 20 -
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 43 -
MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Paper • 2404.06395 • Published • 24
multi-lingual llms
pre-training
-
Pre-training Small Base LMs with Fewer Tokens
Paper • 2404.08634 • Published • 36 -
Ziya2: Data-centric Learning is All LLMs Need
Paper • 2311.03301 • Published • 20 -
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 43 -
MiniCPM: Unveiling the Potential of Small Language Models with Scalable Training Strategies
Paper • 2404.06395 • Published • 24
models 0
None public yet
datasets 0
None public yet