ai
  • Crypto News
  • Ai
  • eSports
  • Bitcoin
  • Ethereum
  • Blockchain
Home»Ai»Meet South Korea’s LLM Powerhouses: HyperClova, AX, Solar Pro, and More
Ai

Meet South Korea’s LLM Powerhouses: HyperClova, AX, Solar Pro, and More

Share
Facebook Twitter LinkedIn Pinterest Email

South Korea is rapidly establishing itself as a key innovator in large language models (LLMs), driven by strategic government investments, corporate research, and open-source collaborations to create models tailored for Korean language processing and domestic applications. This focus helps mitigate dependencies on foreign AI technologies, enhances data privacy, and supports sectors like healthcare, education, and telecommunications.

Government-Backed Push for Sovereign AI

In 2025, the Ministry of Science and ICT initiated a 240 billion won program, selecting five consortia—led by Naver Cloud, SK Telecom, Upstage, LG AI Research, and NC AI—to develop sovereign LLMs capable of operating on local infrastructure.

Regulatory advancements include the Ministry of Food and Drug Safety’s guidelines for approving text-generating medical AI, marking the first such framework globally in early 2025.

Corporate and Academic Innovations

SK Telecom introduced AX 3.1 Lite, a 7 billion-parameter model trained from scratch on 1.65 trillion multilingual tokens with a strong Korean emphasis. It achieves approximately 96% performance on KMMLU2 for Korean language reasoning and 102% on CLIcK3 for cultural understanding relative to larger models, and is available open-source on Hugging Face for mobile and on-device application.

Naver advanced its HyperClova series with HyperClova X Think in June 2025, enhancing Korean-specific search and conversational capabilities.

Upstage’s Solar Pro 2 stands as the sole Korean entry on the Frontier LM Intelligence leaderboard, demonstrating efficiency in matching performance of much larger international models.

LG AI Research launched Exaone 4.0 in July 2025, which performs competitively in global benchmarks with a 30 billion-parameter design.

Seoul National University Hospital developed Korea’s first medical LLM, trained on 38 million de-identified clinical records, scoring 86.2% on the Korean Medical Licensing Examination compared to the human average of 79.7%.

Mathpresso and Upstage collaborated on MATH GPT, a 13 billion-parameter small LLM that surpasses GPT-4 in mathematical benchmarks with 0.488 accuracy versus 0.425, using significantly less computational resources.

Open-source initiatives like Polyglot-Ko (ranging from 1.3 to 12.8 billion parameters) and Gecko-7B address gaps by continually pretraining on Korean datasets to handle linguistic nuances such as code-switching.

Technical Trends

Korean developers emphasize efficiency, optimizing token-to-parameter ratios inspired by Chinchilla scaling to enable 7 to 30 billion-parameter models to compete with larger Western counterparts despite constrained resources.

Domain-specific adaptations yield superior results in targeted areas, as seen in the medical LLM from Seoul National University Hospital and MATH GPT for mathematics.

Progress is measured through benchmarks including KMMLU2, CLIcK3 for cultural relevance, and the Frontier LM leaderboard, confirming parity with advanced global systems.

Market Outlook

The South Korean LLM market is forecasted to expand from 182.4 million USD in 2024 to 1,278.3 million USD by 2030, reflecting a 39.4% compound annual growth rate, primarily fueled by chatbots, virtual assistants, and sentiment analysis tools. Integration of edge-computing LLMs by telecom firms supports reduced latency and enhanced data security under initiatives like the AI Infrastructure Superhighway.

South Korean Large Language Models Mentioned

# Model Developer / Lead Institution Parameter Count Notable Focus
1 AX 3.1 Lite SK Telecom 7 billion Mobile and on-device Korean processing
2 AX 4.0 Lite SK Telecom 72 billion Scalable sovereign applications
3 HyperClova X Think Naver ~204 billion (est.) Korean search and dialogue
4 Solar Pro 2 Upstage ~30 billion (est.) General efficiency on global leaderboards
5 MATH GPT Mathpresso + Upstage 13 billion Mathematics specialization
6 Exaone 4.0 LG AI Research 30 billion Multimodal AI capabilities
7 Polyglot-Ko EleutherAI + KIFAI 1.3 to 12.8 billion Korean-only open-source training
8 Gecko-7B Beomi community 7 billion Continual pretraining for Korean
9 SNUH Medical LLM Seoul National University Hospital undisclosed (~15B est.) Clinical and medical decision support

These developments highlight South Korea’s approach to creating efficient, culturally relevant AI models that strengthen its position in the global technology landscape.


Sources:

  1. https://www.cnbc.com/2025/08/08/south-korea-to-launch-national-ai-model-in-race-with-us-and-china.html
  2. https://www.forbes.com/sites/ronschmelzer/2025/07/16/sk-telecom-releases-a-korean-sovereign-llm-built-from-scratch/
  3. https://www.kjronline.org/pdf/10.3348/kjr.2025.0257
  4. https://www.rcrwireless.com/20250714/ai/sk-telecom-ai-3
  5. https://huggingface.co/skt/A.X-3.1-Light
  6. https://www.koreaherald.com/article/10554340
  7. http://www.mobihealthnews.com/news/asia/seoul-national-university-hospital-builds-korean-medical-llm
  8. https://www.chosun.com/english/industry-en/2024/05/03/67DRPIFMXND4NEYXNFJYA7QZRA/
  9. https://huggingface.co/blog/amphora/navigating-ko-llm-research-1
  10. https://www.grandviewresearch.com/horizon/outlook/large-language-model-market/south-korea


Michal Sutter is a data science professional with a Master of Science in Data Science from the University of Padova. With a solid foundation in statistical analysis, machine learning, and data engineering, Michal excels at transforming complex datasets into actionable insights.

Share. Facebook Twitter Pinterest LinkedIn Tumblr Email

Related Posts

What is DeepSeek-V3.1 and Why is Everyone Talking About It?

août 21, 2025

On the ground in Ukraine’s largest Starlink repair shop

août 21, 2025

Liquid AI Releases LFM2-VL: Super-Fast, Open-Weight Vision-Language Models Designed for Low-Latency and Device-Aware Deployment

août 21, 2025

Material Cultures looks to the past to build the future

août 21, 2025
Add A Comment

Comments are closed.

Top Posts

SwissCryptoDaily.ch delivers the latest cryptocurrency news, market insights, and expert analysis. Stay informed with daily updates from the world of blockchain and digital assets.

We're social. Connect with us:

Facebook X (Twitter) Instagram Pinterest YouTube
Top Insights

Bitcoin Price Watch: Is $112K the Last Line of Defense?

août 21, 2025

Marvel Rivals has officially been given permission to do the opposite of what players want

août 21, 2025

Economist Who Predicted Bitcoin Would Go To $100 Before $100,000 Returns

août 21, 2025
Get Informed

Subscribe to Updates

Get the latest creative news from FooBar about art, design and business.

Facebook X (Twitter) Instagram Pinterest
  • About us
  • Get In Touch
  • Cookies Policy
  • Privacy-Policy
  • Terms and Conditions
© 2025 Swisscryptodaily.ch.

Type above and press Enter to search. Press Esc to cancel.