Article

한국어 언어모델의 정치편향성 측정

송종빈1, 송상헌1,
Jongbeen Song1, Sanghoun Song1,
Author Information & Copyright
1고려대학교
1Korea University
Corresponding author: 부교수 언어학과 고려대학교 02841 서울시 성북구 안암로 145 E-mail: sanghoun@korea.ac.kr

ⓒ Copyright 2026 Language Education Institute, Seoul National University. This is an Open-Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/4.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Received: Mar 01, 2026 ; Revised: Mar 31, 2026 ; Accepted: Apr 13, 2026

Published Online: Apr 30, 2026

ABSTRACT

This study audits the political orientations of seven instruction-tuned Korean large language models (LLMs) amid expanding sovereign-AI deployment. Diverging from Western-centric benchmarks, we evaluate these models using three localized instruments: The Community test, the Hankr Political Compass, and the JoongAng Ilbo’s 2025 Political Orientation Test. Results reveal substantial cross-model dispersion, with no model remaining entirely neutral. While economic orientations generally lean moderately left, social and cultural positions vary widely. Notably, this variation correlates more with developer type and release period than parameter size, suggesting that institutional contexts, training data, and alignment practices leave distinct political fingerprints. Ultimately, this reproducible, Korea-specific audit framework establishes a baseline for evaluating LLM political bias and informs context-sensitive alignment strategies for sovereign AI development.

Keywords: Korean large language models; political bias; value alignment; computational sociolinguistics

References

1.

Bang, Y., Chen, D., Lee, N., & Fung, P. (2024). Measuring political bias in large language models: What is said and how it is said. In Proceedings of the 62nd annual meeting of the association for computational linguistics (Volume 1: Long Papers) (pp. 11142-11159). Association for Computational Linguistics .

2.

Cheng, M., Durmus, E., & Jurafsky, D. (2023). Marked personas: Using natural language prompts to measure stereotypes in language models. arXiv .

3.

Coeckelbergh, M. (2020). AI ethics. MIT Press .

4.

Eom, G., & Kim, D. (2021). Development and application of a comment classifier for online political opinion analysis: Public opinion analysis using KoBERT. Korean Party Studies Review, 20(3), 167-191 .

5.

Feng, S., Park, C. Y., Liu, Y., & Tsvetkov, Y. (2023). From pretraining data to language models to downstream tasks: Tracking the trails of political biases leading to unfair NLP models. In Proceedings of the 61st annual meeting of the association for computational linguistics (Volume 1: Long Papers) (pp. 11737-11762). Association for Computational Linguistics .

6.

Fisher, J., Feng, S., Aron, R., Richardson, T., Choi, Y., Fisher, D. W., Pan, J., Tsvetkov, Y., & Reinecke, K. (2025). Biased LLMs can influence political decision-making. In Proceedings of the 63rd annual meeting of the association for computational linguistics (Volume 1: Long Papers) (pp. 6559-6607). Association for Computational Linguistics .

7.

Jo, S. M. (2025, March 30). 1 in 3 internet users used generative AI last year… Usage rate doubled. Yonhap News. https://www.yna.co.kr/ .

8.

Kim, B., Lee, E., & Na, D. (2023). A new Korean text classification benchmark for recognizing the political intents in online newspapers. arXiv .

9.

Kim, J., & Kim, H. (2025). Responses of artificial intelligence to unethical directive speech acts: Focusing on indirect directive strategies. The Sociolinguistic Journal of Korea, 33(4), 43-80 .

10.

Kim, J., Kim, G., Aiyanyo, I. D., & Lim, H. (2022). Measurement of political polarization in Korean language model by quantitative indicator. In Proceedings of the annual conference on human and language technology (pp. 16-21) .

11.

Manvi, R., Khanna, S., Burke, M., Lobell, D., & Ermon, S. (2024). Large language models are geographically biased. arXiv preprint arXiv:2402.02680 .

12.

Ministry of Science and ICT. (2026, January 15). Announcement of the first phase evaluation results of the proprietary AI foundation model project [Press release]. https://www.msit.go.kr/ .

13.

Motoki, F., Pinho Neto, V., & Rodrigues, V. (2024). More human than human: Measuring ChatGPT political bias. Public Choice, 198(1-2), 3-23. https://doi.org/10.1007/s11127023-01097-2 .

14.

Noh, K. S., Song, S. H., & Oh, E. J. (n.d.). How language models understand honorific mismatches in Korean. Language Research, 60(3), 303-322 .

15.

Oh, D. H. (2026, January 24). Following ChatGPT, Korea is a 'big spender' on Gemini 3... 2nd in global paid subscribers after the U.S. Newsis. https://www.newsis.com/ .

16.

Rettenberger, L., Reischl, M., & Schutera, M. (2025). Assessing political bias in large language models. Journal of Computational Social Science, 8, 42 .

17.

Rozado, D. (2024). The political preferences of LLMs. PLOS ONE, 19(7), e0306621 .

18.

Seo, J., Cho, S., & Park, J. (2025). Political bias in large language models and its implications on downstream tasks. Journal of KIISE, 52(1), 18-28 .

19.

Shin, D. K. (2023). A case study on English test item development training for secondary school teachers using AI tools: Focusing on ChatGPT. Language Research, 59(1), 21-42 .

20.

Taubenfeld, A., Dover, Y., Reichart, R., & Goldstein, A. (2024). Systematic biases in LLM simulations of debates. arXiv preprint arXiv:2402.04049 .