Kaomoji Fixed Translation as Knowledge Hallucination in LLMs: A Case Study on XiaoHongShu
Research Article
Open Access
CC BY

Kaomoji Fixed Translation as Knowledge Hallucination in LLMs: A Case Study on XiaoHongShu

Jinhan Feng 1*
1 Nanjing Jinling High School, Nanjing, China, 210000
*Corresponding author: fengjinhan1013@gmail.com
Published on 5 November 2025
Volume Cover
ACE Vol.203
ISSN (Print): 2755-273X
ISSN (Online): 2755-2721
ISBN (Print): 978-1-80590-515-8
ISBN (Online): 978-1-80590-516-5
Download Cover

Abstract

Large Language Models (LLMs) produce fluent but sometimes unfounded outputs, a phenomenon commonly called hallucination. On the XiaoHongShu (RED) platform, when users input kaomoji—ASCII or Unicode emoticons—the translation tool often returns a stable, seemingly meaningful Chinese phrase even though the input lacks explicit semantic content. This paper examines why LLMs generate such fixed translations. Building on the concept of hallucination snowballing, classifications of hallucination types, and methods for reducing knowledge hallucinations, through case analysis, literature synthesis, and mechanistic review, this paper mainly discusses: (1) LLMs produce consistent translations for semantically null inputs. (2) Which hallucination category best fits this case? (3) How do prompt framing, pretraining co-occurrence patterns, and autoregressive decoding contribute? This study argues that the kaomoji fixed translation is primarily a form of knowledge hallucination reinforced by prefix consistency and statistical co-occurrence. This paper concludes by recommending uncertainty-aware behaviors, prompt-level checks, and data interventions to reduce such errors.

Keywords:

Large Language Models, knowledge hallucination, kaomoji, fixed translation, consistency bias

View PDF
Feng,J. (2025). Kaomoji Fixed Translation as Knowledge Hallucination in LLMs: A Case Study on XiaoHongShu. Applied and Computational Engineering,203,41-45.

References

[1]. Lü, Q., et al. (2024) Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models. Proceedings of the International Conference on Machine Learning (ICML).

[2]. Maynez, J., et al. (2020) On Faithfulness and Factuality in Abstractive Summarization. Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL).

[3]. Shuster, K., et al. (2021) Retrieval-Enhanced Generative Models. Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL).

[4]. Ji, Z., et al. (2023) Survey of Hallucination in Natural Language Generation. ACM Computing Surveys.

[5]. Zhang, M., et al. (2024) How Language Model Hallucinations Can Snowball. Proceedings of the 41st International Conference on Machine Learning (ICML).

[6]. Wu, M., et al. (2024) Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language Models. Proceedings of the International Conference on Machine Learning (ICML).

[7]. Holtzman, A., et al. (2020) The Curious Case of Neural Text Degeneration. Proceedings of the International Conference on Learning Representations (ICLR).

[8]. Pickering, M., & Garrod, S. (2020) Toward a Mechanistic Psychology of Dialogue. Behavioral and Brain Sciences.

[9]. Lee, N., et al. (2022) Prompt Sensitivity in Large Language Models. arXiv: 2212.10559.

[10]. Barbieri, F., et al. (2018) Modelling the Semantics of Emoji. Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).

[11]. Vaswani, A., et al. (2017) Attention is All You Need. Proceedings of the Conference on Neural Information Processing Systems (NeurIPS).

[12]. Evans, O., et al. (2021) TruthfulQA: Measuring How Models Mimic Human Falsehoods. Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL).

[13]. Pagnoni, A., et al. (2021) Understanding Factuality in Abstractive Summarization. Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics (NAACL).

[14]. Lin, S., et al. (2022) Teaching Models to Refuse Unknowns. Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP).

[15]. Zhao, Z., et al. (2023) Revisiting Chain-of-Thought Reasoning. Proceedings of the Conference on Neural Information Processing Systems (NeurIPS).

[16]. Kim, B., et al. (2023) Reducing Hallucination via Data Attribution. Proceedings of the Annual Meeting of the Association for Computational Linguistics (ACL).

Cite this article

Feng,J. (2025). Kaomoji Fixed Translation as Knowledge Hallucination in LLMs: A Case Study on XiaoHongShu. Applied and Computational Engineering,203,41-45.

Data availability

The datasets used and/or analyzed during the current study will be available from the authors upon reasonable request.

About volume

Volume title: Proceedings of CONF-SPML 2026 Symposium: The 2nd Neural Computing and Applications Workshop 2025

ISBN: 978-1-80590-515-8(Print) / 978-1-80590-516-5(Online)
Editor: Marwan Omar, Guozheng Rao
Conference date: 21 December 2025
Series: Applied and Computational Engineering
Volume number: Vol.203
ISSN: 2755-2721(Print) / 2755-273X(Online)