I will continue to search, even if the endless stars make my search hopeless, even if I am alone.
Beijing Haidian, China
👋 Welcome! My name is Zexin Jian (简泽鑫), an undergraduate student at Computer College, National University of Defense Technology (国防科技大学计算机学院), , advised by Dr.Zengkun. I will be a PhD student at Institute of Computing Technology (ICT), Chinese Academy of Sciences (CAS) in Sep 2025. I am lucky to be advised by Dr. Huimin Cui and Dr. Chenxi Wang.
🤔 My research interest includes:
Managed Runtime System: build hard-core systems, managed runtime and big data systems for emerging hardware, such as non-volatile memory and resource-disaggregated datacenter
Machine Learning System: Heterogeneous Computing, AI Compiler
LLM: CTG, Story Generation
📫 If you are seeking any form of academic cooperation, please feel free to email me at zexinjian@gmail.com.
"Attribute-based generation methods are of growing significance in controlling the generation of large pre-trained language models (PLMs). Existing studies control the generation by (1) finetuning the model with attributes or (2) guiding the inference processing toward control signals while freezing the PLM. However, finetuning approaches infuse domain bias into generation, making it hard to generate out-of-domain texts. Besides, many methods guide the inference in its word-by-word generation, pushing the word probability to the target attributes, resulting in less fluent sentences. We argue that distilling controlling information from natural texts can produce fluent sentences while maintaining high controllability. In this paper, we propose GRAdient-guided Controllable rEtrieval (GRACE), a retrieval-augmented generation framework to facilitate the generation of fluent sentences with high attribute relevance. GRACE memorizes the semantic and attribute information from unlabeled corpora and applies a controllable retrieval to obtain desired information. For the generation, we design techniques to eliminate the domain bias from the retrieval results and integrate it into the generation model. Additionally, we propose a gradient-guided generation scheme that iteratively steers generation toward higher attribute relevance. Experimental results and quantities of examples verify the effectiveness of our method."
@inproceedings{GRACE-ACL23,title={GRACE: gradient-guided controllable retrieval for augmenting attribute-based text generation},author={Wen, Zhihua and Tian, Zhiliang and Huang, Zhen and Yang, Yuxin and Jian, Zexin and Wang, Changjian and Li, Dongsheng},booktitle={Findings of the Association for Computational Linguistics: ACL 2023},month=jul,publisher={"Association for Computational Linguistics"},url={"https://aclanthology.org/2023.findings-acl.530/"},pages={8377--8398},year={2023},}
NeurIPS24
Perception of Knowledge Boundary for Large Language Models through Semi-open-ended Question Answering
Zhihua Wen, Zhiliang Tian, Zexin Jian, and 5 more authors
In The Thirty-eighth Annual Conference on Neural Information Processing Systems, Jul 2024
"Large Language Models (LLMs) are widely used for knowledge-seeking purposes yet suffer from hallucinations. The knowledge boundary of an LLM limits its factual understanding, beyond which it may begin to hallucinate. Investigating the perception of LLMs’knowledge boundary is crucial for detecting hallucinations and LLMs’ reliable generation. Current studies perceive LLMs’ knowledge boundary on questions with a concrete answer (close-ended questions) while paying limited attention to semi-open-ended questions that correspond to many potential answers. Some researchers achieve it by judging whether the question is answerable or not. However, this paradigm is not so suitable for semi-open-ended questions, which are usually partially answerable questions containing both answerable answers and ambiguous (unanswerable) answers. Ambiguous answers are essential for knowledge-seeking, but they may go beyond the knowledge boundary of LLMs. In this paper, we perceive the LLMs’ knowledge boundary with semi-open-ended questions by discovering more ambiguous answers. First, we apply an LLM-based approach to construct semi-open-ended questions and obtain answers from a target LLM. Unfortunately, the output probabilities of mainstream black-box LLMs are inaccessible to sample for low-probability ambiguous answers. Therefore, we apply an open-sourced auxiliary model to explore ambiguous answers for the target LLM. We calculate the nearest semantic representation for existing answers to estimate their probabilities, with which we reduce the generation probability of high-probability existing answers to achieve a more effective generation. Finally, we compare the results from the RAG-based evaluation and LLM self-evaluation to categorize four types of ambiguous answers that are beyond the knowledge boundary of the target LLM. Following our method, we construct a dataset to perceive the knowledge boundary for GPT-4. We find that GPT-4 performs poorly on semi-open-ended questions and is often unaware of its knowledge boundary. Besides, our auxiliary model, LLaMA-2-13B, is effective in discovering many ambiguous answers, including correct answers neglected by GPT-4 and delusive wrong answers GPT-4 struggles to identify. "
@inproceedings{LLMBoundary-Nips24,title={Perception of Knowledge Boundary for Large Language Models through Semi-open-ended Question Answering},author={Wen, Zhihua and Tian, Zhiliang and Jian, Zexin and Huang, Zhen and Ke, Pei and Gao, Yifu and Huang, Minlie and Li, Dongsheng},booktitle={The Thirty-eighth Annual Conference on Neural Information Processing Systems},year={2024},}