Share this page:

Characterizing Truthfulness in Large Language Model Generations with Local Intrinsic Dimension

Fan Yin, Jayanth Srinivasa, and Kai-Wei Chang, in ICML, 2024.

Download the full text


Abstract


Bib Entry

@inproceedings{yin2024charactering,
  title = {Characterizing Truthfulness in Large Language Model Generations with Local Intrinsic Dimension},
  author = {Yin, Fan and Srinivasa, Jayanth and Chang, Kai-Wei},
  booktitle = {ICML},
  year = {2024}
}

Related Publications

  1. Control Large Language Models via Divide and Conquer, EMNLP, 2024
  2. Re-ReST: Reflection-Reinforced Self-Training for Language Agents, EMNLP, 2024
  3. Agent Lumos: Unified and Modular Training for Open-Source Language Agents, ACL, 2024
  4. TrustLLM: Trustworthiness in Large Language Models, ICML, 2024
  5. The steerability of large language models toward data-driven personas, NAACL, 2024
  6. AI-Assisted Summarization of Radiologic Reports: Evaluating GPT3davinci, BARTcnn, LongT5booksum, LEDbooksum, LEDlegal, and LEDclinical, American Journal of Neuroradiology, 2024
  7. Understanding and Mitigating Spurious Correlations in Text Classification with Neighborhood Analysis, EACL-Findings, 2024
  8. Few-Shot Representation Learning for Out-Of-Vocabulary Words, ACL, 2019
  9. Learning Word Embeddings for Low-resource Languages by PU Learning, NAACL, 2018
  10. Co-training Embeddings of Knowledge Graphs and Entity Descriptions for Cross-lingual Entity Alignment, IJCAI, 2018
  11. Beyond Bilingual: Multi-sense Word Embeddings using Multilingual Context, ACL RepL4NLP Workshop, 2017