GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub
集合主题趋势排行榜
#

hallucinations

Website
Wikipedia
https://static.github-zh.com/github_avatars/vectara?size=40
vectara / hallucination-leaderboard

#大语言模型#Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents

generative-aihallucinations大语言模型
Python 2.51 k
2 天前
https://static.github-zh.com/github_avatars/EdinburghNLP?size=40
EdinburghNLP / awesome-hallucination-detection

#自然语言处理#List of papers on hallucination detection in LLMs.

hallucinations大语言模型自然语言处理
888
20 天前
https://static.github-zh.com/github_avatars/VITA-MLLM?size=40
VITA-MLLM / Woodpecker

#大语言模型#✨✨Woodpecker: Hallucination Correction for Multimodal Large Language Models

hallucinationhallucinationslarge-language-models大语言模型mllmmultimodal-large-language-modelsmultimodality
Python 636
6 个月前
https://static.github-zh.com/github_avatars/voidism?size=40
voidism / DoLa

Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"

factualityhallucinationslarge-language-models
Python 492
5 个月前
https://static.github-zh.com/github_avatars/MinghuiChen43?size=40
MinghuiChen43 / awesome-trustworthy-deep-learning

#计算机科学#A curated list of trustworthy deep learning papers. Daily updating...

adversarial-machine-learning安全隐私深度学习poisoningfairnessbackdoorownershiprobustnessinterpretable-deep-learningcausalityhallucinationsuncertaintywatermarkingai-alignment
369
5 天前
https://static.github-zh.com/github_avatars/pegasi-ai?size=40
pegasi-ai / feather

AI Testing Toolkit for AI applications

hallucinationsaccuracyragretrieval-augmented-generationsafety安全
Python 318
4 个月前
https://static.github-zh.com/github_avatars/lechmazur?size=40
lechmazur / confabulations

#大语言模型#Hallucinations (Confabulations) Document-Based Benchmark for RAG. Includes human-verified questions and answers.

benchmarkclaudegeminihallucinationsleaderboardllama大语言模型raglanguage-modeldeepseek-r1gemini-proo1o3-mini
HTML 171
4 天前
https://static.github-zh.com/github_avatars/IAAR-Shanghai?size=40
IAAR-Shanghai / UHGEval

#大语言模型#[ACL 2024] User-friendly evaluation framework: Eval Suite & Benchmarks: UHGEval, HaluEval, HalluQA, etc.

benchmarkdatasetevaluation大语言模型ChatGPTgpt-3gpt-4hallucinationslarge-language-modelsqwenhallucinationhuggingfacehuggingface-transformersopenaiopenai-apiceval
Python 167
9 天前
https://static.github-zh.com/github_avatars/PKU-YuanGroup?size=40
PKU-YuanGroup / Hallucination-Attack

#自然语言处理#Attack to induce LLMs within hallucinations

adversarial-attacks大语言模型hallucinations机器学习自然语言处理ai-safety深度学习
Python 155
1 年前
https://static.github-zh.com/github_avatars/ictnlp?size=40
ictnlp / TruthX

#大语言模型#Code for ACL 2024 paper "TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space"

hallucinationslanguage-model大语言模型llm-inferencebaichuanchatglmChatGPTgpt-4hallucinationllamallama2mistralsafetyrepresentationexplainable-aillama3
Python 147
1 年前
https://static.github-zh.com/github_avatars/OpenMOSS?size=40
OpenMOSS / HalluQA

Dataset and evaluation script for "Evaluating Hallucinations in Chinese Large Language Models"

hallucinationslarge-language-modelsquestion-answering
Python 128
1 年前
https://static.github-zh.com/github_avatars/voidism?size=40
voidism / Lookback-Lens

Code for the EMNLP 2024 paper "Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps"

hallucinationslarge-language-modelstext-generationfactuality
Python 126
10 个月前
https://static.github-zh.com/github_avatars/LLAMATOR-Core?size=40
LLAMATOR-Core / llamator

#自然语言处理#Framework for testing vulnerabilities of large language models (LLM).

attack大语言模型自然语言处理Python安全ai-securityred-teamhallucinationsllm-securityrag-evaluation人工智能ragjailbreakowaspred-team-toolsagentvulnerability
Python 119
6 天前
https://static.github-zh.com/github_avatars/rungalileo?size=40
rungalileo / hallucination-index

#大语言模型#Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.

hallucinationslarge-language-models大语言模型llm-evaluationopenairagretrieval-augmented-generation
110
9 个月前
https://static.github-zh.com/github_avatars/X-PLUG?size=40
X-PLUG / mPLUG-HalOwl

mPLUG-HalOwl: Multimodal Hallucination Evaluation and Mitigating

mllmmultimodal-large-language-modelsbenchmarkcontrastive-learninghallucinations
Python 95
1 年前
https://static.github-zh.com/github_avatars/BillChan226?size=40
BillChan226 / HALC

[ICML 2024] Official implementation for "HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding"

hallucinationslarge-language-models
Python 88
6 个月前
https://static.github-zh.com/github_avatars/OpenKG-ORG?size=40
OpenKG-ORG / EasyDetect

#自然语言处理#An Easy-to-use Hallucination Detection Framework for LLMs.

hallucinationsknowledge-graphlarge-language-modelsmultimodal-large-language-models自然语言处理aigcgeneration
Python 59
1 年前
https://static.github-zh.com/github_avatars/hongbinye?size=40
hongbinye / Cognitive-Mirage-Hallucinations-in-LLMs

Repository for the paper "Cognitive Mirage: A Review of Hallucinations in Large Language Models"

hallucinationslarge-language-modelsnatural-language-generationmachine-translation
47
2 年前
https://static.github-zh.com/github_avatars/intuit?size=40
intuit / sac3

#大语言模型#Official repo for SAC3: Reliable Hallucination Detection in Black-Box Language Models via Semantic-aware Cross-check Consistency

blackboxconsistencyhallucinationslarge-language-models大语言模型reliabilitysemantic
Jupyter Notebook 35
5 个月前
https://static.github-zh.com/github_avatars/ChanLiang?size=40
ChanLiang / CONNER

#大语言模型#The implementation for EMNLP 2023 paper ”Beyond Factuality: A Comprehensive Evaluation of Large Language Models as Knowledge Generators“

llm-evaluationhallucinationsemnlp2023large-language-modelsfactualityChatGPTllama
Python 31
1 年前
loading...