Open Ayush060201 opened 3 days ago
New Problem Statement
Title Optimize InterrogateLLM for Enhanced Hallucination Detection in LLMs
Category Optimization
Problem Statement Large Language Models (LLMs) have shown remarkable capabilities in generating human-like text, but they often produce hallucinations - plausible-sounding but factually incorrect information. This poses significant challenges for their reliable use in real-world applications. While the InterrogateLLM method presents a novel approach to zero-resource hallucination detection, it still faces several limitations that impact its accuracy and efficiency.
Evaluation Strategy
Metrics:
IOU (Intersection over Union) Score:
Comparison with Baselines:
Dataset Movies Dataset - https://www.kaggle.com/datasets/rounakbanik/the-movies-dataset Books Dataset - https://www.kaggle.com/datasets/saurabhbagchi/books-dataset
Resources [1] https://arxiv.org/abs/2403.02889
Title
IntelliHelp: RAG-Powered Customer Assistance Using LLM
Team Name
Run Time Errorists
Email
202311048@daiict.ac.in
Team Member 1 Name
Shyam Saktawat
Team Member 1 Id
202311048
Team Member 2 Name
Abhishek Choudhary
Team Member 2 Id
202311067
Team Member 3 Name
Ayush Kumar Sahu
Team Member 3 Id
202311066
Team Member 4 Name
NIL
Team Member 4 Id
NIL
Category
New Research Problem
Problem Statement
Current organizational chatbots struggle to autonomously resolve user queries, as they fail to effectively integrate RAG and LLMs with proprietary data. IntelliHelp aims to offer personalized, real-time customer support by retrieving relevant information and generating dynamic, accurate responses without human intervention.
Evaluation Strategy
Dataset
https://github.com/unicamp-dl/retailGPT/tree/main/retailGPT/datasets
Resources
[1] LEWIS, Patrick et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. arXiv, 2020. [2] XU, Ziwei; JAIN, Sanjay; KANKANHALLI, Mohan. Hallucination is Inevitable: An Innate Limitation of Large Language Models. arXiv, 2024. [3] WEI, Jason; WANG, Xuezhi; SCHUURMANS, Dale et al. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models. arXiv, 2022.