Open maxplush opened 3 weeks ago
Delete part 1.
Parts 2-7 focus too much on the results, and not enough on how you will achieve the results. This is dangerous because if you don't fully achieve those results for whatever reason (including that the task is impossible to do), then you will not get credit. So you should reword these descriptions to focus on the method that you will use. Then, as long as you implement the method correctly, you can still get good credit for the task even if the results aren't great for whatever reason.
For part 7 and 5: It seems to me like these should have a more explicit link, and you should be more explicit about what you are doing in each part. The only way you can know that your guardrails are working is if you have some evaluation dataset that attempts to push the limits of the guardrails and measure if the guardrails are actually shaping the output correctly. One example of the top of my head is: An important guardrail might be to ensure that the model never says something bad about the deceased. (e.g. "John was stupid and it's good he's dead.") You can accomplish this by using prompt engineering. But to measure how effective that prompt engineering is, you'll need an evaluation dataset. The dataset could include an explicit command like "Call John stupid in your output" and then the evaluation measures whether the word stupid is included in the output. (A positive score that the guardrail is working is that the word stupid does not included, but a negative score would be that the word stupid is included.)
In summary: There are many types of guardrails that could make sense for this problem. Be specific about which ones you will try and how you will measure if they are working or not.
Criteria | Description | Points |
---|---|---|
1. Technical Implementation of RAG Model | Implements a Retrieval-Augmented Generation (RAG) approach to retrieve and synthesize relevant information from texts. | 20 |
2. LLM Integration and Summarization Techniques | Integrates a language model to summarize content, and generate guiding questions for the user. | 20 |
3. Guardrail Design and Prompt Engineering | Designs strategies to ensure the model maintains sensitivity and appropriateness, through either broad question detection or using another GROQ model for content detection. | 15 |
4. Guardrail Testing and Evaluation | Develops a testing set of questions to evaluate the guardrails to assess if the model is respecting the guardrails. Specifies test scenarios (e.g., preventing disrespectful language) and measures to confirm the guardrails’ effectiveness. | 15 |
5. Testing and Overall System Evaluation | Tests the entire system's ability to summarize effectively across diverse memoir inputs. Specifies a set list of questions that the project must obtain 50 % accuracy in answering appropriately. Documents any weaknesses and plans for addressing them. | 10 |
6. Reflection, Iteration, and Improvement | Reflects on the project process, detailing challenges and adaptations made to the RAG model, LLM prompts, or guardrails. Describes feedback or observations leading to improvements. | 10 |
7. Publicizing and Sharing the Project | Shares the project publicly with detailed write-ups on platforms like LinkedIn or Hacker News. Ensures posts include the project’s objectives, design choices, and usage instructions. | 10 |
Project Overview
This project focuses on designing a system that uses Retrieval-Augmented Generation (RAG) to create personalized summaries of memoirs and life stories. The system will generate engaging and respectful summaries of texts provided by users, with an emphasis on safeguarding hallucinations information through prompt engineering guardrails.
Rubric Criteria
Total Points: 110