Do I need to tackle hallucinations?
tools: semantic comparison. don't let LLM be the judge
need Agent framework ? Lang chain ?
malicious input detection - does the student input try to manipulate ? idea : semantic comparison to malicious commands/ jailbreaks
Anti spamming - save gpu costs