Do I need to tackle hallucinations?

tools: semantic comparison. don't let LLM be the judge

need Agent framework ? Lang chain ?

malicious input detection - does the student input try to manipulate ? idea : semantic comparison to malicious commands/ jailbreaks

Anti spamming - save gpu costs