Correcting Hallucinations in Large Language Models