Addressing Hallucinations in LLMs: How Guardrails, RAG, and Fact Verification Tools Combat AI Errors