Articles by Mark
4 articles

Technical
How to Perform Hallucination Detection for LLMs
Learn five different methods to detect LLM hallucinations with 90% accuracy for closed-domain question answering.

Technical
The Five Pillars of Trustworthy LLM Testing
From the perspective of a machine learning engineer at Kolena, this article is a continuation in a series of discussions highlighting a testing...

Technical
Quantifying GPT-4’s Hidden Regressions Over Time (Series Part 3)
Kolena's analysis reveals GPT-4's performance regressions in conversational question answering across various data sources, highlighting continuous...

Technical
How Well Do GPT Models Follow Prompts? (Series Part 2)
Explore how different prompt designs impact GPT-3.5 Turbo's performance in text summarization. Learn which prompts yield the most accurate outputs...