Story Details

  • Coping with dumb LLMs using classic ML

    Posted: 2025-01-22 09:25:07

    The blog post explores using traditional machine learning (specifically, decision trees) to interpret and refine the output of less capable or "dumb" Large Language Models (LLMs). The author describes a scenario where an LLM is tasked with classifying customer service tickets, but its performance is unreliable. Instead of relying solely on the LLM's classification, a decision tree model is trained on the LLM's output (probabilities for each classification) along with other readily available features of the ticket, like length and sentiment. This hybrid approach leverages the LLM's initial analysis while allowing the decision tree to correct inaccuracies and improve overall classification performance, ultimately demonstrating how simpler models can bolster the effectiveness of flawed LLMs in practical applications.

    Summary of Comments ( 44 )
    https://news.ycombinator.com/item?id=42790820

    Hacker News users discuss the practicality and limitations of the proposed decision-tree approach to mitigate LLM "hallucinations." Some express skepticism about its scalability and maintainability, particularly with the rapid advancement of LLMs, suggesting that improving prompt engineering or incorporating retrieval mechanisms might be more effective. Others highlight the potential value of the decision tree for specific, well-defined tasks where accuracy is paramount and the domain is limited. The discussion also touches on the trade-off between complexity and performance, and the importance of understanding the underlying limitations of LLMs rather than relying on patches. A few commenters note the similarity to older expert systems and question if this represents a step back in AI development. Finally, some appreciate the author's honest exploration of alternative solutions, acknowledging that relying solely on improving LLM accuracy might not be the optimal path forward.