I’m quite a big fan of perplexity AI, which shows you sources it used to generate the answers. One thing I often do is type a question, glance the automated answer and then jump to the source to see what the users said (basically I use it like a tailored search engine)
Admittedly, there’s nothing stopping the company from throwing up fake sources to “legitimize” their answers, but I think that once models become more open (e.g. AMD’s recent open weights addition is an amazing leap forward) it will be harder to slip in fake sources
Email and message summarization also introduces new problems that don’t happen when using a chatbot for questions and answers since through the process of summarization it removes information from the original text and may remove key information or mischaracterize the message. The ways it may do this stuff isn’t exactly predictable either. It’s also harder since it’s not about proving that something is true or not based on outside sources, it’s about it being accurate to what they said, which may not be provable to outside sources.
I’ve found summarization to be relatively trustworthy. Perplexity does not appear to hallucinate much, and on the odd occasion it does, I dive into the sources it provides.
I’m quite a big fan of perplexity AI, which shows you sources it used to generate the answers. One thing I often do is type a question, glance the automated answer and then jump to the source to see what the users said (basically I use it like a tailored search engine)
Admittedly, there’s nothing stopping the company from throwing up fake sources to “legitimize” their answers, but I think that once models become more open (e.g. AMD’s recent open weights addition is an amazing leap forward) it will be harder to slip in fake sources
Email and message summarization also introduces new problems that don’t happen when using a chatbot for questions and answers since through the process of summarization it removes information from the original text and may remove key information or mischaracterize the message. The ways it may do this stuff isn’t exactly predictable either. It’s also harder since it’s not about proving that something is true or not based on outside sources, it’s about it being accurate to what they said, which may not be provable to outside sources.
I’ve found summarization to be relatively trustworthy. Perplexity does not appear to hallucinate much, and on the odd occasion it does, I dive into the sources it provides.