Is there an extension that warns you when you are wasting time reading ai-generated crap?
Case in point, I was reading an article that claimed to compare kubernetes distros and wasted some good minutes before realizing it was full of crap.
Is there an extension that warns you when you are wasting time reading ai-generated crap?
Case in point, I was reading an article that claimed to compare kubernetes distros and wasted some good minutes before realizing it was full of crap.
If you could detect AI authored stuff, couldn’t you use that to train your LLM?
Suspect it would operate more on the basis of a person confirming that the article is of reasonable quality & accuracy.
So not unlike editors selecting what to publish, what to reject & what to send back for improvements.
If good articles by AI get accepted & poor articles by people get rejected, there may still be impacts, but at face value it might be sufficient for us seeking to read stuff.
It could be used to create a reward model like what is done right now with RLHF.