I thought there were quite a few problems with Watson, but, TBF, I did not follow it closely.
However, I do like the idea of using LLM(s) as another pair of eyes in the system, if you will. But only as another tool, not a crutch, and certainly not making any final calls. LLMs should be treated exactly like you’d treat a spelling checker or a grammar checker - if it’s pointing something out, take a closer look, perhaps. But to completely cede your understanding of something (say, spelling or grammar, or in this case, medicine that people take years to get certified in) to a tool is rather foolish.
I thought there were quite a few problems with Watson, but, TBF, I did not follow it closely.
However, I do like the idea of using LLM(s) as another pair of eyes in the system, if you will. But only as another tool, not a crutch, and certainly not making any final calls. LLMs should be treated exactly like you’d treat a spelling checker or a grammar checker - if it’s pointing something out, take a closer look, perhaps. But to completely cede your understanding of something (say, spelling or grammar, or in this case, medicine that people take years to get certified in) to a tool is rather foolish.
I couldn’t have said it better myself and completely agree. Use as an assistant; just not the main driver or final decision-maker.