Summary MDN's new "ai explain" button on code blocks generates human-like text that may be correct by happenstance, or may contain convincing falsehoods. this is a strange decision for a technical ...
maybe there was good intentions by whoever implemented it
If an executive saying “find ways to use ChatGPT so we can be on the cutting edge” and a developer saying “eh, I guess maybe…” counts as good intentions.
If an executive saying “find ways to use ChatGPT so we can be on the cutting edge” and a developer saying “eh, I guess maybe…” counts as good intentions.