Hallucination is fundamental to how transformer-based language models work. In fact, it's their greatest asset.
There’s always some debate around what style of architecture is best for certain computing applications, with some on the RISC side citing performance per watt and some on the CISC side ...