A company compares two models for a customer-support assista…
A company compares two models for a customer-support assistant. An RNN-based model reads token by token; a Transformer uses self-attention. On long chats the RNN often confuses which product the customer is referring to, especially when the product name appeared much earlier. Which explanation is most plausible?
Read Details