[Text Machine Lab] BERT Busters: Outlier Dimensions that Disrupt Transformers
This is a post I wrote during my time in Text Machine Lab: https://text-machine-lab.github.io/blog/2020/bert-secrets/. It reports on an influential paper on ...
This is a post I wrote during my time in Text Machine Lab: https://text-machine-lab.github.io/blog/2020/bert-secrets/. It reports on an influential paper on ...
What I learned from organizing an introductory course on NLP for linguists at ESSLLI 2019.
A post inspired by an Uber ride with a Trump supporter.
Negative results are hard to publish, and even harder to make well-known. Even when the disproved result is something as pervasive as Mikolov’s word analogies.
With the huge Transformer-based models such as BERT, GPT-2, and XLNet, are we losing track of how the state-of-the-art performance is achieved?