Google’s SMITH Language Model Outperforms BERT in Tests
https://www.searchenginejournal.comGoogle recently published a research paper on a new algorithm called SMITH (Siamese Multi-depth Transformer-based Hierarchical Encoder for Long-Form Document Matching) that it claims outperforms BERT for understanding long queries and long documents. In particular, what makes this new model better is that it is able to understand passages within documents in the same way BERT understands words and sentences, which enables the algorithm to understand longer documents.