[R] Microsoft DeBERTa Tops Human Performance on SuperGLUE NLU Benchmark

  • by

SuperGLUE met its match this week when, for the first time, a new model surpassed human baseline performance on the challenging natural language understanding (NLU) benchmark.

Dubbed DeBERTa (Decoding-enhanced BERT with disentangled attention), the breakthrough Transformer-based neural language model was initially introduced by a team of researchers from Microsoft Dynamics 365 AI and Microsoft Research in June of last year. Recently scaled up to 1.5 billion parameters, DeBERTa “substantially” outperformed the previous SuperGLUE leader — Google’s 11 billion parameter T5 — and surpassed the human baseline with a score of 89.9 (vs. 89.8).

Here is a quick read: Microsoft DeBERTa Tops Human Performance on SuperGLUE NLU Benchmark

The paper DeBERTa: Decoding-enhanced BERT with Disentangled Attention is on arXiv. The team will update their GitHub code repository soon with the latest DeBERTa code and models.

submitted by /u/Yuqing7
[link] [comments]

Leave a Reply

Your email address will not be published. Required fields are marked *