Quantcast
Channel: World Consulting Group
Viewing all articles
Browse latest Browse all 1358

Why Machine Learning Models Should be Smaller in Size?

$
0
0

Recently, analysts at NVIDIA declared MegatronLM, a huge transformer model with 8.3 billion parameters (multiple times bigger than BERT) that accomplished cutting-edge performance on a variety of language tasks.

There are numerous instances of monstrous models being trained to accomplish somewhat higher precision on different benchmarks. In spite of being 24X bigger than BERT, MegatronLM is just 34% better at its language modeling task. As a coincidental trial to exhibit the performance of new hardware, there isn’t a lot of damage here. However, in the long-term, this pattern will cause a couple of issues.

As more artificial intelligence

Management Consulting

The post Why Machine Learning Models Should be Smaller in Size? appeared first on World Consulting Group.


Viewing all articles
Browse latest Browse all 1358

Trending Articles



<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>