Yandex Publishes YaLM 100B. It's the Largest GPT-Like Neural Network in Open Source
In recent years, large-scale transformer-based language models have become the pinnacle of neural networks used in NLP tasks. They grow in scale and complexity every month, but training such models requires millions of dollars, the best experts, and years of development. That's why only major IT companies have access to this state-of-the-art technology.
https://medium.com/yandex/yandex-publishes-yalm-100b-its-the-largest-gpt-like-neural-network-in-open-source-d1df53d0e9a6