GPT-3

SambaNova GPT
OpenAI’s GPT-3 Vs SambaNova System’s GPT

California-based SambaNova Systems recently announced the launch of its own GPT language model to be available in a data-as-a-service model.

NVIDIA, Microsoft Introduce New Language Model MT-NLG With 530 Billion Parameters, Leaves GPT-3 Behind
NVIDIA, Microsoft Introduce New Language Model MT-NLG With 530 Billion Parameters, Leaves GPT-3 Behind

MT-NLG has 3x the number of parameters compared to the existing largest models – GPT-3, Turing NLG, Megatron-LM and others.

GPT-3 books
All Recent Books Written By GPT-3

Generative Pre-trained Transformer 3, or GPT-3, is an autoregressive language model that can produce human-like…

Baidu Launches World’s Largest Dialogue Generation Model With 11 Billion Parameters
Baidu Launches World’s Largest Dialogue Generation Model With 11 Billion Parameters

PLATO-XL is trained on a high-performance GPU cluster with 256 NVIDIA Tesla V100 32G GPU cards.

Are Larger Language Models Less Truthful?

Researchers at the University of Oxford and OpenAI have recently created a dataset called TruthfulQA that contains questions some humans might answer incorrectly due to false beliefs or misconceptions.

Google Introduces New Architecture To Reduce Cost Of Transformers
Google Introduces New Architecture To Reduce Cost Of Transformers

Primer’s improvements can be attributed to two simple modifications — squaring ReLU activations and adding a depthwise convolution layer after each Q, K, and V projection in self-attention.

Facebook’s Answer To GPT-3, Textless NLP
Facebook’s Answer To GPT-3, Textless NLP

GSLM uses the latest breakthroughs in representation learning, allowing it to work directly from raw audio signals, without any text or labels.

Bugged: The Ugly Side Of No Code AI Platforms

Snippet: Copilot is based on the OpenAI Codex family of models. Codex models begin with a GPT-3 model, and then fine-tune it on code from GitHub.

Foundation Models
Risks Of Using Foundational Models Such As GPT-3

Compared to most other machine learning models, foundation models are characterised by a vast increase in training data and complexity

Jurassic
Jurassic-1 vs GPT-3 vs Everyone Else

With its 178 billion parameters, Jurassic-1 is slightly bigger (3 billion more) than GPT-3.

With A Rush To Create Larger Language Models, Are We Beating Their Purpose

Language Models trained on large, uncurated, static datasets from the Web encode hegemonic views that are harmful to marginalised populations.

What’s Bigger Than GPT3? It’s “Jurassic”

AI21 Studio allows developers to easily customise a private version of Jurassic-1 models, shortening time to production and lowering costs.