Subscribe to Our Newsletter

Success! Now Check Your Email

To complete Subscribe, click the confirmation link in your inbox. If it doesn't arrive within 3 minutes, check your spam folder.

Ok, Thanks

Amazon may be training an ambitious 2T-parameter LLM

Anonymous sources claim that Amazon is financing training for a 2 trillion parameter LLM codenamed 'Olympus', intending to compete with OpenAI and Google's most powerful offerings. Details of the project are scarce, and there seems to be no specific timeline for the release.

Ellie Ramirez-Camara profile image
by Ellie Ramirez-Camara
Amazon may be training an ambitious 2T-parameter LLM
Photo by Marques Thomas / Unsplash

According to people familiar with the project, Amazon is financing efforts to train a 2 trillion parameter large language model that can compete with OpenAI and Google's best offerings. OpenAI has never officially disclosed details on its GPT family of models. However, news site Semafor claimed they had "spoken to eight people familiar with the inside story" and published that GPT-4 had 1 trillion parameters. Since then, the unconfirmed estimate has taken off. If true, it would mean that Amazon's model, codenamed 'Olympus', would be twice as large as GPT-4 and one of the largest models in training.

The report originally appeared in Reuters and details that Rohit Prasad, the former head of Alexa, leads the team behind Olympus. Prasad now reports directly to CEO Andy Jassy. He is unifying the AI efforts across the company by bringing together researchers previously employed at Alexa AI and the Amazon science team to work on training models. According to the two anonymous sources, there isn't a specific timeline for the model's release.

Acquiring enough computing power to train large-sized models is costly. As a result, Amazon recently stated that it "would increase investment in LLMs and generative AI while cutting back on fulfillment and transportation in its retail business." The expectation is that once released, Olympus will join Amazon's smaller model Titan on Amazon Bedrock, the platform available for Amazon Web Services (AWS) users to build and scale generative AI applications. Bedrock allows developers to deploy high-quality proprietary and open-source models such as Anthropic's Claude, Cohere's Command, Stable Diffusion XL, and soon, Meta's Llama 2. Olympus would be the largest model to join the lineup at Bedrock.

Ellie Ramirez-Camara profile image
by Ellie Ramirez-Camara

Data Phoenix Digest

Subscribe to the weekly digest with a summary of the top research papers, articles, news, and our community events, to keep track of trends and grow in the Data & AI world!

Success! Now Check Your Email

To complete Subscribe, click the confirmation link in your inbox. If it doesn’t arrive within 3 minutes, check your spam folder.

Ok, Thanks

Read More