• About us
  • Privacy Policy
  • Terms and Conditions
  • Contact us
News Hacker
  • Home
    • Home 2
    • Home 3
    • Home 4
    • Home 5
    • Home 6
  • News
  • Politics
  • Science
  • Sports
  • World
  • About us
  • Terms and Conditions
  • Privacy Policy
  • Contact us
No Result
View All Result
  • Home
    • Home 2
    • Home 3
    • Home 4
    • Home 5
    • Home 6
  • News
  • Politics
  • Science
  • Sports
  • World
  • About us
  • Terms and Conditions
  • Privacy Policy
  • Contact us
No Result
View All Result
News Hacker
No Result
View All Result
Home Technology

Google’s PaLM 2 uses nearly five times more text data than predecessor

anmolkumarengineer by anmolkumarengineer
May 17, 2023
in Technology
0
Google’s PaLM 2 uses nearly five times more text data than predecessor
0
SHARES
0
VIEWS
Share on FacebookShare on Twitter


  • Google’s PaLM 2 large language model is using nearly five times the amount of text data for training as its predecessor LLM, CNBC has learned.
  • In announcing PaLM 2 last week, Google said the model is smaller than the earlier PaLM but uses a more efficient “technique.”
  • The lack of transparency around training data in artificial intelligence models is becoming an increasingly hot topic among researchers.

Sundar Pichai, chief executive officer of Alphabet Inc., during the Google I/O Developers Conference in Mountain View, California, on Wednesday, May 10, 2023.

David Paul Morris | Bloomberg | Getty Images

Google’s new large language model, which the company announced last week, uses almost five times as much training data as its predecessor from 2022, allowing its to perform more advanced coding, math and creative writing tasks, CNBC has learned.

PaLM 2, the company’s new general-use large language model (LLM) that was unveiled at Google I/O, is trained on 3.6 trillion tokens, according to internal documentation viewed by CNBC. Tokens, which are strings of words, are an important building block for training LLMs, because they teach the model to predict the next word that will appear in a sequence.

Google’s previous version of PaLM, which stands for Pathways Language Model, was released in 2022 and trained on 780 billion tokens.

While Google has been eager to showcase the power of its artificial intelligence technology and how it can be embedded into search, emails, word processing and spreadsheets, the company has been unwilling to publish the size or other details of its training data. OpenAI, the Microsoft-backed creator of ChatGPT, has also kept secret the specifics of its latest LLM called GPT-4.

The reason for the lack of disclosure, the companies say, is the competitive nature of the business. Google and OpenAI are rushing to attract users who may want to search for information using conversational chatbots rather than traditional search engines.

But as the AI arms race heats up, the research community is demanding greater transparency.

Since unveiling PaLM 2, Google has said the new model is smaller than prior LLMs, which is significant because it means the company’s technology is becoming more efficient while accomplishing more sophisticated tasks. PaLM 2, according to internal documents, is trained on 340 billion parameters, an indication of the complexity of the model. The initial PaLM was trained on 540 billion parameters.

Google didn’t immediately provide a comment for this story.

Google said in a blog post about PaLM 2 that the model uses a “new technique” called “compute-optimal scaling.” That makes the LLM “more efficient with overall better performance, including faster inference, fewer parameters to serve, and a lower serving cost.”

In announcing PaLM 2, Google confirmed CNBC’s previous reporting that the model is trained on 100 languages and performs a broad range of tasks. It’s already being used to power 25 features and products, including the company’s experimental chatbot Bard. It’s available in four sizes, from smallest to largest: Gecko, Otter, Bison and Unicorn. 

PaLM 2 is more powerful than any existing model, based on public disclosures. Facebook’s LLM called LLaMA, which it announced in February, is trained on 1.4 trillion tokens. The last time OpenAI shared ChatGPT’s training size was with GPT-3, when the company said it was trained on 300 billion tokens at the time. OpenAI released GPT-4 in March, and said it exhibits “human-level performance” on many professional tests.

LaMDA, a conversation LLM that Google introduced two years ago and touted in February alongside Bard, was trained on 1.5 trillion tokens, according to the latest documents viewed by CNBC.

As new AI applications quickly hit the mainstream, controversies surrounding the underlying technology are getting more spirited.

El Mahdi El Mhamdi, a senior Google Research scientist, resigned in February over the company’s lack of transparency. On Tuesday, OpenAI CEO Sam Altman testified at a hearing of the Senate Judiciary subcommittee on privacy and technology, and agreed with lawmakers that a new system to deal with AI is needed.

“For a very new technology we need a new framework,” Altman said. “Certainly companies like ours bear a lot of responsibility for the tools that we put out in the world.”

— CNBC’s Jordan Novet contributed to this report.

WATCH: OpenAI CEO Sam Altman calls for A.I. oversight

Previous Post

IPL 2023: Marcus Stoinis, Ravi Bishnoi help LSG seal victory against MI | Cricket

Next Post

AAP’s ‘Put On Hold’ Circular After Supreme Court Order On Delhi Lt Governor

anmolkumarengineer

anmolkumarengineer

Next Post
AAP’s ‘Put On Hold’ Circular After Supreme Court Order On Delhi Lt Governor

AAP's 'Put On Hold' Circular After Supreme Court Order On Delhi Lt Governor

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Follow us on Google news

Checkout Best Deals&Offers for INDIA only

Best Gaming Headphones

Best 5G phones

Best Gaming Laptops

Recent News

A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

July 31, 2023
Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

July 31, 2023
Tragic Suicide Bombing at Election Rally in Pakistan Leaves 54 Dead

Pakistan news: Tragic Suicide Bombing at Election Rally in Pakistan Leaves 54 Dead

July 31, 2023
Exciting News: 'Made in Heaven Season 2' Trailer Release Date Announced

Exciting News: ‘Made in Heaven Season 2’ Trailer Release Date Announced

July 31, 2023
  • Trending
  • Comments
  • Latest
CBSE Class 10 and 12 Result 2023: How to Access and Download Mark Sheets on DigiLocker

CBSE Class 10 and 12 Result 2023: How to Access and Download Mark Sheets on DigiLocker

May 11, 2023
CBSE Class 10 and Class 12 Board Exam Results 2023: All You Need to Know

CBSE Class 10 and Class 12 Board Exam Results 2023: All You Need to Know

May 7, 2023
Cricket World Cup 2023 Tickets – The Free Media

Cricket World Cup 2023 Tickets – The Free Media

July 4, 2023
Bengaluru techie dead after car enters flooded K R Circle underpass

Bengaluru techie dead after car enters flooded K R Circle underpass

May 22, 2023
Gaming Smartwatch 2023

Gaming Smartwatch 2023

0
Best Headphones under 5000

Best Headphones under 5000

0
5G Phone Under 10000

5G Phone Under 10000

0
Realme 5G phone under 20000

Realme 5G phone under 20000

0
A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

July 31, 2023
Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

July 31, 2023
Tragic Suicide Bombing at Election Rally in Pakistan Leaves 54 Dead

Pakistan news: Tragic Suicide Bombing at Election Rally in Pakistan Leaves 54 Dead

July 31, 2023
Exciting News: 'Made in Heaven Season 2' Trailer Release Date Announced

Exciting News: ‘Made in Heaven Season 2’ Trailer Release Date Announced

July 31, 2023
News Hacker

Welcome to The Newshacker, your go-to source for reliable, unbiased and up-to-date news around the world. Our mission is to provide our readers with the most important and relevant news stories of the day, delivered in an engaging and informative manner.

Follow Us

Browse by Category

  • Bags
  • Business
  • Electronics
  • Entertainment
  • Food
  • Gadget
  • Health
  • India
  • Lifestyle
  • Mobile
  • Mobile phones
  • Movie
  • News
  • Politics
  • Review
  • Science
  • Shoes
  • Smartwatch
  • Sports
  • Technology
  • Top Stories
  • Uncategorized
  • World

Recent News

A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

A New Cybersecurity Threat: FraudGPT Unleashed on the Darknet

July 31, 2023
Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

Tragic Loss: French Daredevil Remi Lucidi Falls to His Death While Attempting High-rise Stunt

July 31, 2023
  • About
  • Terms and Conditions
  • Privacy & Policy
  • Contact

© 2023 Thenews Hacker - All rights reserved by Thenews Hacker.

No Result
View All Result
  • Home
    • Home 2
    • Home 3
    • Home 4
    • Home 5
    • Home 6
  • News
  • Politics
  • Science
  • Sports
  • World
  • About us
  • Terms and Conditions
  • Privacy Policy
  • Contact us

© 2023 Thenews Hacker - All rights reserved by Thenews Hacker.

Go to mobile version