Welcome to this week’s roundup of triple-distilled artisanal AI news.

This week, corporations that scrape your data free of charge complained that others were stealing it.

Big Tech is making its own AI chips to compete with NVIDIA.

And a bunch of recent AI models are flying off the shelves.

Let’s dig in.

AI data hunger games

AI’s insatiable hunger for data continues as corporations scramble for text, audio, and video to coach their models.

It’s an open secret that OpenAI almost actually scraped YouTube video content to coach its text-to-video model Sora. YouTube CEO Neal Mohan warned OpenAI that it could have breached its terms of service. Well, boohoo. Have you seen how cool Sora is?

You’ve got to think that Google will need to have cut some ethical corners to coach its models. Having YouTube cry foul over the “rights and expectations” of the creators of videos on its platform is a bit wealthy.

A better take a look at Big Tech’s tussle over AI training data reveals how Google amended its Google Docs privacy policies to make use of your data. Meanwhile, OpenAI and Meta proceed to push the legal and ethical boundaries within the pursuit of more training data.

I’m unsure where the information got here from for the brand new AI music generator Udio, but folks have been prompting it with some interesting ideas. This is proof AI needs to be regulated.

This is wild.

Udio just dropped and it’s like Sora for music.

The music are insane quality, 100% AI.

1. “Dune the Broadway Musical” pic.twitter.com/hzt7j32jIV

— Min Choi (@minchoi) April 10, 2024

Chips ahoy

All that controversial data needs to be processed and NVIDIA hardware is doing most of that. Sam takes us through NVIDIA’s rags-to-riches story from 1993 (when it is best to have bought shares) to now (once you would have been wealthy) and it’s fascinating.

While corporations keep lining up to purchase NVIDIA chips hot out of the oven, Big Tech is attempting to wean itself off its chips.

Intel and Google unveiled recent AI chips to compete with NVIDIA, despite the fact that they’ll still be buying NVIDIA’s Blackwell hardware.

Release the models!

It’s crazy that just over a yr ago, OpenAI had the one models getting any real attention. Now there’s a relentless stream of recent models from the Big Tech usual suspects and smaller startups.

This week we saw three AI models released inside 24 hours. Google’s Gemini Pro 1.5 now has a 1M token context window. Big context is great, but will its recall be pretty much as good as Claude 3?

There were interesting developments with OpenAI enabling API access to GPT-4 with vision, and Mistral just gave away one other powerful smaller model.

Politician turned Meta exec Nick Clegg spoke at Meta’s AI event in London to champion open-source AI. Clegg also said that Meta expects Llama 3 to roll out very soon.

During discussions around AI disinformation, he bizarrely downplayed AI’s role in attempts to influence recent major elections.

Does this guy even read the news we report here at Daily AI?

What issues of safety?

Geoffrey Hinton, considered the godfather of AI, was so concerned over AI safety that he quit Google. Meta’s Yann LeCun says there’s nothing to fret about. So which is it?

A Georgetown University study found that just 2% of AI research is concentrated on safety. Is that because there’s nothing to fret about? Or should we be concerned that researchers are focused on making more powerful AI with little thought to creating it protected?

Should ‘move fast and break stuff’ still be AI developers’ rallying cry?

The trajectory of AI development has been exponential. In an interview this week, Elon Musk said he expects AI could also be smarter than humans by the tip of 2025.

AI expert Gary Marcus doesn’t agree and he’s willing to place money on it.

Dear @elonmusk,

$1 million says your latest prediction – that AI can be smarter than any individual human by the tip of 2025 – is incorrect.

Game? I can suggest some rules on your approval.

Best wishes,
Gary

P.S. Note that in some respects (but not all) computers have been…

— Gary Marcus (@GaryMarcus) April 9, 2024

xAI is facing the identical NVIDIA chip shortage challenge many others are. Musk says the 20,000 NVIDIA H100s the corporate has will complete Grok 2’s training by May.

Guess what number of GPUs he says they’ll have to train Grok 3.

Anthropic streaks ahead

Anthropic says it develops large-scale AI systems in order that they can “study their safety properties on the technological frontier.”

The company’s Claude 3 Opus is actually on the frontier. A brand new study shows the model blows the remainder of the competition away in summarizing book-length content. Even so, the outcomes show that humans still have the sting in some respects.

Anthropic’s latest tests show that Claude LLMs have turn out to be exceptionally persuasive, with Claude 3 Opus generating arguments as persuasive as those created by humans.

People who play down the AI safety risks often say that you may simply pull the plug if an AI went rogue. What if the AI was so persuasive that it could persuade you not to do this?

Claude 3’s big claim to fame is its massive context window. Anthropic released a study that shows large context LLMs are vulnerable to a “many-shot” jailbreak technique. It’s super easy to implement and so they admit they don’t know find out how to fix it.

In other news…

Here are another clickworthy AI stories we enjoyed this week:

  • The virtual AI for Customer Success Summit 2024 kicks off.
  • Fake AI law firms are sending DMCA threats to generate website positioning gains.
  • Stability releases its multilingual lightweight Stable LM 2 12B LLM with impressive benchmark results.
  • Microsoft researchers propose “Visualization-of-Thought” prompting to enable LLMs to make use of spatial reasoning.
  • AI could possibly be as consequential to the economy as electricity, says JPMorgan Chase CEO, Jamie Dimon.
  • An AI-operated fighter jet will fly the Air Force Secretary in a test of the US military’s future drone warplanes.
  • The speed of AI development is outpacing risk assessment.
  • Meta announced its next generation of the Meta Training and Inference Accelerator (MTIA) chip.

And that’s a wrap.

Do you care that OpenAI, and certain others, could have used your YouTube content to coach their AI? If Altman released Sora free of charge then I’m guessing all could be forgiven. Google may disagree.

Do you think that Musk is being overly optimistic together with his AI intelligence predictions? I hope he’s right, but I’m a little bit uneasy that only 2% of AI research goes into safety.

How crazy is the quantity of AI models we’re seeing now? Which one is your favorite? I’m hanging onto my ChatGPT Plus account and hoping for GPT-5. But Claude Pro is looking very tempting.

Let us know which article stood out for you and keep sending us links to any juicy AI stories we could have missed.

The post DAI#34 – Data grabs, hot chips, and plus-size models appeared first on DailyAI.


This article was originally published at dailyai.com