Google, OpenAI and Mistral released latest versions of their cutting-edge AI models in only 12 hours.

Meta can even be in the combination with its upcoming Llama-3 model, and OpenAI’s highly anticipated GPT-5 is within the pipeline.

What began as a purely area of interest category dominated by ChatGPT is now flooded with alternatives that transcend Big Tech and the open- and closed-source divide.

Google Gemini Pro 1.5

Google’s Gemini Pro 1.5 caused a stir and led to advances in long-context understanding that challenge Claude 3 Opus, which leads the best way on this category.

With the flexibility to process as much as 1 million tokens, Gemini Pro 1.5 can process large amounts of data concurrently, including 700,000 words, an hour of video, or 11 hours of audio.

Its Mixture of Experts (MoE) architecture increases efficiency and performance through the use of specialized models for specific tasks.

Google’s list of Gemini models is sort of complex, but essentially the most powerful model for typical tasks is ranked here.

New version of GPT-4 Turbo

OpenAI then released a new edition, GPT-4 Turbo, with superior math and image processing.

An X post states: “GPT-4 Turbo with Vision is now generally available within the API. Vision requests can now also use JSON mode and performance calls.”

OpenAI expects to release GPT -5 soon, in addition to its text-to-video model Sora, which currently has no significant competitors (although that can change).

Mixtral 8x22B

Perhaps the largest surprise, nevertheless, got here from mistralwho boldly released their Mixtral 8x22B model as a free downloadable 281GB file via torrent.

With a powerful 176 billion parameters and a context length of 65,000 tokens, this open source model based on the Apache 2.0 license is claimed to outperform Mistral’s predecessor Mixtral 8x7B, which had already outperformed competitors akin to Llama 2 70B in various benchmarks.

Mixtral 8x22B’s advanced MoE architecture enables efficient calculations and improved performance in comparison with previous iterations.

Meta Llama 3 is coming

Not to be left behind, reports suggest that Meta could release a small version of its highly anticipated Llama 3 model as early as next week, with the total open source model still scheduled for July.

Llama 3 is anticipated to be available in a wide range of sizes, from very small models that compete with Claude Haiku or Gemini Nano to larger, fully responsive and reasoning models that compete with GPT-4 or Claude 3 Opus.

Model multiplication

A generative AI ecosystem once dominated by ChatGPT is now flooded with alternatives.

Virtually every major technology company is involved, either directly or through significant investments. And the more players join the fight, the more hope that one faction can dominate the market dwindles.

We at the moment are also seeing the gap closing between closed source models from OpenAI, Anthropic, Google, etc. and closed source alternatives from Mistral, Meta and others.

Open source models are still hardly accessible to the overall population, but that can probably change too.

Do any of those models represent an actual advance in machine learning, or are they simply more of the identical but higher? It is dependent upon who you ask.

Some, like Elon Musk, predict that AI will surpass human intelligence inside a yr.

Others, like Meta’s chief scientist Yann LeCun argue that AI is miles behind us by way of robust intelligence measurements.

LeCun declared in February about current LLMs: “So principally they’ll’t invent latest things. They will reflect roughly every little thing they were trained on from public data, meaning you may get it from Google. People have said, ‘Oh my God, we’d like to control LLMs because they’re going to be so dangerous.’ That’s just not true.”

Meta goals to create an “object-driven” AI that higher understands the world and attempts to plan and reason about it.

“We are working hard to determine methods to make these models not only talk, but additionally think and plan.” . . have memory,” explained Joelle Pineauthe vice chairman of AI research at Meta.

OpenAI Chief Operating Officer Brad Lightcap also said his company is concentrated on improving AI’s ability to grasp and handle more complex tasks.

“We will begin to see AI that may tackle more complex tasks in additional sophisticated ways,” he said at a recent event. “I feel over time we’ll see the models turn out to be longer, in some way more complex tasks, and that implicitly requires improving their considering skills.”

As 2024 approaches summer, the AI ​​community and society at large might be watching closely to see what groundbreaking developments emerge from the laboratories of those tech giants.

By the tip of the yr there might be quite a various selection.


This article was originally published at dailyai.com