🧮Google's AI Agent Solves a 50-Year Math Puzzle

When machines start writing better machines, what happens next?

Welcome, AI Visionaries!

Google unleashed an AI that evolves its own code, breaking a half-century-old math wall. TikTok turned still photos into uncanny living moments, testing the limits of reality and creativity. Meanwhile, Anthropic’s Claude got a serious upgrade to rethink and fix itself like a human coder. And OpenAI just threw open its safety reports, daring the world to judge its trustworthiness.

Have you noticed that AI has stopped playing catch-up and started rewriting the rules?

In today’s Generative AI Newsletter:

• Google breaks new ground with AlphaEvolve
• TikTok brings photos to life
• Anthropic makes Claude smarter and more independent
• OpenAI shows its safety cards

🧬 Google’s AlphaEvolve Just Solved a 50-Year-Old Math Problem

Image Credit: Google

Google has unleashed a new class of AI agent: AlphaEvolve, a coding system that combines Gemini models with evolutionary strategies to generate and refine algorithms.

Key Developments:

  • AlphaEvolve combines multiple Gemini models, using Gemini Flash for fast ideation and Gemini Pro for critical analysis. Each code candidate is tested, evaluated, and evolved.

  • It cracked Strassen’s algorithm, improving a classic 1969 matrix multiplication method for the first time in over 50 years.

  • It outperformed state-of-the-art math solvers, matching top solutions on 75% of 50+ open math problems, and producing entirely new, superior answers in 20%.

  • Inside Google, it’s optimizing data center schedules, accelerating model training (including for Gemini itself), and contributing to chip design workflows.

  • It builds code through an evolutionary loop, generating algorithm candidates, evaluating them through execution, and mutating high-potential solutions for further improvement.

AlphaEvolve shows that when LLMs are wired into evolutionary systems, AI starts producing novel scientific knowledge. The future of research might depend less on individual insight and more on algorithmic exploration.

🧟 TikTok’s ā€˜AI Alive’ Brings Photos to Life — Magical or Just Plain Creepy?

Image Credit: Tiktok

Your static photo can now blink, breathe, and move on its own. TikTok has launched AI Alive, a new feature that transforms still images into animated clips with just a text prompt. The results range from surprisingly lifelike to deeply unsettling, and people are already split between calling it art or nightmare fuel.

What’s New:

  • AI Alive animates photos: using simple prompts like ā€œmake eyes blinkā€ or ā€œadd moving cloudsā€ to generate short videos from still images.

  • Only real-time photos work: users must shoot pictures through the Story Camera. Uploading from the gallery is blocked to prevent misuse.

  • Multi-layered moderation: scans photos, prompts, and final videos before anything goes public. All content is labeled as AI-generated and disappears after 24 hours unless saved.

  • Results vary wildly: Basic prompts like subtle movement look realistic. But complex requests often trigger bizarre or creepy distortions, especially on faces.

  • Creative potential is huge: Artists are animating portfolios, small businesses are spicing up product shots, and users are already swapping prompt formulas to get the best results.

TikTok’s rollout of AI Alive marks a new chapter in consumer AI video tools. It leapfrogs filters and effects by letting users animate reality itself. As faces come to life with a few words, the line between creativity and creepiness is starting to blur and TikTok seems fine dancing on that edge.

🧠 Claude Is About to Get Way Smarter

Image Credit: Anthropic

Anthropic is preparing to launch upgraded Claude models in the coming weeks, including new versions of Sonnet and Opus. These models promise deeper reasoning, smarter tool use, and the ability to fix their own mistakes without human help.

Details:

  • Models can switch between reasoning and tool use, allowing them to step back, reassess, and self-correct.

  • For coding tasks, they test outputs, identify bugs, reason through fixes, and rewrite code autonomously.

  • A new model, codenamed Neptune, is in safety testing. Some speculate the name hints at a Claude 3.8 release.

  • Anthropic has also launched a bug bounty program aimed at probing Claude’s adherence to safety principles.

Anthropic has lagged behind in model releases this year, but this drop could reset the board. If these systems can truly revise and recover like human thinkers, we’re not just watching AI automate tasks, we’re also watching it learn how to invent.

šŸ” OpenAI Opens the Books on Model Safety

Image Credit: OpenAI

OpenAI has launched a public Safety Evaluations Hub to show how its models handle dangerous prompts, factual accuracy, and adversarial testing. The dashboard offers a clearer view into model behavior, following rising pressure for transparency.

Key Details:

  • Metrics span four key areas: harmful content refusal, jailbreak resistance, hallucination rates, and instruction-following behavior.

  • The hub provides comparative data across OpenAI models, including their ability to reject unsafe queries and stay grounded in truth.

  • OpenAI says the dashboard will be updated regularly, aligning with a broader push to communicate more openly about safety.

  • The move follows recent criticism around a GPT 4o release and broader concerns that OpenAI’s safety practices are opaque.

As model capabilities race ahead, trust depends on knowing how these systems behave under pressure. This hub is a meaningful transparency step, but the real test will be whether OpenAI stays consistent, thorough, and honest in what it shares.

šŸš€ Boost your business with us—advertise where 10M+ AI leaders engage

🌟 Sign up for the first AI Hub in the world.

šŸ“² Our Socials

Reply

or to participate.