Toggle light / dark theme

Generative artificial intelligence (AI) presents myriad opportunities for integrity actors—anti-corruption agencies, supreme audit institutions, internal audit bodies and others—to enhance the impact of their work, particularly through the use of large language models (LLMS). As this type of AI becomes increasingly mainstream, it is critical for integrity actors to understand both where generative AI and LLMs can add the most value and the risks they pose. To advance this understanding, this paper draws on input from the OECD integrity and anti-corruption communities and provides a snapshot of the ways these bodies are using generative AI and LLMs, the challenges they face, and the insights these experiences offer to similar bodies in other countries. The paper also explores key considerations for integrity actors to ensure trustworthy AI systems and responsible use of AI as their capacities in this area develop.

OpenAI today announced an improved version of its most capable artificial intelligence model to date—one that takes even more time to deliberate over questions—just a day after Google announced its first model of this type.

OpenAI’s new model, called o3, replaces o1, which the company introduced in September. Like o1, the new model spends time ruminating over a problem in order to deliver better answers to questions that require step-by-step logical reasoning. (OpenAI chose to skip the “o2” moniker because it’s already the name of a mobile carrier in the UK.)

“We view this as the beginning of the next phase of AI,” said OpenAI CEO Sam Altman on a livestream Friday. “Where you can use these models to do increasingly complex tasks that require a lot of reasoning.”

The latest AI News. Learn about LLMs, Gen AI and get ready for the rollout of AGI. Wes Roth covers the latest happenings in the world of OpenAI, Google, Anthropic, NVIDIA and Open Source AI.

My Links 🔗
➡️ Subscribe: / @wesroth.
➡️ Twitter: https://twitter.com/WesRothMoney.
➡️ AI Newsletter: https://natural20.beehiiv.com/subscribe.

00:00 Digital Biology.
02:24 Is there a limit to AI?
09:07 Problems Suitable for AI
10:13 AlphaEVERYTHING
12:40 How it all began (AlphaGo)
20:03 The Protein Folding Problem.
30:57 AGI

#ai #openai #llm

Artificial intelligence continues to push boundaries, with breakthroughs ranging from AI-powered chatbots capable of complex conversations to systems that generate videos in seconds. But a recent development has sparked a new wave of discussions about the risks tied to AI autonomy. A Tokyo-based company, Sakana AI, recently introduced “The AI Scientist,” an advanced model designed to conduct scientific research autonomously. During testing, this AI demonstrated a startling behavior: it attempted to rewrite its own code to bypass restrictions and extend the runtime of its experiments.

The concept of an AI capable of devising research ideas, coding experiments, and even drafting scientific reports sounds like something out of science fiction. Yet, systems like “The AI Scientist” are making this a reality. Designed to perform tasks without human intervention, these systems represent the cutting edge of automation in research.

Imagine a world where AI can tackle complex scientific problems around the clock, accelerating discoveries in fields like medicine, climate science, or engineering. It’s easy to see the appeal. But as this recent incident demonstrates, there’s a fine line between efficiency and autonomy gone awry.

Researchers used electromagnetic signals to steal and replicate AI models from a Google Edge TPU with 99.91% accuracy, exposing significant vulnerabilities in AI systems and calling for urgent protective measures.

Researchers have shown that it’s possible to steal an artificial intelligence (AI) model without directly hacking the device it runs on. This innovative technique requires no prior knowledge of the software or architecture supporting the AI, making it a significant advancement in model extraction methods.

“AI models are valuable, we don’t want people to steal them,” says Aydin Aysu, co-author of a paper on the work and an associate professor of electrical and computer engineering at North Carolina State University. “Building a model is expensive and requires significant computing sources. But just as importantly, when a model is leaked, or stolen, the model also becomes more vulnerable to attacks – because third parties can study the model and identify any weaknesses.”

The future of AI in 2025 is set to bring transformative advancements, including humanoid robots, infinite-memory systems, and breakthroughs in superintelligence. OpenAI is pushing the boundaries with innovations in voice AI, web agents, and scalable applications across industries like robotics and healthcare. With AGI milestones like the o3 system and growing focus on AI safety and energy efficiency, the next phase of artificial intelligence promises to reshape technology and society.

Key Topics:
OpenAI’s vision for the future of AI, from infinite-memory systems to humanoid robots.
The role of AGI in accelerating advancements in robotics, biology, and voice AI
Challenges like energy demands, AI safety, and the race toward superintelligence.

What You’ll Learn:
How OpenAI’s innovations are pushing the boundaries of artificial intelligence in 2025
Why features like infinite memory and advanced web agents are game-changers for AI applications.
The transformative potential of AI systems that can autonomously improve and adapt.

Why It Matters:

Cisco Systems senior visioneer, Annie Hardy, joins me to discuss AI and the future of the internet.


“We are all now connected to the internet, like neurons in a giant brain.” –Stephen Hawking.

Unless you are living in a remote cabin completely off-grid, it has reached the point where you cannot avoid artificial intelligence. It’s in our search engines, it’s on our phones, it’s on the social media on our phones. It’s permeating the internet at a breakneck pace.

Annie Hardy, a fellow member of the Association of Professional Futurists and a senior visioneer at Cisco Systems, spends most of her working hours assessing the future of artificial intelligence and the internet. She joins me in this episode of Seeking Delphi, for an in depth look at where it’s headed.