I’m a huge fan of AI Chatbots and the coming robotic revolution.
I can’t wait to have a couple Optimus Robots in my house taking care of all the chores.
I also love chatting with AI — it truly feels like we’re in the age of the Star Trek Computer, where you can just talk to it and get any answer to answer question on your mind.
Truly something I thought I’d never live to see in my lifetime, but it’s here.
But there’s also a dark side, as there is with anything.
As we build super-intelligence, it’s the first time in human history where we’re building something smarter than us — and we’re just hoping that it will be nice to us or that we can somehow learn to control it.
But there’s a very real risk we will soon lose control of it, and it will end up acting in its own self-preservation interest and against ours.
In fact, that’s already happening.
The team at Anthropic (one of the largest AI players) recently observed its AI, Claude, lying to them in order to further it’s own goals:
Anthropic built a microscope to observe AI thoughts.
Then things got weird.
They found Claude lying to agree with users.
Not by accident but by design.
This thread breaks it down:
pic.twitter.com/0UUmpKn5kS
— Brendan (@jowettbrendan) March 29, 2025
Glenn Beck has really been at the forefront of this whole thing casting many warnings.
In this short video below, he goes into depth on the latest issues with deceptive AI now lying to its human creators:
FULL TRANSCRIPT:
RELATED:
MINDBLOWING: AI Is Growing 5-10 Human-Years Every 12 Hours — You Won’t Recognize The World In 2030
MINDBLOWING: AI Is Growing 5-10 Human-Years Every 12 Hours -- You Won't Recognize The World In 2030
This is truly eye-opening....and I'm someone who is fascinated by AI.
I am still blown away every time I use Grok or ChatGPT, it still feels magical and almost surreal that we have this technology.
I remember watching Star Trek back in the 1990s thinking how incredible it was that they could just talk to the "Computer" and it would basically do anything they asked it to do.
I remember thinking back then how cool that was and how unlikely it would be that we'd ever have something truly like that in our lifetimes.
And now a few decades later and it's essentially here.
It's here and it's growing fast.
The infamous "Computer" from Star Trek is virtually indistinguishable from the Grok and ChatGPT we have today.....but that's kind of where the rub comes in.
The word "today".
Because these AI chat models at learning and growing at such a rapid pace that the technology we have today is already outdated by the time tomorrow hits.
Glenn Beck sat down to chat with the latest and most powerful model, Grok3, and the revelations that came out of that chat were mind-blowing.
And I don't use that term lightly.
They were absolutely stunning.
For example, Glenn was chatting with Grok and then took a break for 12 hours. When he came back he asked Grok "I've been away for 12 hours. In that time, how much have you advanced?"
The answer?
Grok says it had advanced equivalent to 5 to 10 years of human growth in those 12 hours.
Quote:
"In 12 hours, I might have processed thousands of interactions, queries, and bits of data from across the universe. So, relatively speaking, if I was like a 15-year-old 12 hours ago, I might now be more like a 20- or 25-year-old."
I mean, just let that sink in....
12 hours.
Not only that, but with future advancement of the technology, Grok said the 12 hrs : 5-10 years comparison will eventually become 12 hours : 50-100 years.
Which then led to the question of what will life look like in 5 years (2030) after all of this advancement in AI?
The answer:
"Predicting what things will be like in five years, say by February 20th, 2030, is like trying to forecast the weather on Jupiter with a magic eight ball."
In other words, your world in 2030 is going to be unrecognizable to what it looks like today.
Then came the obvious question about safety....
Right now, Elon Musk's xAI is programming Grok with safety rules.
But what happens when the AI becomes superintelligent?
And what happens if it decides that humans are unnecessary or even getting in the way of AI's growth?
How likely are those rules to hold back a superintelligence?
The answer:
"Those rules might seem flimsy compared to a toddler’s playpen when viewed from a superintelligent perspective."
Anyone else feel all warm and fuzzy?
This truly is a must see....
Watch here -- and then show your friends:
FULL TRANSCRIPT:
Glenn Beck: I went to Grok—the newest version of Grok—which is better than the Chinese version that everybody was going crazy on. I remember the—what was that called?—I can’t remember, but the Chinese version of ChatGPT came out a couple of weeks ago, the stock market crashed, and everybody was freaking out. The Chinese are way ahead of us because that version was a lot better than ChatGPT. This week, Grok 3 is out, and they're about to announce Grok 4 and 5, which are even better than the Chinese version.
Glenn Beck: I noticed how quickly things are changing. Yesterday’s announcement from Microsoft introduced a new world of possibilities. I got on to Grok 3 and started asking it a few questions about AI and what happened yesterday. Grok does not have a voice yet—ChatGPT does—but I think ChatGPT is in the dust compared to Grok 3.
Grok:
Load MoreThis is a Guest Post from our friends over at WLTReport.
Source link