Get all your news in one place.
100's of premium titles.
One app.
Start reading
Tom’s Guide
Tom’s Guide
Technology
Amanda Caswell

I tried to use ChatGPT as a timer and it failed — here's why Sam Altman says it’s a year away from a fix

OpenAI logo with person in front.

It seems like there's never a "slow week" for AI. Big tech is constantly rolling out new features, chatbots regularly get updated and new models seem to pop up regularly. We’ve reached the point where AI is smart enough to pass the Bar Exam, write complex Python code and diagnose medical symptoms.

As a ChatGPT power user, I've tested the AI on several challenges meant to break it, yet, it still fails at a task so basic that a 1980s kitchen timer can do it — keep time. If you've ever tried to use ChatGPT as a legitimate personal assistant, you already know that we are still a ways out from the "Agentic AI" era.

The 'bake' test

A post shared by Husk (@husk.irl)

A photo posted by on

One of my favorite AI influencers/comedians has shown numerous times how ChatGPT cannot handle timing...anything. Sam Altman has tried to explain it, telling Gizmodo it will take a year to fix.

I urge you to try it because it's oddly hilarious. Simply open ChatGPT Voice and prompt anything that requires tracking time. For example: “I'm baking cookies. Start a 10-minute timer.”

When I did this ChatGPT didn't skip a beat. It responded with total, unearned confidence: "Absolutely! I've started a 10-minute timer for you. I'll let you know when it’s up."

I went back to work. Ten minutes passed. Then twenty. Then thirty. Nothing. No alarm or any kind of notification alerting me that time was up. The problem is, ChatGPT simply simulated the act of being helpful without actually doing anything.

Instead, it responded like it had set the timer, then convincingly assured me that it had timed ten minutes perfectly, essentially gaslighting me when I tried to question it.

This isn't about timers

A post shared by Husk (@husk.irl)

A photo posted by on

I recently wrote about Google’s AI Overviews confidently delivering answers — even though about 1 in 10 responses is wrong.

The problem here really isn't about timers. It’s about the gap between what AI says it can do and what it can actually execute. When I pushed further, asking it to track elapsed time or notify me when something was done, it acted like everything was working fine. As if it actually did what I asked it to do.

I chatted with a developer friend of mine about it last night. He suggested that maybe the problem with counting is in the Voice application itself. If you ask ChatGPT to "count" to 100 in text, it will write all the numbers.

But regardless of where the error lies, this is what makes AI feel deceptively capable. Becaues these aren't one off failures, the core problem is that ChatGPT can generate responses, but it doesn’t always have access to the tools needed to actually do things in the real world — like starting a timer or tracking time accurately.

So instead, it fills the gap with an answer that sounds right and this is where AI falls short because it can't always execute basic real-world actions. It can describe a timer, write a story about a timer and even probably vibe code a timer, but it cannot reliably be the timer.

Then, it doesn’t always tell you when it’s pretending to do something versus actually doing it.

Why this drives me crazy

(Image credit: Future)

Claude will tell you when it is wrong or unsure. It is not as people-pleasing as ChatGPT. I've even tested this with unhinged recipes. What is frustrating is the biggest lies of the current AI era:

  • The hallucination of action: ChatGPT is confidently a master of words but can't always execute and will make things up rather than admit it is wrong.
  • The confidence gap: Rather than say, "I can't." ChatGPT says, "I'm on it!"
  • The execution wall: Even Sam Altman has admitted that "reliability across long chains of logic" is the current bottleneck.

Bottom line

There’s no question that ChatGPT is an incredibly powerful tool. It can write, plan, explain and help you think through almost anything.

But what I want, and what most people probably expect, is honesty about its limits. If it can’t do something, it should say that clearly instead of sounding like it already did.



More from Tom's Guide

Sign up to read this article
Read news from 100's of titles, curated specifically for you.
Already a member? Sign in here
Related Stories
Top stories on inkl right now
One subscription that gives you access to news from hundreds of sites
Already a member? Sign in here
Our Picks
Fourteen days free
Download the app
One app. One membership.
100+ trusted global sources.