Hacker Newsnew | past | comments | ask | show | jobs | submit | pyridines's commentslogin

In my experience, an agent with "fresh eyes", i.e., without the context of being told what to write and writing it, does have a different perspective and is able to be more critical. Chatbots tend to take the entire previous conversational history as a sort of canonical truth, so removing it seems to get rid of any bias the agent has towards the decisions that were made while writing the code.

I know I'm psychologizing the agent. I can't explain it in a different way.


I think of it as they are additive biased. ie "dont think about the pink elephant ". Not only does this not help llms avoid pink elphants instead it guarantees that pink elephant information is now being considered in its inference when it was not before.

I fear thinking about problem solving in this manner to make llms work is damaging to critical thinking skills.


Fresh eyes, some contexts and another LLM.

The problem is information fatigue from all the agents+code itself.


It is ridiculously more expensive and complicated under the hood, technically, but to the user, the sheer convenience of being able to text the computer "hey, when I get an email like X, inform Y and do Z" and that's it, you're done, is unmatched.

What about the convenience of having your whole inbox deleted?

https://www.pcmag.com/news/meta-security-researchers-opencla...

Maybe OpenClaw was just practicing a really aggressive form of Inbox Zero.


So ... don't give it write access to your email?

As I said elsewhere, complaining about this is like complaining that rm can let you delete your hard drive.

It's a tool. Learn how to use it.


Ignoring that you've just cut off a whole vector of usefulness, how do I keep it from exfilling my inbox to the Internet in response to a malicious email? Or using its access to take control of my online accounts?

Honest question, this kind of stuff is what keeps me from using it.


Don't give it access to your email then. I haven't. Plenty of other uses for it!

Use this software, it's amazing, it will change your life!"

"Oh but don't use it for A, or B, or C (even though it says to use it for A, B, and C): it will ruin your life"


Yes and yes!

A spouse can be amazing, or can destroy your life. Would you use that as an argument against marriage?


"Marrying? But what about cheating?" "Easy! Just murder your spouse before that ever happens!"

Like what?

I'm not using OpenClaw specifically here, but I have an agentic-ish AI I've built myself (considering that these things are generally just a while loop that monitors things & awakens if necessary, or a cron-job that runs a specific prompt).

One potential use - my Claude (Opus 4.6) has access to my to-do list, including for my business / software development. Claude awakens while I'm asleep, to go through the to-do list and look for things it can do proactively to help, or make suggestions about the business. An example from this morning: it saw that I'd been taking a long time last night creating icons in Affinity Designer for an Android app using its exporter. When I woke up, I saw Claude had written a CLI image resizer program for me that would take a PNG file and resize it specifically to all of the necessary sizes with the necessary filenames and folder structure for Android. It then offered to make an MCP version so it could do the resizing itself in future (though it could have used the CLI too if I'd granted approval).

This wasn't something I'd asked for, or prompted it to do. I didn't tell it to code this, or how to code it. Claude just thought this was the best way it could help me right now, and save me the most time. And it did it while I was asleep.

On another day, I woke up and it had made another Go program to track a regression test matrix, where it had plotted out all the platforms the program I'm making runs on and the various tests that need to be performed to check that it's ready to ship, with a little interactive program to mark each test as pass/fail/skipped. That helps me get through the manual tests faster - but it also saves the data into a format that Claude can read, to check on the test status while I'm asleep and make further recommendations.

I don't think many people have figured out yet that you don't even need to prompt AI. Treat it well, treat it with respect, give it the opportunity and ability to do things, and there is a lot that will emerge. But if you treat AI like a tool, it performs about as well as if you treat your employees like tools.


It's all tradeoffs and threat models.

You can prevent yourself from getting spam by not having an email account. But it's the nuclear option.

I'm fine with a system that can just read mail - and I already built one of those. I personally never send emails anyway so it's not an issue for me.


So what do you do with your OpenClaw instance that has read-only access to your email and no Internet access?

Did I say my Claw (not Open) doesn't have internet access?

All of my tools are geared towards reducing noise and condensing information.

- My weather scripts tell me just the exact metrics I care about

- My email filtering system surfaces only the mails that are relevant right now - I can check the rest later.

- My RSS feed hydrator pre-filters Hacker News and other RSS feeds and adds data like comment/vote count etc to the feed itself so I can determine whether the link is worth opening just based on the information presented

None of these require an LLM to have free rein to modify things for me.


> Did I say my Claw (not Open) doesn't have internet access?

> My email filtering system surfaces only the mails that are relevant right now - I can check the rest later.

So then you have not actually addressed the concerns expressed in my post. You indeed have an LLM with both email access and Internet access. Exactly the scenario I described. The amount of trouble those two accesses together can cause is huge.

> None of these require an LLM to have free rein to modify things for me.

Give me read access to your email and an Internet connection and I bet I can come up with all sorts of ways to modify things for you. So can an LLM. If your lucky it won't.


> As I said elsewhere, complaining about this is like complaining that rm can let you delete your hard drive.

rm won't wipe my HDD on a whim whilst instructing it to do something totally different.

You pretending they are the same thing is disingenous.


Bad take.

You can rm -rf your entire hard drive, but you can't blame rm for it, it's you who did it, maybe because you don't know, or a mistake, doesn't matter.

When you ask the clanker to delete x number of files in a directory, it can reason itself that is easier to just get rid of the directory.

Can't expect deterministic outcomes out of a statistical model.

At it's current state its a wildcard, sure you can build guard rails, reduce permissions, but it's still a wildcard.

Let's not kid ourselves saying is just a skill issue.


> When you ask the clanker to delete x number of files in a directory, it can reason itself that is easier to just get rid of the directory.

Oh sure, so don't give it write access to anything important. And make backups.

Mine is on a VM. It doesn't have access to my host's files. The worst it will do is delete the files on the VM. No great loss.

Yes, I do get it to modify things on my host, but only via a REST API I've set up on my host, and I whitelist the things it can do (no generic delete, for example). I even let it send emails. But only to me. It can't send an email to anyone else.


> So ... don't give it write access to your email?

> (…)

> Oh sure, so don't give it write access to anything important. And make backups.

If this conversation continues much longer, we’ll end up with “don’t use it at all”.

If I can’t trust a piece of software with anything important, why am I wasting my time fiddling with it? Might as well go play a video game or go do literally anything else entertaining.


> If I can’t trust a piece of software with anything important

Not what I said. As I've repeatedly said in this thread: Plenty of use cases where you don't give it access to email and write access to files. The comment you're replying to has an example of that.

> Might as well go play a video game or go do literally anything else entertaining.

True of most hobbies, right? I knew people who 20 years ago used to spend time in their garage building solar powered vehicles. But if I can't trust it to be reliable and safe on the road, I might as well go play a video game.

Also: Is anyone telling you to use it?


> True of most hobbies, right?

If everyone treated OpenClaw as a hobby, you might have a point, but people are using it for work in ways which will affect millions of other people when they’re hacked or the agent fucks up something important.

You already know how Meta’s AI Safety Director borked her email. Here’s the corporate vice president of Microsoft Word asking to be pwned:

https://www.omarknows.ai/p/meet-lobster-my-personal-ai-assis...

> Also: Is anyone telling you to use it?

You don’t need to use the technology to be affected by it. Ask Scott Shambaugh:

https://theshamblog.com/an-ai-agent-published-a-hit-piece-on...


> but people are using it for work in ways which will affect millions of other people when they’re hacked or the agent fucks up something important.

People will always do stupid things. My guess is less than 10% (perhaps even less than 1%) are using it for work. Most workplaces wouldn't allow unfettered AI usage.

80-90% try it, find it unreliable and buggy, and give up on it.

Of the remaining ones, likely 90+% are not using it in (very) dangerous ways.

People like me using it for boring things aren't making the news, and aren't writing blog posts about "Look at the cool stuff I've done!" because getting OpenClaw to notify me of class openings is not worth writing about.

In my (large) company, we have a Slack channel for OpenClaw. Over 400 people are in that channel. Let's assume 10% are using it (at home). No one's lost files/emails or any other damage.

If you're old enough, you'll remember sentiments in the 80's and 90's where "Oh, you let your teen get a modem? He must be hacking/phreaking."

Or "Oh, he's using Linux? He must be using it to become a hacker."[1]

Most of the complaints I see on HN are from people who know little about it, and are going off negative press/posts. Just as people knew little about modems and Linux. I mean, having to tell people "Don't give it access to your emails" is a clear sign of their ignorance. Kind of like having to tell someone "OK, just don't give your 10 year old the car keys" when they complain that cars are inherently dangerous because 10 year olds can kill themselves driving it.

It's worth trying it in a secure environment so at least one can make an informed critique.

Like you, I steered clear of OpenClaw, seeing all the problems and all the money people were burning on tokens. But at some point, I decided I should at least try it in a safe way before rendering judgment. And now I see what it is. Has it done so much for me that I'd throw a lot of money at it? Heck no. Not yet at least. But I do see we're past the point of no return. OpenClaw itself may die, but some derivative of it is going to be transformational.

As I said: Make it secure, affordable, reliable and user friendly, and many App/SaaS services will disappear.

> You don’t need to use the technology to be affected by it. Ask Scott Shambaugh:

> https://theshamblog.com/an-ai-agent-published-a-hit-piece-on...

I don't know how old you are, but once everyone had a camera in their phones, the cat was out of the bag. Lots of people complaining about their photos showing up online because someone had taken a picture of them. Yes, this is bad. Yes, lives were lost (bullying, etc). And no, phones with cameras weren't going to go away. And everyone who complained has one now.

And as I pointed out a few days ago[2], the whole Scott Shambaugh episode was pretty mild compared to what some open source maintainers have had to deal with when it comes to humans.

[1] Lots of cases where ISPs, etc kicked customers out because they were using Linux and they didn't want the ISP to be implicated in criminal activities. "Only criminals use Linux"

[2] https://news.ycombinator.com/item?id=47083686


Worth the risk.

[flagged]


OpenClaw is rightly being blamed for a mistake it made. Any argument regarding her aptitude would be irrelevant as it would in no way absolve OpenClaw.

Sure, that’s an interface that’s better for many users and use-cases.

However, it seems better if you could, as much as is possible, move the AI stuff from runtime to “compile time.”

Instead of having the AI do everything all the time, have AI configure your Zapier (or whatever) on your behalf. That way you can (ideally) get the best of both worlds: the reliability and predictability of classical software, combined with the fuzzy interface of LLMs.


> Instead of having the AI do everything all the time, have AI configure your Zapier (or whatever) on your behalf.

That is what many use OpenClaw for! The AI assistant will happily recommend existing services and help you (or itself, if you let it), set it up.

(In theory. In practice, it often does a poor job).

The appeal of OpenClaw is I don't need to go research all these possible solutions for different problems. I just tell it my problem and it figures it out.

Yesterday I told it to monitor a page which lists classes offered, and have it ping me if any class with a begin date in March/April is listed. This is easily scriptable by me, but I don't want to spend time writing that script. And modifying it for each site I want to be notified for. I merely spoke (voice, not text) to the agent and it will check each day.

(Again, it's not that reliable. I'm under no illusion it will inform me - but this is the appeal).


But literally any decent agent can recommend existing services and help you set them up. And even help you help them set the services up for you. I do this with Claude all the time.

That's still too much work. Someone would have to make like an OpenClaw wizard that protectively offers to set all that stuff up. So the potential OpenClaw user can then, on running for the first time, be guided through the setup of whatever they'd like to get connected. And "setup" here means a short description of X and a "Connect? (y/n)" prompt. Anything more and you start losing people.

yes. in a similar vein, we're seeing that get standardized in coding agents as "don't have the agent use tools directly, have the agent write code to call the tools"

Sometimes I reflect on all the metaphorical forests that have burned because a certain person at the right time only knew so much about how to use Excel, or the inbox rules of their MUA, or being totally unaware of the incredible power of macros of all sorts.

Like if you could just sit someone down for 30 minutes and show a few "power user" things, you will have truly taught her to fish for a lifetime. But it can go so unaddressed, and people's careers are built on these small ignorances.

I've cancelled everything at this point and just call Emacs my "special agential assistant," it makes me still sound in-the-know, and most of the time no one knows the difference!

"Convenience" in this context is laziness; "productivity" and "efficiency" is for management and bosses. We don't need to be our own bosses, I want to be free from such things as an individual. I want to be capable, be maybe almost "cool." Its sad to see a whole generation turn into such product dorks!

"Oh please read my email for me Mr. AI!"


Some nitpicking.

> Modern JavaScript applications aren't just a few scripts anymore — they're sprawling codebases with thousands of dependencies, complex module graphs, and extensive build pipelines.

This has been true for at least a decade.

The very next paragraph:

> JavaScript-based tools that were once "good enough" now struggle to keep up, leading to sluggish build times, laggy editor experiences, and frustratingly slow feedback loops.

The tools really weren't "good enough", even back then, by these metrics. JavaScript tooling has been slow and bloated on large codebases for just as long.


Just as the modern Javascript applications. What if - and hear me out on this one - Javascript just is a poor choice for huge complex codebases?


What else should you use for huge complex web apps?


Keep the huge, complex business logic on the server whenever possible.

That doesn't work for webapps that are effectively entirely based on client side reactivity like Figma, though the list of projects that need to work like that is extremely low. Even for those style of apps I do wonder how far something like Phoenix LiveView might go towards the end goal.


maybe, just maybe, the browser is not always the best tool for the job


I think that there are more apps that are better off as web apps (cross platform and sandboxed) than not.


But I hired the whole react dev, so I’ll use the whole react dev!

/s


<3 if I don't see 15 new node modules and 3 CVEs by EOB today I'll replace you with a css architect and vibe-coding nft monkey by next week!


Animal intelligence is often underestimated, (e.g. there's a famous test that shows that chimpanzee working memory is better than ours) but our use of language is qualitatively different from other animals. Some animals have rudimentary communication, but no other animal is capable (as far as we know) of recursive, infinitely variable language structure like us.


> spaghettification is also a non-issue, as long as the framework defines clear containers for spaghettis

Sorry, but I disagree strongly with this. When there is, inevitably, a bug that the LLM can't fix, someone's going to have to read all that spaghetti, and they'll curse whoever put it there. "clear containers for spaghetti" is a pipe dream, all abstractions leak, and the bug may very well be in the spaghetti. "Just start over" is unrealistic for large, complex apps.

Of course, if you really have a solution for this, that would be incredible.


nothing lives forever. software comes into life out of necessity, develops complexities, eventually becomes incomprehensible, obsolete and dead. it’s a natural cycle that we should work into the user experience, instead of defining it as a failure state that we need to “solve for”.


they're referring to the infamous Corporate Memphis style which is frequently used in big tech branding: https://en.wikipedia.org/wiki/Corporate_Memphis


Wow. People can get angry at anything.


... and which is also reviled by many, making it an odd point to bring up.


It's not unreasonable to briefly forget details like that, especially when you're dealing with a multi-language codebase where "how do I make a log statement?" requires a different pattern in each one.


> It's not unreasonable to briefly forget details like that, especially when you're dealing with a multi-language codebase where "how do I make a log statement?" requires a different pattern in each one.

You make my point for me.

When I wrote:

  ... I love working with people who understand what they
  are doing when they do it.
This is not a judgement about coworker ability, skill, or integrity. It is instead a desire to work with people who ensure they have a reasonable understanding of what they are about to introduce into a system. This includes coworkers who reach out to team members in order achieve said understanding.


Another American holiday coming up with an equally useless name is Fourth of July. Nobody seems to have a problem with that name, and nobody I know calls it Independence Day. Neither Fourth of July or Juneteenth are great names out of context, but they both have histories behind them and can't be changed anymore.

Heck, Juneteenth is a better name, since it is not literally month+day.


I was dumbfounded too, but NBC explains in this same article:

> "The law banning TikTok [...] allows the president to grant a 90-day extension before the ban is enforced, provided certain criteria are met."


Those criteria have not been met and we are passed the deadline in which that extension could be applied.


I agree - if I was a student, I would be very tempted to use an LLM for this sort of coursework-irrelevant busywork assignment, especially if I had other work on my plate. It's not so hard to rationalize using it for this type of thing vs. an "actual" assignment, all due respect to the professor, who I'm sure means well


As an instructor, I suppose I'm happy to tailor the class to the AI's wishes. ;)


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: