Agents Aren't Coworkers, Embed Them in Your Software

(feldera.com)

37 points | by gz09 2 hours ago

8 comments

  • solid_fuel 4 minutes ago
    "Agents" can't think and LLMs aren't sentient. They aren't suited to be your coworker, but they also aren't suited for generation computational tasks. The chat interface is all that there is and their behavior in chat is not deterministic or bounded enough to be useful in most applications. They mimic tokens in reply to the tokens you give them, and that is all.

    You know what's a bad idea from an engineering (that thinky thing we used to do as part of building software) perspective?

    Building a dependency on an expensive remote API into your system.

    This isn't just me bloviating, I've been down this road before. In my case I had a project using LLMs to automatically edit videos provided by Hollywood content owners. It seemed like a decent application, but LLMs are structurally unsuited for dealing with user data like this. The way that the prompt is evaluated means there is no separation between system and user input, so once you start dealing with a wide variety of topics you pretty quickly run into walls.

    One example - ChatGPT refusing to summarize and pick a top segment from a news program because it contained references to a murder-suicide, and both murder and suicide are included in the many prohibited topics that are filtered in ChatGPT replies. This was through their API, not the regular user interface, so it is in theory as unrestricted as access gets. But because the LLM cannot be trusted to behave properly around the topic, they have to filter anything which touches it.

    Structurally, I don't see a way this can be overcome - LLMs by design mix the entire prompt together, it's not like a parameterized SQL query where you can isolate the user and system data. That means that a long or bold enough user input is often enough to outweigh the system prompt, and that causes the LLM to veer into unpredictable territory.

  • iot_devs 1 hour ago
    > Give an agent the right interfaces and it becomes less conversational and more ambient. It no longer needs to constantly ask, explain, summarize, and negotiate. It can stay in the background, react to changes, and make steady progress with less supervision and less noise. That is closer to Weiser’s vision: calm technology, but for machines.

    I tend to agree quite a bit.

    I created a ambient background agent for my projects that does just that.

    It is there, in the background, constantly analysing my code and opening PRs to make it better.

    The hard part is finding a definition of "better" and for now it is whatever makes the longer and type checker happy.

    But overall it is a pleasure to use.

    • stingraycharles 1 hour ago
      Just take a look at the pull requests / issues opened of a repository that’s popular with LLM agents, to understand how well that works.

      If there’s one take away it’s that these agents need more, not less, oversight. I don’t agree at all with the “just remove a few tools and you can remove the human from the loop” approach. It just reduces the blast radius in case the agent gets it wrong, not the fact that it gets it wrong.

      • iot_devs 59 minutes ago
        Yeah, but my projects are personal and not popular.

        I crafted the AI loop to do exactly what I would be doing by manually.

        Out of 10 PRs, 6 to 7 gets merged. The other simply get closed.

        • stingraycharles 14 minutes ago
          Yeah my experience is that this works for a short time and then after a few weeks your codebase is a complete disaster.
  • ori_b 1 hour ago
    I'd pay more for deterministic, explainable, and fast software without agents. The value of computers is that they do tasks repeatably, reliably, and at blinding speed.

    This stuff is negative value.

    • efskap 51 minutes ago
      Right and modulo agents, they're just describing event-driven architecture like Lambda.
  • apsurd 49 minutes ago
    Ambient agents premise lands and is thought provoking.

    But the more you read the article the more the point is lost. The prescriptions given aren't ambient?

        CLI: a good command-line interface makes it easy for an agent loop to interact with your system and saves tokens.
        Specs: Declarative configs, schemas, manifests. Artifacts that state the desired outcome, not the steps.
        Reconciliation loops: you declare the target state, let the system continuously converge toward it. Detect if something drifts.
    
    (seems you're talking to the AI above (and you'll need to refine just like a conversation), it's just not synchronously in chat)

    The gripe seems to be specifically with being able to chat with the AI. Yes, ideally the AI just knows to do stuff. But the chat interface is also the reason every Bob and Sarah has chatGPT in their pocket. It's also just growing pains.

  • skybrian 38 minutes ago
    I like using them for coding, but I'm wary of making software that depends on an unreliable, expensive remote API. I'd rather have the agent write code and have no runtime dependency.

    It might be nice to have something simple and cheap for basic text classification, but I'm not sure what to use. (My websites are written in Deno.)

  • leobuskin 1 hour ago
    > Agentic management software is all the hype today: What started with Moltbot and OpenClaw now has a lot of competition: ZeroClaw, Hermes, AutoGPT etc.

    Moltbot is OpenClaw, AutoGPT was born significantly before. I just couldn’t read after the first paragraph, I’ve lost the trust entirely, whatever/whoever wrote it.

    • simonw 1 hour ago
      Hermes agent dates back to at least September last year too, pre-dating Moltbot/OpenClow by a couple of months https://github.com/NousResearch/hermes-agent/commit/17608c11...
    • stingraycharles 1 hour ago
      It’s marketing. They’re selling some change management solution, so obviously they advocate for showing AI agents only changes, rather than the full context.

      Doesn’t mean it’s a good idea, though.

  • orliesaurus 1 hour ago
    not yet coworkers*
    • claysmithr 50 minutes ago
      you wouldn't download a coworker
      • doubled112 30 minutes ago
        I would, along with a car for the coworker to drive me around in.
        • electroglyph 5 minutes ago
          but should you drive or walk to the car wash?
  • WhoffAgents 55 minutes ago
    [dead]