I am tired boss…

  • harmbugler@piefed.social
    link
    fedilink
    English
    arrow-up
    27
    arrow-down
    1
    ·
    24 hours ago

    Sounds like this is in the same space as OpenWebUI? It would be good to have some more choice there.

    • Rekall Incorporated@piefed.social
      link
      fedilink
      English
      arrow-up
      25
      arrow-down
      5
      ·
      18 hours ago

      At last a useful comment thread about the actual functionality in question.

      While I am not moving back to reddit, Threadiverse is just terrible on any nuanced conversations on modern ML tools and approaches.

      The tech is not the issue here. It has legitimate use cases and it is here to stay (this is not a blockchain pump and dump scheme ala Web 3.0). The issue here are American tech companies and broader support for crime/corruption in US society (as of today, doesn’t mean that this can’t/won’t change in 20-30 years). We need truly independent open source systems and tools.

      I am aware OpenWebUI is based in San Francisco and Mozilla Foundation is based in the US. I am always on the lookout for alternatives.

    • CyberSeeker@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      22
      arrow-down
      1
      ·
      edit-2
      22 hours ago

      It sounds like a step further than open-webui; it’s an enterprise grade client-server model for access to agents, workflows, and centralized knowledge repositories for RAG.

      In addition to local chatbot for executive/admin use, I can see this being the backend for developers running Cursor or some other AI enhanced IDE, with local knowledge stores holding proprietary documents and running against local large models.

      I am also curious about time share and prioritization of resources; I assume it would queue simultaneous requests. Presumably this would let you more effectively pool local compute, rather than providing A100 GPUs to each developer that may sit unused when they’re not working.

      Edit: Somewhat impressively, this whole stack does not even include a local inference provider; so it does everything except local models right now, and requests are forwarded to cloud inference providers (Anthropic, OpenAI, etc). But it does have the backend started for rate limiting and queuing, and true “fully offline/local” is on the roadmap, just not there yet.

      • ikidd@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        2 hours ago

        Where are you seeing that it’s useable for RAG? I’ve gone through the github and not seeing anyting very specific that way.

        Edit: good lord their documentation is shit. Spend a few tokens on a proper mkdocs site or something.

      • CyberSeeker@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        1
        ·
        edit-2
        22 hours ago

        After reading through the GitHub docs, the most impressive thing is that they open sourced their Thunderbolt coding agent for Claude Code. There are quite a few skills available for implementation planning, dependency/build environment setup, coding, linting/cleanup, QA, and managing agent pull requests. Pretty good examples if you are looking at building Claude Code skills.

    • FauxLiving@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      2
      ·
      22 hours ago

      It looks like it, it doesn’t seem to be doing any self-referencing/agentic things out of the box so the end user would need to build a bit to cover their specific use cases.

      They seem to be aiming more at the startup/small company demographic than at self-hosters. This is just based on a skim of the repo and their product page, I haven’t looked at it too hard yet.