• tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    6 months ago

    KoboldAI has models trained on erotica (Erebus and Nerybus). It has the ability to spread layers across multiple GPUs, so as long as one is satisfied with the output text, in theory, it’d be possible to build a very high-powered machine (like, in wattage terms) with something like four RX 4090s and get something like real-time text generation. That’d be like $8k in parallel compute cards.

    I’m not sure how many people want to spend $8k on a locally-operated sex chatbot, though. I mean, yes privacy, and yes there are people who do spend that on sex-related paraphernalia, but that’s going to restrict the market an awful lot.

    Maybe as software and hardware improve, that will change.

    The most obvious way to cut the cost is to do what has been done with computing hardware for decades, like back when people were billed for minutes of computing time on large computers in datacenters – have multiple users of the hardware, and spread costs. Leverage the fact that most people using a sex chatbot are only going to be using the compute hardware a fraction of the time, and then have many people use the thing and spread costs across all of them. If any one user uses the hardware 1% of the time on average, that same hardware cost per user is now $80. I’m pretty sure that there are a lot more people who will pay $80 for use of a sex chatbot than $8000.