• 2 Posts
  • 133 Comments
Joined 11 months ago
cake
Cake day: April 27th, 2024

help-circle
  • As a fellow Futo user: it’s not great out of the box. My biggest recommendations are:

    • under Languages and models, download all the voice models (if you use those), transformers, and wordlists you can for your languages
    • if you use multiple languages, set the check on “multilingual typing” for ALL of those languages
    • this is probably the biggest one: in text prediction -> Advanced Parameters, DRASTICALLY change the values. The original ones are 3.4 and 4.0 for LLM strength and autocorrect threshold, mine are currently set at 28.5 and 0.8, respectively. This takes the autocorrect from “occasionally working” to “as good as SwiftKey” for me.
    • Keyboard and Typing -> Long Press -> Show hints. Could not find that for ages so thought I’d add it here.

    Also, two super useful shortcuts: you can press the space-bar and move your finger around to move the pointer; and the same for backspace to fine-control what to delete.

    Hope this helps, but if not… What additional gripes do you habe with it?












  • I am a bit confused tbh 😅

    The link you send links to docker projects, the link I sent is the second one of those. Seems pretty straightforward?

    But to be fair, I have never used docker for any of this. In my nix config, it’s literally just:

        services.prowlarr.enable = true;
        services.prowlarr.openFirewall = true;
    

    There’s not really anything you need to configure host-side. Prowlarr needs to be able to communicate with sonarr and radarr (same as jackett), but otherwise it’s basically stateless.








  • I have one big frustration with that: Your voice input has to be understood PERFECTLY by TTS.

    If you have a “To Do” list, and speak “Add cooking to my To Do list”, it will do it! But if the TTS system understood:

    • Todo
    • To-do
    • to do
    • ToDo
    • To-Do

    The system will say it couldn’t find that list. Same for the names of your lights, asking for the time,… and you have very little control over this.

    HA Voice Assistant either needs to find a PERFECT match, or you need to be running a full-blown LLM as the backend, which honestly works even worse in many ways.

    They recently added the option to use LLM as fallback only, but for most people’s hardware, that means that a big chunk of requests take a suuuuuuuper long time to get a response.

    I do not understand why there’s no option to just use the most similar command upon an imperfect matching, through something like the Levenshtein Distance.