

Regular users can use Gemini, Deepseek, Meta AI, and there will probably be many more services in the future.
Regular users can use Gemini, Deepseek, Meta AI, and there will probably be many more services in the future.
NFS gives me the best performance. I’ve tried GlusterFS (not at home, for work), and it was kind of a pain to set up and maintain.
If it works, I don’t update unless I’m bored or something. I also spread things out on multiple machines, so there’s less chance of stuff happening like you describe with the charts feature going away. My NAS is pretty much just a NAS now.
You can probably backup your configs/data, upgrade, then deploy jellyfin again, restore, and reconfigure. You should probably backup your data on your ZFS pool. But, I recently updated to the latest TrueNas Scale from ~5 year old FreeBSD version of TrueNas and the pools still worked fine (none of the “apps” or jails worked, obviously). The upgrade process even ported my service configurations over. I didn’t care about much of the data in the pools, so only backed up the most important stuff.
I personally use a dual core pentium with 16GB of RAM. When I first installed TrueNas (FreeNas back then), I only had 8GB of RAM, but that proved to be not enough to run all the services I wanted, so I would suggest 12-16GB. Depending on the services you want to run any multi-core x86 CPU that allows 16GB of RAM to be used should be adequate. I believe TrueNas recommends ECC RAM, but I don’t think using consumer grade RAM and hardware has caused me any problems. I’m also using an old SSD for the system drive, which I is recommended now (I used to use 2 mirrored USB thumb drives, buy that’s not recommended anymore). Very importantly, make sure the HDD(s) you get are not shingled drives; made that mistake initially, and performance was ridiculously bad.
Yeah. If you’re a minor you have to take Drivers Ed that requires a couple hours of driving with an instructor. If you’re an adult, you can just take the written and driving test. I think I just drove around the block, and did a reverse parking test for my driving test. Depending on where you live, roundabouts are not common here. I don’t think I saw one IRL until I was in my late 20s when I moved to a different state.
Yeah, I was disappointment when I bought a very expensive Galaxy S22 to replace my old Moto G whose charging port wore out,. The S22 had worse battery life, camera, and no noticeable performance improvements. Recently, my S22 stopped charging, and I just bought a “Mint”-grade used Pixel 6 and installed GrapheneOS on it. Happy so far, and it’s nice to be able to block network access to all apps, including Google’s.
Marginal cost doesn’t always decrease. More people buying gold or whatever won’t decrease the price of gold. The cheapest way to feed cattle is to just let them graze, but there isn’t enough land on Earth for everyone to eat as much beef as Americans, even if using intensive agriculture to grow feed (which degrades the soil over time and results in large amounts of greenhouse gas emissions). I don’t think there’s enough land on Earth to maintain the current human population for very long. I.e. I think we are in the overshoot phase of a boom and bust population dynamic. Saw this graphic a while back, and it’s wild how much of the biomass we’ve took over:
Some of the “open” models seem to have augmented their training data with OpenAI and Anthropic requests (I. E. they sometimes say they’re ChatGPT or Claude). I guess that may be considered piracy. There are a lot of customer service bots that just hook into OpenAI APIs and don’t have a lot of guardrails, so you can do stuff like ask a car dealership’s customer service to write you Python code. Actual piracy would require someone leaking the model.
Some people’s aversion of algorithms on the fediverse kind of reminds me of people’s aversion of GMO food. Genetically modifying rice to contain more vitamin D is probably good; genetically modifying vegetables to contain more cyanide would probably be bad. Algorithms don’t have to be built to maximize “engagement;” they can be designed to maximize other metrics, or balance multiple metrics, or be user-customizable.
IMO, Mastadon is much worse off for their refusal to implement any kind of algorithm outside their “explore” feed. When I tried using Mastodon, search was unhelpfully in chronological order, and my home feed just got overtaken by the people that post the most. In contrast, Lemmy’s handling of algorithms is pretty good, imo.
As bad as search engines are now, they’d be even worse if they just gave you results in chronological order.
I’m curious if ByteDance could just create a new legal entity and call it TikTak or something.
If you have to verify children’s identity, you have to verify everyone’s identity. This is part of KOSA. https://www.eff.org/deeplinks/2024/12/kids-online-safety-act-continues-threaten-our-rights-online-year-review-2024
Oldest I got is limited to 16GB (excluding rPis). My main desktop is limited to 32GB which is annoying, because I sometimes need more. But, I have a home server with 128GB of RAM that I can use when it’s not doing other stuff. I once needed more than 128GB of RAM (to run optimizations on a large ONNX model, iirc), so had to spin up an EC2 instance with 512GB of RAM.
I just use Joplin, encrypted, and synced through dropbox. Tried logseq, but never really figured out how to use its features effectively. The notebook/note model of Joplin seems more natural to me. My coding/scripting stuff mostly just goes into git repos.
The PC I’m using as a little NAS usually draws around 75 watt. My jellyfin and general home server draws about 50 watt while idle but can jump up to 150 watt. Most of the components are very old. I know I could get the power usage down significantly by using newer components, but not sure if the electricity use outweighs the cost of sending them to the landfill and creating demand for more newer components to be manufactured.
CK2 - 400h
Fallout NV (guessing most of this has been TTW) - 190h
Stellaris - 180h
Xcom 2 - 140h
GTA 5 - 99h
Cities Skylines - 95h
Skyrim - 90h
Civ5 - 85h
Xcom - 83h
The other games I’ve played are pretty much the standard play-through times. (< 70h)
I’m loading up on vacuum tubes.
Last time I looked it up and calculated it, these large models are trained on something like only 7x the tokens as the number of parameters they have. If you thought of it like compression, a 1:7 ratio for lossless text compression is perfectly possible.
I think the models can still output a lot of stuff verbatim if you try to get them to, you just hit the guardrails they put in place. Seems to work fine for public domain stuff. E.g. “Give me the first 50 lines from Romeo and Juliette.” (albeit with a TOS warning, lol). “Give me the first few paragraphs of Dune.” seems to hit a guardrail, or maybe just forced through reinforcement learning.
A preprint paper was released recently that detailed how to get around RL by controlling the first few tokens of a model’s output, showing the “unsafe” data is still in there.
Looked it up, and looks like a local criminal justice reform non-profit bailed him out earlier this year. Looks like he’s still awaiting trial though, is on an ankle monitor, and they keep resetting dismissal hearing dates.
Yeah, the company I was working at got bought out and then they layed the entire tech team and pretty much everyone else. Co-founded a business with coworkers, but it’s not bringing in any revenue and not sure it ever will bring in very much, so have been applying to jobs. Only got a few interviews, then ghosted afterwards. I’m guessing a part of it is I have a criminal charge pending, and the first thing you see on Google when you search my name and town is one of those mugshot websites. Maybe I should go into construction, lol.
Oh, I forgot about Claude. Last time I tried it, it seemed on par or even better that ChatGPT-4o (but was missing features like browsing).