Yep. In fact, Amazon devices can connect to other Amazon devices over their Sidewalk meshnet and get the wifi password that way. I’m never getting anything from Amazon more complicated than a screwdriver.
Yep. In fact, Amazon devices can connect to other Amazon devices over their Sidewalk meshnet and get the wifi password that way. I’m never getting anything from Amazon more complicated than a screwdriver.
Webtoon is still shitty in other ways. When they adapt a property, they want it their way, regardless of the author’s original vision. I’ve seen several stories that originated on Royal Road get Webtoon adaptations, and the adaptations always seem to change or leave out important parts of the story, making characters look stupid or just completely replacing entire sets of characters, forcing the story to diverge substantially when inevitably something they got rid of turns out to have been critically important to where the author was taking things. They turn great stories into middling slop every single time.
Tweet not found, not even when I change the URL to go directly to Twitter. Was it deleted?
Try KittyToy (itch.io).
I don’t listen to many podcasts, but those two are pretty great.
Router-level VPN is going to be more difficult to configure and cause more problems than just having it on all your devices. There are some games where online play just refuses to work if connecting through a VPN. Some mobile apps are the same. When a website blocks your currently selected server, and the usual solution is switching to another server, that’s going to be more difficult and more tedious when it’s configured at the router level. In addition, if you do something like using a self-hosted VPN in order to connect remotely to a media server on your home network, that becomes more difficult if your home router is on a different VPN.
If you’re trying to keep local devices in the building from phoning home and being tracked, a PiHole or router-level firewall might be a better solution. I think if you’re running a pfsense or opnsense router and are a dab hand with VLANs then maybe you could get what you’re looking for with router-level VPN, but it’s a huge hassle otherwise. Just put Mullvad on your computers and phones and call it a day.
PCIe gen 5 is for the PCIe slots and NVMe storage slots, but they’re backwards compatible; you can put a gen 3 component in a gen 5 slot and it will work at gen 3 speeds. Similarly, if you put a gen 5 component in a gen 4 slot, it will be limited to gen 4 speeds. Right now there’s very little appreciable difference between gen 4 and gen 5 unless you’re spending a lot of money on the component (GPU/storage). Another thing to note is that Gen 5 requires that both the CPU and motherboard support it; a CPU with gen 4 support in a gen 5 motherboard will limit all the slots to gen 4 speeds.
RAM is a totally different standard that must be matched exactly for what the motherboard has; if it’s a DDR5 motherboard then you have to use DDR5 RAM or it won’t even fit in the slots. You can get a PCIe gen 5 motherboard and just use gen 4 SSDs or GPUs, that’s perfectly fine and leaves you room to upgrade later.
Seems mostly fine to me, I game all the time on Linux (Bazzite gang 🤘) with a 3900X + 7900GRE, haven’t had any significant issues aside from needing to make sure clock speeds were configured correctly on the GPU. Two ram sticks is the way to go with these systems as sometimes they don’t support 4 sticks at full speed.
You’re right that GPU passthrough is definitely more for tinkering or advanced users with very specific needs (usually professionals who need Windows/Nvidia and choose to run it in a VM rather than dual-boot), with a budget to match. For a gamer couple, having fully separate systems is going to be much less hassle and more resilient against failure.
The one thing I would recommend changing is the power supply, it’s unironically the most important component in the computer because if it fails it can kill everything else, and the System Power 10 is known enough for being low-quality that discussions of that come up in web searches. Poor quality power supplies can damage your hardware and otherwise cause weird, intermittent issues even if everything seems to work fine most of the time, and will fail and shut off the computer when a good power supply would have just kept on chugging. Seasonic and Corsair are considered the best brands and have 10 year warranties - they’re more expensive, but they’re worth it. You want 80+ Gold or better these days, this is a buy once, cry once component.
If you don’t have a UPS, I would also recommend getting one at some point, either one big shared unit (if they’ll be close together) or two individual units. Having backup power will allow you to shut down the computers gracefully during a power outage, and prevents the worst-case scenario where the power goes out while the computer is installing updates and it turns into a brick.
At least their username is accurate!
Web ads are a security risk that even the FBI has acknowledged, so your friends should be aware that having uBlock Origin installed is nearly as important as having virus protection.
Regarding profiles, having two is generally recommended - your main profile with no Google services, and a secondary profile only for apps that absolutely require Google Play Services. Personally, I just dump everything in one profile and deny nearly every permission to anything Google, and on top of the sandboxing that’s enough of an improvement over stock Android that I don’t bother with two profiles.
They advertise E2EE as a feature
They can call it E2EE as much as they want, but it’s a lie. It’s encrypted in transit and at rest, at least on the user’s device, but unlike true E2EE, they can decrypt and view any conversation they want to.
I wouldn’t trust any phone with GrapheneOS preloaded unless it was directly sold by GrapheneOS themselves. Especially not from a site that phrases things in an almost uncanny way.
He did at the beginning, but he helped them get what they wanted in the end, and I think that counts for something.
“We’re thankful that the Biden administration played the long game on sick days and stuck with us for months after Congress imposed our updated national agreement,” Russo said. “Without making a big show of it, Joe Biden and members of his administration in the Transportation and Labor departments have been working continuously to get guaranteed paid sick days for all railroad workers.
“We know that many of our members weren’t happy with our original agreement,” Russo said, “but through it all, we had faith that our friends in the White House and Congress would keep up the pressure on our railroad employers to get us the sick day benefits we deserve. Until we negotiated these new individual agreements with these carriers, an IBEW member who called out sick was not compensated.”
Archives are ideal for identifying sneaky behavior like that. You never know when an admin might have the ability to delete or edit something without anyone noticing.
Looking into it, ZOOM Platform is run by the Jordan Freeman Group, which has a similar mission statement to GOG, that being a DRM-free games platform that also touches up and sells old games. They seem to have acquired publishing rights to the FlatOut trilogy, along with either source code or enough resources to touch them up without it. I didn’t know about them until this post, so it’s nice to see that GOG isn’t the only DRM-free old-games platform around.
You don’t need to add the exe of whatever mod tool to Steam, use Steam Tinker Launch. It lets you add an exe to run instead of the game, concurrent with the game, or injected after the game is up, and it will run in the same prefix that Proton uses for that game. It also has tools for installing and using several mod managers, and generally a ton of good features for tinkering with the game.
The main issue I haven’t solved is getting something like the Nexus mods “open in manager” to work. My guess is I might have to install, run, and configure a web browser inside the prefix, but that sounds really annoying so I haven’t tried it.
Technically, any model trained on LAION-5B before December 2023 was trained on CSAM.
But yeah, I expect any porn model trained on a sufficient diversity of adult actors could be used to make convincing CP even without having it in the training data. AI image generation is basically the digital equivalent of a chainsaw - a tool for a particular messy job that can really hurt people if used incorrectly. You wouldn’t let a typical kid run around unattended with one, that’s for sure.
You can add swipe (glide?) typing into HeliBoard. From their github readme:
- Glide typing (only with closed source library ☹️)
- library not included in the app, as there is no compatible open source library available
- can be extracted from GApps packages (“swypelibs”), or downloaded here (click on the file and then “raw” or the tiny download button)
The only reason HeliBoard doesn’t include this themselves is presumably legal liability plus their dedication to the app not having any network permissions at all.
You’re entirely correct, but in theory they can give it a pretty good go, it just requires a lot more computation, developer time, and non-LLM data structures than these companies are willing to spend money on. For any single query, they’d have to get dozens if not hundreds of separate responses from additional LLM instances spun up on the side, many of which would be customized for specific subjects, as well as specialty engines such as Wolfram Alpha for anything directly requiring math.
LLMs in such a system would be used only as modules in a handcrafted algorithm, modules which do exactly what they’re good at in a way that is useful. To give an example, if you pass a specific context to an LLM with the right format of instructions, and then ask it a yes-or-no question, even very small and lightweight models often give the same answer a human would. Like this, human-readable text can be converted into binary switches for an algorithmic state machine with thousands of branches of pre-written logic.
Not only would this probably use an even more insane amount of electricity than the current approach of “build a huge LLM and let it handle everything directly”, it would take much longer to generate responses to novel queries.