There was an EU-wide one that gota lot of its funding redirected to AI stuff recently that you might be thinking of.
There was an EU-wide one that gota lot of its funding redirected to AI stuff recently that you might be thinking of.
That’s more likely to be the tool assuming it’s running on a case-insensitive filesystem than it is Windows breaking anything. If you mount networked storage running on a case-sensitive machine, that’s something that’s worked fine in Windows for a very long time.
No, that is an entirely unrelated bad decision. It being okay to not have a popup to opt out of secure boot when it does its one job and notices you’re about to run insecure code in kernel mode doesn’t make every other user-hostile thing Microsoft ever does magically okay.
It’s upstream GRUB that’s decided the older GRUB versions are insecure and not to be trusted. Microsoft just propagated that to machines running distros that weren’t shipping patched GRUB builds yet. Up-to-date Debian wouldn’t be affected provided that they downstreamed fixes quickly.
https://fedia.io/m/linux@lemmy.ml/t/1111595/-/comment/6916699 says that Debian’s GRUB wasn’t affected, but another part of the boot sequence was.
You can’t trust users to make informed decisions about cybersecurity as most users don’t have the necessary background knowledge, so won’t think beyond this popup is annoying me and has a button to make it go away and I am smart and therefore immune to malware. Microsoft don’t want Windows to have the reputation for being infested with malware like it used to have, and users don’t want their bank details stolen. If something’s potentially going to be a bad idea, it’s better to only give the decision to people capable of making it an informed decision. That’s why we don’t let children opt into surgery or decide whether to have ice cream for dinner, and have their parents decide instead.
The comment you’re quoting was replying to someone suggesting a warning popup, and saying it would be a bad idea, rather than suggesting the secure boot UEFI option should be taken away. You need at least a little bit more awareness of the problem to know to toggle that setting.
If you’re doing things properly, you’ll know your Microsoft account password or have it in a password manager (and maybe have other account recovery options available like getting a password reset email etc.), and have a separate password for the PC you’re locked out of, which would be the thing you’d forgotten. If someone isn’t computer-literate, it’s totally plausible that they’d forget both passwords, have no password manager, and not have set up a recovery email address, and they’d lose all their data if they couldn’t get into their machine.
Is that enough to mitigate how much worse bare Google is than it was ten years ago, back when they were winning against SEO bots? In my experience, it hasn’t been, but I’ve not done enough AI-aided web searches to have a good sample size.
If you give a chip more voltage, its transistors will switch faster, but they’ll degrade faster. Ideally, you want just barely enough voltage that everything’s reliably finished switching and all signals have propagated before it’s time for the next clock cycle, as that makes everything work and last as long as possible. When the degradation happens, at first it means things need more voltage to reach the same speed, and then they totally stop working. A little degradation over time is normal, but it’s not unreasonable to hope that it’ll take ten or twenty years to build up enough that a chip stops working at its default voltage.
The microcode bug they’ve identified and are fixing applies too much voltage to part of the chip under specific circumstances, so if an individual chip hasn’t experienced those circumstances very often, it could well have built up some degradation, but not enough that it’s stopped working reliably yet. That could range from having burned through a couple of days of lifetime, which won’t get noticed, to having a chip that’s in the condition you’d expect it to be in if it was twenty years old, which still could pass tests, but might keel over and die at any moment.
If they’re not doing a mass recall, and can’t come up with a test that says how affected an individual CPU has been without needing to be so damaged that it’s no longer reliable, then they’re betting that most people’s chips aren’t damaged enough to die until the after warranty expires. There’s still a big difference between the three years of their warranty and the ten to twenty years that people expect a CPU to function for, and customers whose parts die after thirty-seven months will lose out compared to what they thought they were buying.
It wasn’t me who you replied to originally - I agree that it’s most likely AMD are just being super cautious given historically how many times bad news for their competitors has been falsely equated by the press as equivalent to a minor issue they’ve had, and the delay moving things after the microcode update and therefore making launch-day benchmarking more favourable is just a bonus.
You’ve misunderstood. The original release date was set, then Intel announced the microcode update, which was after the original release date, then AMD announced that they’d be delaying the release date, and that new release date is after the microcode update.
It’s at least common on forums as bots love making accounts with non-megacorp email addresses on PhpBB and MyBB forums. Typically, there aren’t people signing up the same services with business emails as personal ones, so if ones expecting not to be used by businesses want to fight spam, it’s generally pretty effective and consequence-free to block email providers not known to have effective anti-bot measures built in.
That would be annoying for people who work on files with a double extension for legitimate reasons, e.g. .tar.gz
, and (this can’t be stressed strongly enough) Windows users do not pay attention to warning popups, so it wouldn’t actually help. Despite it being eighteen years since Windows Vista released, and therefore vanishing unlikely that any given software was written assuming that Windows didn’t have a permissions system, it’s still most people’s first troubleshooting step to try and run things as admin, and you still get loads of people (including ones who should know better, e.g. ones who also use Linux and would never log in as root) who disable UAC as one of the first things they do when setting up a windows install, and end up running everything as the equivalent of root just to suppress the mildly annoying pop-up when something asks for elevated permissions.
So, your proposed popup:
Typically Windows applications bundle all their dependencies, so Chocolatey, WinGet and Scoop are all more like installing a Flatpak or AppImage than a package from a distro’s system package manager. They’re all listed in one place, yes, but so’s everything on FlatHub.
It’s been the default since 2015 when Windows 10 launched, although there was an obvious button to opt out during first-time setup back then which was then respected permanently. It’s got gradually less prominent over time, and maybe the article’s just doing a really bad job of explaining that it’s no longer something where your initial preference is permanent and it’ll change back to the default every so often.
It’s a silly flag to use as it only works when running 32-bit Windows applications on 64-bit Windows, and if you’re compiling from source, you should also have the option to just build a 64-bit binary in the first place. It made a degree of sense years ago when people actually used 32-bit Windows sometimes (which was usually just down to OEMs installing the wrong version on prebuilt PCs could have supported 64-bit) if you really wanted to only have one binary or you consumed a precompiled third party library and had to match its architecture.
They tried. UWP and the Windows Store did loads to boost security and make the source of apps verifiable, but people hated it and barely used it, so the holes they were supposed to patch stayed open. The store itself did have the problem that part of its raison d’être was to try and take a cut of the sales of all software for Windows, like Apple do for iOS, and UWP made certain things a pain or impossible (sometimes because they were inherently insecure), but UWP wasn’t tied to the store and did improve even though it’s barely used.
They update on two Tuesdays a month, and have done that at least since XP. Even with the most reboot-keen settings, the update doesn’t happen until the time of day you’re least likely to be using the machine based on when you typically do it. It tells you when that time will be and gives you several hours of notice with a popup with the option to delay. Depending on the variant of Windows you’re using, you have settings to delay a forced reboot for up to a week (Home), a month (Pro) or forever (Enterprise). Obviously, that’s not enough to make sure no one ever gets updates forced on them when they don’t want them, and it would be nice if there was a way to distinguish users who know what they’re doing from users who don’t so people who do could be given more power to control if and when they install updates, but it is enough to ensure that checking the equipment before you use it is enough, potentially two weeks in advance.
It doesn’t necessarily work that way, though. If tests tell you you broke something immediately, you don’t have time to forget how anything works, so identifying the problem and fixing it is much faster. For the kind of minor bug that’s potentially acceptable to launch a game with, if it’s something tests detect, it’s probably easier to fix than it is to determine whether it’s viable to just ignore it. If it’s something tests don’t detect, it’s just as easy to ignore whether it’s because there are no tests or because despite there being tests, none of them cover this situation.
The games industry is rife with managers doing things that mean developers have a worse time and have the opposite effect to their stated goals. A good example is crunch. It obviously helps to do extra hours right before a launch when there’s the promise of a holiday after the launch to recuperate, but it’s now common for games studios to be in crunch for months and years at a time, despite the evidence being that after a couple of weeks, everyone’s so tired from crunch that they’re less productive than if they worked normal hours.
Games are complicated, and building something complicated in a mad rush because of an imposed deadline is less effective than taking the time to think things through, and typically ends up failing or taking longer anyway.
It’s a 250W+ part running at around 1V, so it’s going to draw a lot of current. Power is supplied via many pins on the back of the CPU, and they’re connected to many traces, so it’s not putting all that current through just one. It still puts out a lot of heat anyway, which is why modern motherboards have large heat sinks, sometimes with fans, on their VRMs.
It’s easy to get pressured into thinking it’s your responsibility. There’s also the risk that an unhappy company will make a non-copyleft clone of your project, pump resources into it until it’s what everyone uses by default, and then add proprietary extensions so no one uses the open-source version anymore, which, if you believe in the ideals of Free Software, is a bad thing.