Well, five times zero women checks out.
Well, five times zero women checks out.
The undocked Switch is in the same ballpark for raw power as the 360 and PS3, so as long as they’ve managed to sufficiently unfuck the game’s nightmare spaghetti code, should be just fine.
That’s incredible. Certified “Directive #4” moment.
Well, the ones based on Chromium aren’t, anyway. I’ve heard some major criticisms of Safari in the last few years, for what that’s worth.
The NHS’ virtual appointment service in the UK doesn’t support Firefox either, only Chrome, Safari and Edge. The dark days of “please view this website in Internet Explorer 6” are creeping closer to the present again. I hate the modern internet.
Funnily enough, one of the few legitimately impactful non-enterprise uses of AVX512 I’m aware of is that it does a really good job of accelerating emulation of the Cell SPUs in RPCS3. But you’re absolutely right, those things are very funky and implementing their functions is by far the most difficult part of PS3 emulation.
Luckily, I think most games either didn’t do much with them or left programming for them to middleware, so it would mostly be first- and second-party games that would need super-extensive customisation and testing. Sony could probably figure it out, if they were convinced there was sufficient demand and potential profit on the other side.
There’s even rumours that the next version of Windows is going to inject a bunch of AI buzzword stuff into the operating system. Like, how is that going to make the user experience any more intuitive? Sounds like you’re just going to have to fight an overconfident ChatGPT wannabe that thinks it knows what you want to do better than you do, every time you try opening a program or saving a document.
The Xbox 360 was based on the same weird, in-order PowerPC 970 derived CPU as the PS3, it just had three of them stuck together instead of one of them tied to seven weird Cell units. The TL;DR of how Xbox backwards compatibility has been achieved is that Microsoft’s whole approach with the Xbox has always been to create a PC-like environment which makes porting games to or from the Xbox simpler.
The real star of the show here is the Windows NT kernel and DirectX. Microsoft’s core APIs have been designed to be portable and platform-agnostic since the beginning of the NT days (of course, that isn’t necessarily true of the rest of the Windows operating system we use on our PCs). Developers could still program their games mostly as though they were targeting a Windows PC using DirectX since all the same high-level APIs worked in basically the same way, just with less memory and some platform-specific optimisations to keep in mind (stuff like the 10MB of eDRAM, or that you could always assume three 3.2GHz in-order CPU cores with 2-way SMT).
Xbox 360 games on the Xbox One seem to be run through something akin to Dolphin’s “Übershaders” - in this case, per-game optimised modifications of an entire Xenon GPU stack implemented in software running alongside the entire Xbox 360 operating environment in a hypervisor. This is aided by the integration of hardware-level support for certain texture and audio formats common in Xbox 360 games into the Xbox One’s CPU design, similarly to how Apple’s M-series SoCs integrate support for x86-style memory ordering to greatly accelerate Rosetta 2.
Microsoft’s APIs for developers to target tend to be fairly platform-agnostic - see Windows CE, which could run on anything from ARM handhelds to the Hitachi SH-4 powered Sega Dreamcast. This enables developers who are mostly experienced in coding for x86 PCs running Windows to relatively easily start writing programs (or games) for other platforms using those APIs. This also has the beneficial side-effect of allowing Microsoft to, with their collective first-hand knowledge of those APIs, create compatibility layers on an x86 system that can run code targeted at a different platform.
Yeah, Windows’ bullshit is what drove me to Linux in the first place. I only have it on my gaming system, and only because Discord’s stupid screensharing doesn’t transmit audio on Linux, NVIDIA’s drivers for Linux suck balls (going AMD next time now that their cards are good again) and there are a couple of games my friends play that have issues on Linux. I’ve never run into a game on my everyday laptop that Linux couldn’t run, and the Steam Deck will take basically whatever you throw at it.
Windows is a barely-functional rat’s nest of code spaghetti that falls apart at complete random. Sometimes your audio drivers will just stop working for no apparent reason. Sometimes your computer will just refuse to connect to the internet until you do a clean install. Windows Update apparently runs Prime95 in its spare time and so does the Antimalware Service Executable. I hate using it so much. I wish Windows would just curl up and die.
A Libertarian candidate for the US Senate back in 2018 proposed giving shotguns to homeless people for self-defence.
They already have alt-tech, which had kind of a headstart on the Fediverse.
The Fediverse is home to a lot of young, tech-minded people distrustful of major corporations. The younger generations are more likely to come out as transgender due to greater awareness and acceptance of gender identity and dysphoria, and a decentralised, open platform is naturally going to appeal to communists, syndicalists and other left-wingers who don’t want some billionaire buying the next website they get comfortable on. And funnily enough, there are a surprising number of trans people in the tech sector, to the point where trans-flag socks have become a meme among programmers.
This is the first I’ve heard of the JPEG XL format, but it sounds pretty good!
Hopefully it doesn’t get misused by websites to mangle lossless compressed images with so much compression they’re barely visible to save a few kilobytes, though.
This is a total affront to the ethos of the web and everyone involved in drafting this awful proposal should be publicly shamed. Stick sandwich boards on each of them saying “I tried to build the Torment Nexus”, chain them together and march them through the streets while ringing a bell and chanting “shame”.
In the case of the Surface Go family, there isn’t really anything comparable from other companies. It’s unironically the best compact tablet I’m aware of that you can put Linux on, and it runs Pop!_OS without issue once you disable Secure Boot. The only better Linux tablet for me would be an iPad Mini, but you can’t put Linux on one of those and even if you could it’s ARM-based so most proprietary apps won’t work on it.
In general, your tablet options for something smaller and handier than full-size 2-in-1s are pretty limited if you don’t want to be running iPadOS, so excluding Microsoft’s devices from the running if you want to put Linux on your tablet is pointless. Yeah, buying a Surface Laptop to put Linux on there is a bit weird, but I can see the Surface Pro family yielding a good ARM Linux tablet some day.
USB-C hubs all seem to be dodgy crap made by anonymous Chinese companies and resold through various companies, including the likes of Apple. There’s an absolute dearth of hubs made by actual reputable firms.
VR possibly needs to get as high as 16K to be truly perfect for the human eye, so it could be useful there. But 16K headsets are a long way off, and refresh rate also needs to improve since the human eye can perceive up to 2,000 frames per second, so I feel like refresh rate will be the priority over pixel density by the time we get even 8K headsets.
IMAX 1570 film is equivalent to somewhere between 12K and 18K resolution, though as you say that’s only going to matter on a really big display or if you’re creating content for future VR headsets. As for frame rate, more than 24 FPS doesn’t really seem popular in cinema - I remember that Hobbit film getting quite a negative reaction when it tried 48 FPS, and then there’s the Soap Opera Effect. The colour characteristics of different kinds of film also appeal to some artistic filmmakers.
None of those points really justify the significantly more expensive and complicated workflow that comes with analogue film, mind. I do wish digital hadn’t taken over so soon because a whole bunch of media is now stuck with 480p digital as the best it’s realistically ever going to look barring AI upscaling, but I can’t say I really blame the producers for making that decision because digital is far less of a pain in the ass.
Depends on the film itself being used. You can get a pretty insane level of detail on the professional-grade stuff. It’s how we’re able to do 4K remasters of stuff from decades ago. Digital still has a fair bit to go before it can fully outshine analogue film.
Kbin. Kbity. Like kitty but kbin. Kbity!