If we stop doing business with SpaceX, we immediately demolish most of our capability to reach space, including the ISS until Starliner quits failing. Perhaps instead of trying to treat this as a matter of the free market we should recognize it as what it is - a matter of supreme economic and military importance - and force the Nazi fucker out.
sleep_deprived
I'd be interested in setting up the highest quality models to run locally, and I don't have the budget for a GPU with anywhere near enough VRAM, but my main server PC has a 7900x and I could afford to upgrade its RAM - is it possible, and if so how difficult, to get this stuff running on CPU? Inference speed isn't a sticking point as long as it's not unusably slow, but I do have access to an OpenAI subscription so there just wouldn't be much point with lower quality models except as a toy.
Well they said .NET Framework, and I also wouldn't be surprised if they more or less wrapped that up - .NET Framework specifically means the old implementation of the CLR, and it's been pretty much superseded by an implementation just called .NET, formerly known as .NET Core (definitely not confusing at all, thanks Microsoft). .NET Framework was only written for Windows, hence the need for Mono/Xamarin on other platforms. In contrast, .NET is cross-platform by default.
Honestly, after DOS2, I'd play a Larian game in any setting just based on them being the devs - and that goes double after BG3. Their handle on storytelling and environments is so good I'd trust it would be enjoyable even in a setting I'm not interested in.
This is a use-after-free, which should be impossible in safe Rust due to the borrow checker. The only way for this to happen would be incorrect unsafe code (still possible, but dramatically reduced code surface to worry about) or a compiler bug. To allocate heap space in safe Rust, you have to use types provided by the language like Box
, Rc
, Vec
, etc. To free that space (in Rust terminology, dropping it by using drop()
or letting it go out of scope) you must be the owner of it and there may be current borrows (i.e. no references may exist). Once the variable is drop
ed, the variable is dead so accessing it is a compiler error, and the compiler/std handles freeing the memory.
There's some extra semantics to some of that but that's pretty much it. These kind of memory bugs are basically Rust's raison d'etre - it's been carefully designed to make most memory bugs impossible without using unsafe
. If you'd like more information I'd be happy to provide!
Even as an (older) zoomer in the US, this was never a thing for me. No one cared what phone you used. If you had an Android you wouldn't be in iMessage group chats but no one judged you for it.
The GPU I used is actually a 1080, with a (rapidly declining in usefulness) Intel 4690k. But I suppose laptop vs desktop can certainly make all the difference. What I really want is GPU virtualization, which I've heard AMD supports, but I'm not about to buy a new GPU when what I've got works fine.
My experience with single GPU passthrough on Proxmox to a media VM was pretty positive, especially for it being an old Nvidia card. Even as someone doing it for the first time, it just took about 10 minutes to figure out the passthrough itself and another ~15 to figure out some driver issues. And it's worked perfectly since then. All in all much better than what I'd expected.
Besides rendering bugs that may or may not be Safari's fault, I wanted to get uBlock Origin on an iPhone but it's not available, IIRC because the content blocking API is more restrictive than what uBlock is designed for.
The CAD files needed for making an expansion card are open source so you could make your own if you have the expertise. It's been tried before, but to my knowledge nobody's done it. I could be wrong though.
I was even kind of interested, but then I got further in the main quest and figured out what the ending is...
Then I felt like there was no point to anything I did.
Reminds me of The Gourmet in Skyrim