

Software to allow partitioning gpu resources among multiple virtual machines instead of just assigning the one PCIe device to a single VM. Very useful for having a single GPU do 3D acceleration on a host and multiple guests at the same time.
Software to allow partitioning gpu resources among multiple virtual machines instead of just assigning the one PCIe device to a single VM. Very useful for having a single GPU do 3D acceleration on a host and multiple guests at the same time.
Either Linus or Greg K-H, likely after feedback from many others.
We should be looking at his given reasons, not making assumptions based on some ineffable set of considerations that he might have.
Christof’s given reason of complexity is sensible, it’s also one already considered when allowing R4L in the first place; adding rust language support has been deemed worth the additional complexity.
It definitely has its roots in Debian, but when you need to use that weird closed source application for work, if it has a “supported” (for a given value of support) Linux distro it’ll be Ubuntu.
I personally prefer straight Debian myself, or something entirely different but when asked for a recommendation by friends it’s Ubuntu.
I had a similar issue, low frequency or crashes after gaming for a while, turns out the fan control wasn’t setting the fan correctly and the GPU was overheating. Utilisation didn’t show it, but frequency did.
Restarting the game could be giving it enough time to cool off?
Cost to manufacture is not more than wages, but cost to purchase a good is always more than the total cost of labour needed to produce it, so long as profit exists.
The money isn’t free so much as redistributed from taxation elsewhere, think of it as the same as subsidising industry except only to the workers of that industry (instead giving it to owners and expecting the savings to trickle downwards). You could also consider it an income tax rebate with more fine-grained control of who gets it.
It doesn’t seem particularly ground-breaking of a concept; I see the value in investing money into necessary but unprofitable industry though my concern is that if you subsidise wages of a business with a profit incentive, management may lower wages to compensate.
Man if I was concerned about sinking the time to make a configuration for the compositor with a bus factor of 1 man-child, and a toxic community; I can’t imagine anybody investing the time to make a compositor is going to want to hitch themselves to that cart.
The compositor is really solid and makes for a great user experience but I’ll be fucked if every word vaxry writes doesn’t make me want to move to sway or niri.
My parents treated my device access something they had to keep a keen eye on. They were good at manually making sure I wasn’t sitting around having my brain rot, but their spying on what I was doing into my teens left me with some trust issues.
They briefly tried to use technological solutions to control my access and monitor me, but all that served was to make me very good at circumventing them. Outsourcing parenting to a computer program doesn’t work, and kids notice when you try.
They’re not trivializing, just noting that the different things you need to discuss for kernel development compared with other work. It is very different in a lot of ways, and does shape your perspective. I also find it interesting.
For the same reason spoken languages often have semantic structures that make a literal translation often cumbersome and incorrect, translating nontrivial code from one language into another without being a near expert in both langauges, as well as being an expert in the project in question, can lead to differences in behaviour varying from “it crashes and takes down the OS with it”, to “it performs worse”.
A rather overly simplistic view of filesystem design.
More complex data structures are harder to optimise for pretty much all operations, but I’d suggest the overwhelmingly most important metric for performance is development time.
Yes, but note that neither the Linux foundation nor OpenZFS are going to put themselves in legal risk on the word of a stack exchange comment, no matter who it’s from. Even if their legal teams all have no issue, Oracle has a reputation for being litigious and the fact that they haven’t resolved the issue once and for all despite the fact they could suggest they’re keeping the possibility of litigation in their back pocket (regardless of if such a case would have merit).
Canonical has said they don’t think there is an issue and put their money where their mouth was, but they are one of very few to do so.
Opengear in Brisbane; development teams often use Linux.
Brand new anything will not show up with amazing performance, because the primary focus is correctness and features secondary.
Premature optimisation could kill a project’s maintainability; wait a few years. Even then, despite Ken’s optimism I’m not certain we’ll see performance beating a good non-cow filesystem; XFS and EXT4 have been eeking out performance for many years.
License incompatibility is one big reason OpenZFS is not in-tree for Linux, there is plenty of public discussion about this online.
I’ll also tac on that when you use cloud storage, what do you think your stuff is stored on at the end of the day? Sure as shit not Bcachefs yet, but it’s more likely than not on some netapp appliance for the same features that Bcachefs is developing.
In addition to the comment on the mentioned better hardware flexibility, I’ve seen really interesting features like defining compression & deduplication in a granular way, even to the point of having a compression algo when you first write data, and then a different more expensive one when your computer is idle.
This is actually a feature that enterprise SAN solutions have had for a while, being able choose your level of redundancy & performance at a file level is extremely useful for minimising downtime and not replicating ephemeral data.
Most filesystem features are not for the average user who has their data replicated in a cloud service; they’re for businesses where this flexibility saves a lot of money.
Regarding 1: if you open up dmesg after it happens and you see an error regarding “No edid read”, your GPU is having a hard time automatically getting the monitor’s edid over display port. My 7800xt has this issue.
If your monitor setup doesn’t change much, you can manually set the edid on a per output basis. Here is a good guide.
Also, regarding 3: you may need to set your amdgpu feature mask in your kernel parameters.
I remember old Tesla and Firepro drivers had a jank, proprietary alternative to SR-IOV but didn’t think any vendor (except Intel with i915’s GVT-g) had an implementation for their consumer devices.