

There is a theory that Ghislaine Maxwell is/was a Reddit mod.


There is a theory that Ghislaine Maxwell is/was a Reddit mod.


Stupid BAM Broadcom BAM Legacy BAM Wireless BAM Drivers BAM
I just read that recently. Let me see if I can run that source back down.
Edit: All in one CompTIA server plus certification exam guide second edition exam SK0-005 McGraw-Hill Daniel LaChance 2021 Page 138. In the table there it says that SATA is not designing for constant use.
Edit 2:
https://www.hp.com/us-en/shop/tech-takes/sas-vs-sata
Reliability:
SAS: Designed for 24/7 operation with higher >mean time between failures (MTBF), often 1.6 million hours or more SATA: Suitable for regular use but not as robust as SAS for constant, heavy workloads, with MTBF typically around 1.2 million hour
They are saying that SAS is a better option with a longer MTBF, but I don’t expect my drives to last 5 years, much less 136.
My own two cents here is that you probably don’t want to use SATA ZFS JBOD in an enterprise environment, but that’s more based on enterprise lifecycle management than utility.



Here is my combination lab and workbench. I have been busy trying to buy/sell/trade computers that I have become significantly behind on cleaning as I go. I also just got the network rack:

I haven’t had time between work, hustling, and home maintenance to finish getting the cabling managed or the NAS:

The goal is to get the NAS in the rack, UPS to the items in the rack, the 3D printer under the bench, and the monitors on the wall and off the bench. Then I’ll start in on plastic organizers for the bits and parts that clutter my bench.


I only ask because I work in a HP environment and run off a G3 sff myself.


Is that an EliteDesk 800 G3, G4, or G5?
I have been toying with the idea of using USB storage, but my concern is that the controllers are not meant to be used that heavily. Supposedly SATA controllers are also not built for the abuse I have been throwing them in my machines, and I don’t want to push it.


Valvonta means supervision in Finnish. That led me to search for the Finnish word for weekdays, arkisin. The word for weekend, viikonloppu, didn’t make sense here, but Saturday is Lauantai.
I failed to understand the connection to LLMs. Thanks!
How would you transmit information without communicating it? In some ways avoiding communication is a form of communication, along with the way you dress, your body language and EEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEE.


Yeah, I can see that. I’m thinking of streaming assets and code on demand, similar to how an optical disk works. It’s a terrible waste of resources, and they can be grabbed if they are not cryptographically secured.


That would protect the IP, but the response time is terrible. Pinging Google.com I get a response time of about 80 ms. At that delay, everything would feel spongy and laggy.


SAAS. You never install the entire application. Large parts of the engine never run locally.


That makes more sense. I was literally thinking, “We have all at home.”


Ani.social has an all option. It’s not a /c/, more of a firehose.
I knew a flat Earth believer about ten years ago. I’d estimate hundreds at least.


Obviously not.


More to your point, the hypervisor is OS-specific. On my end, it’s more of a skill issue but I have run into issues getting Linux virtual machine tools running outside of VirtualBox. The software was looking for explicit infrastructure I was struggling to provide. I eventually figured out qEMU, but that’s besides the point.
I’m going to counter and say that it might be possible for these solutions to work on Linux, but you will first need to satisfy the hardware requirements. Once you meet that then it’s possible to try with a Windows compatibility layer, but then you are shooting in the dark.
Lastly, while the crackers themselves have a reputation to uphold, they and their software can both be compromised. The article is very clear about the risks. Most people running Linux are not using secure boot, and the hypervisor is an unauditabe container.


People ruin everything.
Hey, I’m not sure where you got your factor of 5 years, but it was a number I pulled out my ass. I’m a repair depot I typically didn’t see drives that live much longer than 17k hours (just under 2 years). That didn’t mean that they always fall at that age, only that systems that came through had about that much time on them max.
Regarding the 136 vs 150 million numbers, those numbers are pure bullshit. MTBF is a raw calculation of how long it will take these devices to fall based on operational runtime over how many failures were experienced in the field. They most likely applied a small number of warranty failures over a massive number of manufacturing runs and projected that it would take that long for about half their drives to fall.
In reality, you will see failure spikes in the lifetime of a product. The initial failures will spike and drop off. I recall reading either the data surrounding this article or something similar when they realized that the bathtub curve may not be the full picture. They just updated it again for numbers from up to last year and you can see that it would be difficult to project an average lifetime of 20 years, much less 150.
My last thought on this is that when Backblaze mentions consumer vs enterprise drives they are possibly discussing SATA vs SAS. This comes from the realization that enterprise workstation drives are still just consumer drives with a part number label on them (seen in Dell and HP Enterprise equipment). Now, they could be referring to more expensive SATA drives, but I can’t imagine that they are using anything but SAS at this point in their lifecycle.