Jason2357, jason2357@lemmy.ca
Instance: lemmy.ca
Joined: 3 years ago
Posts: 0
Comments: 163
Posts and Comments by Jason2357, jason2357@lemmy.ca
Posts by Jason2357, jason2357@lemmy.ca
Comments by Jason2357, jason2357@lemmy.ca
Yes, good guess. I do expect everyone to raise prices owing to the cost of ram and other hardware.
If you are willing to have all your accounts linkable by the signup email, you dont need a temp or relay email service at all. Just sign up for a single random email address somewhere that lets you do it anonymously.
I am happy with addy.io and havnt had any issues. You could maintain a few providers - throw in ddg and mozillas services - so you can try from lots of domains I guess.
My vps is operated by an EU company, running on hardware in Montreal, powered entirely by Quebec’s abundant hydroelectric grid. EU Canada all the way.
Are you a native french speaker? Maybe you heard it differently from me, but while I am all for nuance, lets not sanewash people and take them at their word.
I use plenty of software where the developers are not primarily focused on security, but his line of reasoning sounds just plain dangerous for an OS developer. Maybe he phrased it bad, but that would be up to him to clarify and we shouldnt do that for him.
I usually treat them by using an extremely well established library where someone else has spent the requisite years crying over every stupid edge case of csv reading. Rolling your own csv reader is a bit like encryption. Until someone hands you a file that rejects all sanity and you start fking with regex. Lol.
The delimiter isn’t really the issue. Its that there are lots and lots of weird edge cases that break reading csvs. If you use commas, at minimum, you need to escape commas in the data, or quote strings that might contain commas… But now you have to deal with the possibility of a quote character or your escape character in the data.
Then you have the fact that csvs can be written with so many different character encodings, mangling special characters where they occur.
Aaand then you have all the issues that come with lack of metadata - good formats will at least tell you the type of data in each column so you dont have to guess them.
Lets see, its also really annoying to include any binary data in a csv, theres no redundancy or parity checks to catch currupted data, and they arent compressed so you need to tack on compression if you want efficient storage, but that means you always have to read the whole csv file for any task.
Oh, that brings me to the joys of modern columnar formats where you can read selected columns super fast without reading the whole file.
Oh god, I really kept going there. Sorry. Its been a year.
It was some sort of weird database frontend the contractor used. It was very limited.
Honestly, dont take anyones recommendation. It takes 10 minutes to create a bootable USB for a Linux distro once you get the hang of it. Try a handful of different “easy” distros and desktops on a Saturday morning and pick one that seems to work well on your computer and that you find you like. What you find intuitive isnt necessarily good for another, etc. A little time invested in shopping will pay off later (which is true for a lot of things).
I’m not a fan, but thats extreme. The Ubuntu desktop will boot to the DE on half a gig of ram, and can open basic desktop apps with 1 or 2. Its the websites, containered apps, and more complex applications that Ubuntu is worried about UX disappointment from naive users (which is their segment). Windows 11 requires many times that just to get to a desktop and open a text file in notepad. They are not the same.
All of the default software that comes with the Ubuntu desktop will run reasonably well with 2Gb. Its the websites and electon apps (i.e., websites) that will make it swap. That and modern users that want to keep dozens of programs or websites open -which users 10 or 20 years ago may have known not to do.
Which is exactly what Ubuntu is doing. The desktop and even most native desktop applications that come with it will run just fine with 1 or 2GB of ram. If you used it like a 90s computer for 90s computer tasks, it will work fine.
In practice, however, users will open a web browser to some “modern” websites or a couple electron apps and have a very bad experience.
P.s. in the above quagmire, the only solution is choose to keep only the most important un-clean column per csv, and make it the last column in the file so you have predictable columns. If you need more, then write separate csvs. Computers are stupid.
God I hate csv with the fire of a thousand suns.
Contractors never seem to know how to write them correctly. Last year, one even provided “csv”s that were just Oracle error messages. lol. Another told me their system could not quote string columns nor escape commas or use anything but commas as their separator, so there were unpredictable numbers of commas in the rows when the actual data contained commas. Total nightmare. And so much of my data has special character issues because somewhere in the pipeline a text encoding was wrong and there is exactly one mangled character in 5 million lines for me to find.
Give me the data as closely to the source data as you can. If it is a database, then a database dump or access to a clone of your database is the best option by far. I don’t care how obscure your shit is, Ill do the conversion myself.
For intermediate data, something like parquet or language specific formats like Rdata or pickle files. Maaaaybe very carefully created csv files for archival purposes, but even then, I think parquet is safe for the long haul nowadays.
I will have to yield to your experience then. I mainly thought of it as a naive type of sensible argument, given people were not all that concerned about tracking and particularly browser fingerprinting. I guess back then, the main thing was web developers who used flash needed to check for it. But those people were anti-open web back then and deserved to be ignored by the browser makers.
I am guessing you were strongly in the open web camp back then. I am glad we sort of won that particular battle, even if we lost so many others.
There are actual open-source alternatives if you want to do more than just inspect “code available” claude code. For example, https://github.com/Aider-AI/aider
Now that would be an interesting development for them to actually argue - which they wont.
I am surprised that Github even allows you to create a public fork of a repo with an all-rights-reserved License.md. I would have actually expected that UI element to be disabled or restrict it to a private fork.
You want open source, people, work on open source. “Code available” software is better than fully proprietary, sure, you can inspect it if you want, but ffs, don’t give them your development time.
^ comments like this are why Lemmy is great.
That might have been a sensible argument 20 years ago. Mozilla has spent the last 5 or so slowly stripping most of that out for “anti-fingerprinting” without breaking website layout.
Yeah, the description is misleading because anyone reading it is thinking desktop software. But… hear me out. I know that all the surveillance capitalism companies do this, but in this case it is literally pairing what is mostly corporate IT policy data (browser, hardware, OS, and extensions) with employee name, title, and employer. That does technically fit the definition of corporate espionage, and I am always open to getting more people -especially people with some levers in government- onto “our side” of the Internet privacy conflict.
Yes, good guess. I do expect everyone to raise prices owing to the cost of ram and other hardware.
If you are willing to have all your accounts linkable by the signup email, you dont need a temp or relay email service at all. Just sign up for a single random email address somewhere that lets you do it anonymously.
I am happy with addy.io and havnt had any issues. You could maintain a few providers - throw in ddg and mozillas services - so you can try from lots of domains I guess.
My vps is operated by an EU company, running on hardware in Montreal, powered entirely by Quebec’s abundant hydroelectric grid. EU Canada all the way.
Are you a native french speaker? Maybe you heard it differently from me, but while I am all for nuance, lets not sanewash people and take them at their word.
I use plenty of software where the developers are not primarily focused on security, but his line of reasoning sounds just plain dangerous for an OS developer. Maybe he phrased it bad, but that would be up to him to clarify and we shouldnt do that for him.
I usually treat them by using an extremely well established library where someone else has spent the requisite years crying over every stupid edge case of csv reading. Rolling your own csv reader is a bit like encryption. Until someone hands you a file that rejects all sanity and you start fking with regex. Lol.
The delimiter isn’t really the issue. Its that there are lots and lots of weird edge cases that break reading csvs. If you use commas, at minimum, you need to escape commas in the data, or quote strings that might contain commas… But now you have to deal with the possibility of a quote character or your escape character in the data.
Then you have the fact that csvs can be written with so many different character encodings, mangling special characters where they occur.
Aaand then you have all the issues that come with lack of metadata - good formats will at least tell you the type of data in each column so you dont have to guess them.
Lets see, its also really annoying to include any binary data in a csv, theres no redundancy or parity checks to catch currupted data, and they arent compressed so you need to tack on compression if you want efficient storage, but that means you always have to read the whole csv file for any task.
Oh, that brings me to the joys of modern columnar formats where you can read selected columns super fast without reading the whole file.
Oh god, I really kept going there. Sorry. Its been a year.
It was some sort of weird database frontend the contractor used. It was very limited.
Honestly, dont take anyones recommendation. It takes 10 minutes to create a bootable USB for a Linux distro once you get the hang of it. Try a handful of different “easy” distros and desktops on a Saturday morning and pick one that seems to work well on your computer and that you find you like. What you find intuitive isnt necessarily good for another, etc. A little time invested in shopping will pay off later (which is true for a lot of things).
I’m not a fan, but thats extreme. The Ubuntu desktop will boot to the DE on half a gig of ram, and can open basic desktop apps with 1 or 2. Its the websites, containered apps, and more complex applications that Ubuntu is worried about UX disappointment from naive users (which is their segment). Windows 11 requires many times that just to get to a desktop and open a text file in notepad. They are not the same.
All of the default software that comes with the Ubuntu desktop will run reasonably well with 2Gb. Its the websites and electon apps (i.e., websites) that will make it swap. That and modern users that want to keep dozens of programs or websites open -which users 10 or 20 years ago may have known not to do.
Which is exactly what Ubuntu is doing. The desktop and even most native desktop applications that come with it will run just fine with 1 or 2GB of ram. If you used it like a 90s computer for 90s computer tasks, it will work fine.
In practice, however, users will open a web browser to some “modern” websites or a couple electron apps and have a very bad experience.
P.s. in the above quagmire, the only solution is choose to keep only the most important un-clean column per csv, and make it the last column in the file so you have predictable columns. If you need more, then write separate csvs. Computers are stupid.
God I hate csv with the fire of a thousand suns.
Contractors never seem to know how to write them correctly. Last year, one even provided “csv”s that were just Oracle error messages. lol. Another told me their system could not quote string columns nor escape commas or use anything but commas as their separator, so there were unpredictable numbers of commas in the rows when the actual data contained commas. Total nightmare. And so much of my data has special character issues because somewhere in the pipeline a text encoding was wrong and there is exactly one mangled character in 5 million lines for me to find.
Give me the data as closely to the source data as you can. If it is a database, then a database dump or access to a clone of your database is the best option by far. I don’t care how obscure your shit is, Ill do the conversion myself.
For intermediate data, something like parquet or language specific formats like Rdata or pickle files. Maaaaybe very carefully created csv files for archival purposes, but even then, I think parquet is safe for the long haul nowadays.
I will have to yield to your experience then. I mainly thought of it as a naive type of sensible argument, given people were not all that concerned about tracking and particularly browser fingerprinting. I guess back then, the main thing was web developers who used flash needed to check for it. But those people were anti-open web back then and deserved to be ignored by the browser makers.
I am guessing you were strongly in the open web camp back then. I am glad we sort of won that particular battle, even if we lost so many others.
There are actual open-source alternatives if you want to do more than just inspect “code available” claude code. For example, https://github.com/Aider-AI/aider
Now that would be an interesting development for them to actually argue - which they wont.
I am surprised that Github even allows you to create a public fork of a repo with an all-rights-reserved License.md. I would have actually expected that UI element to be disabled or restrict it to a private fork.
You want open source, people, work on open source. “Code available” software is better than fully proprietary, sure, you can inspect it if you want, but ffs, don’t give them your development time.
^ comments like this are why Lemmy is great.
That might have been a sensible argument 20 years ago. Mozilla has spent the last 5 or so slowly stripping most of that out for “anti-fingerprinting” without breaking website layout.
Yeah, the description is misleading because anyone reading it is thinking desktop software. But… hear me out. I know that all the surveillance capitalism companies do this, but in this case it is literally pairing what is mostly corporate IT policy data (browser, hardware, OS, and extensions) with employee name, title, and employer. That does technically fit the definition of corporate espionage, and I am always open to getting more people -especially people with some levers in government- onto “our side” of the Internet privacy conflict.