Google ads are full of crime and most web users should block
them. If you don’t believe the FBI, or Malwarebytes,
believe Google. Their 2024
Ads Safety Report is out (Search
Engine Land covered it) and things do not look good. The report is
an excellent example of some of the techniques that big companies use to
misrepresent an ongoing disaster as somehow improving, so I might as
well list them. If I had to do a corporate
misinformation training session, I’d save this PDF for a reading
assignment.
release bad news when other news is happening This
was a big news week for Google, which made it the best time to release
this embarrassing report. Editors aren’t going to put their Google
reporter to work on an ad safety story when there’s big news
from the Federal courthouse.
counting meaningless numbers Somehow our culture
teaches us to love to count,
so Google gives us a meaningless number when the meaningful numbers
would look crappy.
Last year, we continued to invest heavily in making our LLMs more
advanced than ever, launching over 50 enhancements to our models which
enabled more efficient and precise enforcement at scale.
The claim is that Google continued to invest heavily and
that’s the kind of statement that’s relatively easy to back up with a
number that has meaningful units attached. Currency
units, head count, time units, even lines
of code. Instead, the count is enhancements which could be
almost anything. Rebuild an existing package with different compiler
optimizations? Feed an additional data file to some ML system? What this
looks like from the outside is that the meaningful numbers are going in
the wrong direction (maybe some of the people
who would have made them go up aren’t there any more?) so they
decided to put out a meaningless number instead.
control the denominator to juice the ratio Only
takes elementary school math to spot this, but might be easy to miss if
you’re skimming.
Our AI-powered models contributed to the detection and enforcement of
97% of the pages we took action on last year.
Wow, 97%, that’s a big number. But it’s out of pages we took
action on which is totally under Google’s control. There are a bunch
of possible meaningful ratios to report here, like
(AI-flagged ads)/(total ads)
(ads removed)/(AI-flagged ads)
(bad ad impressions)/(total ad impressions)
and those could have been reported as a percentage, but it looks like
they wanted to go for the big number.
pretend something that’s not working is working The
AI models contributed to 97% of the actions, but
contributed isn’t defined. Does it count as contributed
if, say, human reviewers flagged 1,000 ads, the AI flagged
100,000 ads, and 970 ads were flagged by both? If AI were flagging ads
that had been missed by other methods, this would have been the place to
put it. The newsworthy claim that’s missing is the count of bad ads
first detected by AI before getting caught by a human reviewer.
Contributed to the detection could be a lot of things. (If this
were a report on a free trial of an AI-based abuse detection service,
contributed wouldn’t get me to upgrade to the paid plan.)
report the number caught, not the number that get
through Numbers of abusers caught is always the easiest number
to juice. The simplest version is to go home at lunch hour, code up the
world’s weakest bot, start it running from a non-work IP address, then
go back to work and report some impressive numbers.
To put this into perspective: we suspended over 39.2 million accounts
in total, the vast majority of which were suspended before they ever
served an ad.
Are any employees given target numbers of suspensions to issue? Can
anyone nail their OKRs by raising the number of accounts suspended? If
this number is unreliable enough that a company wouldn’t use it for
management, it’s not reliable enough to pay attention to. They’re also
reporting the number of accounts, not individuals or companies. If some
noob wannabe scammer writes a script to POST the new account form a
million times, do they count for a million?
don’t compare to last year Here’s the graph of bad
ads caught by Google in 2024.
The total number isn’t as interesting as the individual, really
problematic categories. The number caught for enabling dishonest
behavior went down from about 20 million in 2023 to under 9 million
in 2024.
Did the number of attempts at dishonest behavior with Google ads
really go down by more than half in a single year? Or did Google catch
fewer of them? From the outside, it’s fairly easy to tell that Google
Ads is understaffed and the remaining employees are in the weeds,
but it’s hard to quantify the problem. What’s really compelling about
this report is that the staffing situation has gotten bad enough that
it’s even showing up in Google’s own hand-picked numbers. In general
when a report doesn’t include how a number has changed since the last
report, the number went in the wrong direction and there’s no good
explanation for why. And the number of ads blocked or removed for
misinformation went from 30 million in 2023 to (checks notes)
zero in 2024. Yes, misinformation
has friends in high places now, but did all of the sites worldwide
that run Google ads just go from not wanting to run misinformation to
being fine with it?
report detection, not consequences Those numbers on
bad ads are interesting, but pay attention to the text. These are
numbers for ads blocked or removed, and repeat
offenders drive the bulk of tech support scams via Google Ads. Does
an advertiser caught doing misrepresentation in one ad get to
keep going with different ads?
don’t compare to last year, part 2 The previous two
graphs showed Google’s bad ads/good site problem, so here’s how they’re
doing on their good ad/bad site problem. Here’s 2024:
1.3 billion pages taken action against in
2024
And 2023:
2.1 billion pages taken action against in
2023
Ad-supported AI
slop is on the way up everywhere, making problem pages easier to
create at scale, but Google somehow caught 800 million fewer pages than
in 2023. How many pages they took action against isn’t even a good
metric (and I would be surprised if anyone is incentivized based on it).
Some more useful numbers would be stuff like
What percentage of advertisers had their ad run on a page that
later had action taken against it?
How much money was paid out to sites that were later removed for
violating the law or Google policy?
A real Ad Safety Report would help an advertiser answer
questions about how likely they are to sponsor illegal content when they
buy Google ads. And it would help a publisher understand how likely they
are to have an ad for malware show up on their pages. No help from this
report. Even though from the outside we can see that Google runs a bunch of
ads on copyright-infringing sites, not only does Google not report
the most meaningful numbers, they’re doing worse than before on the less
meaningful numbers they do choose to report.
Google employees, (yes, both FTEs and TVCs) are doing a
lot of good work trying to do the right thing on the whole ads/crime
problem, but management just isn’t staffing and funding the ad safety
stuff at the level it needs. A company with real competition would have
had to straighten this situation out by now, but that’s not
the case for Google. Google’s services like Search are both free and
overpriced—users don’t pay in money, but in over-exposure to fraud and
malware risks that would be lower in a competitive market. If a future
Google
breakup works, one of the best indicators of success will be more
meaningful, and more improved, metrics in future ad safety reports.
Flaming
Fame. by George Tannenbaum. We don’t see shitty work and say
that’s shitty. It’s worse than that. We simply don’t see it at
all.
LG
TVs’ integrated ads get more personal with tech that analyzes viewer
emotions by Scharon Harding. The company plans to incorporate a
partner company’s AI tech into its TV software in order to interpret
psychological factors impacting a viewer, such as personal interests,
personality traits, and lifestyle choices. (What happens when you do
a Right to Know for the family TV?)
Former
Substack creators say they’re earning more on new platforms that offer
larger shares of subscription revenue by Alexander Lee. Since
leaving Substack, some writers’ subscriber counts have plateaued over
the past year, while others have risen — but in both cases, creators
said that their share of revenue has increased because Ghost and Beehiiv
charge creators flat monthly rates that scale based on their subscriber
counts, rather than Substack’s 10 percent cut of all transaction
fees.
The
Mediocrity of Modern Google by Om Malik. What’s particularly
ironic is that today’s Google has become exactly what its founders
warned against in their 1998 paper: an advertising company whose
business model fundamentally conflicts with serving users’
needs.
Git
turns 20: A Q&A with Linus Torvalds by Taylor Blau. So I was
like, okay, I’ll do something that works for me, and I won’t care about
anybody else. And really that showed in the first few months and
years—people were complaining that it was kind of hard to use, not
intuitive enough. And then something happened, like there was a switch
that was thrown.
I’m not an expert on electric cars, so I don’t know enough to
criticize some of the hard parts of the design of a Tesla. But when they
get obvious
stuff like getting out without power wrong, that’s a pretty good
sign to stay away.
How
the U.S. Became A Science Superpower by Steve Blank. Post war, it
meant Britain’s early lead was ephemeral while the U.S. built the
foundation for a science and technology innovation ecosystem that led
the world—until now.
Here’s another privacy paradox for people who collect them.
On the web, the average personalized ad is probably better than
the average non-personalized ad. (The same ad campaigns that have a
decent budget for ad creative also have a budget for targeting
data.)
But users who block personalized ads, or avoid personalization by
using privacy tools and settings, are, on average, better
off than users who get personalized ads.
There’s an expression in finance: Picking
Up Nickels In Front Of A Steam Roller. For some kinds of investing
decisions, the investor is more likely to make a small gain than to lose
money in each individual trade. But the total expected return over time
is negative, because a large loss is an unlikely outcome of each trade.
The decision to accept personalized ads or try to avoid them might be a
similar bet.
For example, a typical positive outcome of getting personalized ads
might be getting better shoes, cheaper. There’s a company in China that
is working the personalized ad system really well. Instead of paying for
high production value ads featuring high-profile athletes in the USA,
they’re just doing the incremental data-driven marketing thing. Make
shoes, experiment with the personalized ad system, watch the numbers,
reinvest in both shoe improvements and improvements to the personalized
ads. For customers, the shoe company represents the best-case scenario
for turning on the personalized ads. You get a pair of shoes from China
for $40 that are about as good as the $150 shoes from China that you
would get from a a big-name brand. (The shoes might even be made by the
same people out of the same materials.) I don’t need to link to the
company, just turn on personalized ads and if you want the shoes they’ll
find you.
That example might be an outlier on the win-win side, though. On
average, personalized (behaviorally targeted) ads are likely
to be associated with lower quality vendors and higher product prices
compared to competing alternatives found among search results. (Mustri
et al.) but let’s pretend for a minute and say you figured out how
to get targeted in the best possible way and come out on the winning
side. That’s pretty sweet—personalized ads save you more than a
hundred bucks on shoes, right?
Here comes the steamroller, though.
In recent news, Baltimore
sues 2 sportsbooks over alleged exploitative practices. Some people
are likely to develop a gambling problem, and if you don’t know in
advance whether or not you’re one of them, should you have the
personalized ads turned on? You stand to lose a lot more than you would
have gained by getting the cheap shoes or other miscellaneous stuff. It
is possible that machine learning on the advertising or recommended
content side could know more about you than you do, and the negative
outcomes from falling for an online
elder fraud scheme tend to be much larger than the positive outcomes
from selecting the best of competing legitimate products.
People’s reactions to personalization are worth watching, and reflect
more widely held understanding of how information works in markets than
personalized ad fandom does. The fact that Google may have used this
data to conduct focused ad campaigns targeted back to you was disclosed as if
it was a security issue, which makes sense. Greg Knauss writes,
Blue Shield says that no bad actor was involved, but is that
really true? Shouldn’t a product that, apparently by default, takes
literally anything it can—privacy be damned—and tosses it into the old
ad-o-matic not be considered the output of a bad actor? Many people
(but not everybody)
consider being targeted for a personalized ad as a threat in itself.
More:personalization
risks
Bonus links
What If We Made Advertising
Illegal? by Kōdō Simone. The traditional argument
pro-advertising—that it provides consumers with necessary
information—hasn’t been valid for decades. In our information-saturated
world, ads manipulate, but they don’t inform. The modern advertising
apparatus exists to bypass rational thought and trigger emotional
responses that lead to purchasing decisions. A sophisticated machine
designed to short-circuit your agency, normalized to the point of
invisibility. (Personally I think it would be hard to come up with a
law that would squeeze out all incentivized communication intended to
cause some person to purchase some good or service, but it would be
possible to regulate the information flows in the other
direction—surveillance of audience by advertiser and intermediaries—in a
way that would mostly eliminate surveillance advertising as we know it:
Big Tech
platforms: mall, newspaper, or something else?)
Meta
secretly helped China advance AI, ex-Facebooker will tell Congress
by Ashley Belanger. In her prepared remarks, which will be delivered
at a Senate subcommittee on crime and counterterrorism hearing this
afternoon, Wynn-Williams accused Meta of working hand in glove
with the Chinese Communist Party (CCP). That partnership allegedly
included efforts to construct and test custom-built censorship tools
that silenced and censored their critics as well as provide the CCP
with access to Meta user data—including that of Americans.
(And if they’re willing to do that, then the elder fraud ads on Facebook
are just business as usual.)
Protecting
Privacy, Empowering Small Business: A Path Forward with S.71 (A
privacy law with private right of action gets enforced based on what
makes sense to normal people in a jury box, not to bureaucrats who think
it’s normal to read too many PDFs. Small businesses are a lot better off
with this common-sense approach instead of having to feed the
compliance monster.)
This
startup just hit a big milestone for green steel production by Casey
Crownhart. Boston Metal uses electricity in a process called molten
oxide electrolysis (MOE). Iron ore gets loaded into a reactor, mixed
with other ingredients, and then electricity is run through it, heating
the mixture to around 1,600 °C (2,900 °F) and driving the reactions
needed to make iron. That iron can then be turned into steel. Crucially
for the climate, this process emits oxygen rather than carbon
dioxide…
There are two kinds of PDFs. Some have real embedded text that you
can select in a PDF reader, and some are just images.
The second kind is what I sometimes get in response to a CCPA/CPRA
Right to Know. Some companies, for whatever reason, want to make it
harder to do automated processing of multiple RtKs. This should make
privacy researchers more likely to look at them, because what are they
hiding and they must be up to something.
But the PDF still needs to get run through some kind of OCR. Tesseract OCR has been
giving me pretty good results, but it needs to be fed images, not
PDFs.
So I have been feeding the PDFs to pdf2image—in
Python code, and then passing the images to Tesseract. But it turns out
that Tessaract works a lot better with higher resolution images, and the
default for pdf2image is 200 DPI. So I’m gettting a lot more accurate
OCR by making the images oversized with the dpi named
parameter:
I might tweak this and try 300 DPI, or also try passing
grayscale=True to preserve more information. Some other
approaches to try next, if I need them.
Anyway, Meta
(Facebook) made some of their info easy to parse (in JSON format)
and got some of us to do research on them. Some of the other interesting
companies, though, are going to be those who put in the time to
obfuscate their responses to RtKs.
Related
OCRmyPDF is an
all-in-one tool that adds a text layer to the PDF. Uses Tessaract
internally. When possible, inserts OCR information as a “lossless”
operation without disrupting any other content. Thanks to Gaurav Ujjwal for the link. (I’m
doing an OCR step as part of ingesting PDFs into a database, so I don’t
need to see the text, but this could be good for PDFs that you actually
want to read and not just do aggregated reporting on.)
Example of where
GDPR compliance doesn’t get you CCPA compliance: This is the mistake
that Honda
recently made. CCPA/CPRA is not just a subset of GDPR. GDPR allows a
company to verify an objection to processing, but CCPA does not allow a
company to verify an opt out of sale. (IMHO the EU should harmonize by
adopting the California good-faith, reasonable, and documented belief
that a request to opt-out is fraudulent standard for objections to
processing.)
New
Report: Many Companies May Be Ignoring Opt-Out Requests Under State
Privacy Laws - Innovation at Consumer ReportsThe study examined
40 online retailers and found that many of them appear to be ignoring
opt-out requests under state privacy laws. (A lot more companies are
required to comply with CCPA/CPRA than there are qualified compliance
managers. Even if companies fix some of the obvious problems identified
in this new CR report, there are still a bunch of data transfers that
are obvious detectable violations if a GPC flag wasn’t correctly set for
a user in the CRM system. You can’t just fix the cookie—GPC also has to
cover downstream usage such as custom audiences and server-to-server
APIs.)
Bonus links
EU
may “make an example of X” by issuing $1 billion fine to Musk’s social
network by Jon Brodkin at Ars Technica. (A lot of countries don’t
need to raise their oen tariffs in order to retaliate against the USA’s
tariffs. They just need to stop letting US companies slide when they
violate laws over there. If they can’t rely on the USA for regional
security, there’s no reason not to. Related: US
Cloud soon illegal? at noyb.eu)
Big
Tech Backed Trump for Acceleration. They Got a Decel President
Instead by Emanuel Maiberg and Jason Koebler at 404 Media. Unless
Trump folds, the tariffs will make the price of everything go up.
Unemployment will go up. People will buy less stuff, and companies will
spend less money on advertising that powers tech platforms. The tech
industry, which has thrived on the cheap labor, cheap parts, cheap
manufacturing, and supply chains enabled by free and cheap international
trade, will now have artificial costs and bureaucracy tacked onto all of
this. The market knows this, which is why tech stocks are eating
shit. (Welcome to the weak men create hard times phase—but
last time we had one of these, the dismal Microsoft monopoly days are
when we got the web
and Linux scenes that evolved into today’s Big Tech. Whatever
emerges from the high-unemployment, import-denied generation, it’s going
to surprise us.)
Signal
sees its downloads double after scandal by Sarah Perez on
TechCrunch. Appfigures chalks up the doubling of downloads to the old
adage all press is good press, as the scandal increased Signal’s
visibility and likely introduced the app to thousands of users for the
first time. (Signal is also, according
to traders on Manifold Markets, the e2e messaging program least
likely to provide message content to US law enforcement. Both Apple, the
owner of iMessage, and Meta, the owner of WhatsApp, have other
businesses that governments can lean on in order to get cooperation.
Signal just has e2e software and reputation, so fewer points of
leverage.)
YouTube
removes ‘gender identity’ from hate speech policy by Taylor Lorenz
(In the medium term, a lot of the moderation changes at Big Tech are
going to turn into a recruiting challenge for hiring managers in
marketing departments. If an expected part of working in marketing is
going to be mandatory involvement in sending money to weird, creepy
right-wing dudes, that means you’re mostly going to get to hire…weird,
creepy right-wing dudes.) Related: slop
capitalism and dead internet theory by Adam Aleksic. Our best way
of fighting back? Spend as little time on algorithmic media as possible,
strengthen our social ties, and gather information from many different
sources—remembering that the platforms are the real enemy.
Some good news last week: Meta
settles UK ‘right to object to ad-tracking’ lawsuit by agreeing not to
track plaintiff. Tanya O’Carroll, in the UK, has settled a case with
Meta, and the company must stop using her data for ad targeting when
she uses its services. It’s not a change for everyone, though, since
the settlement is just for one person. O’Carroll said she is unable
to disclose full details of the tracking-free access Meta will be
providing in her case but she confirmed that she will not have to pay
Meta.
The Open Rights Group now has a Meta opt-out
page that anyone in the UK can use to do an opt out under the UK
GDPR.
If you use any Meta products – Facebook, Instagram, Meta Quest or VR,
Threads or WhatsApp – you can use our tool to request that they no
longer collect or process your data for advertising. This is known as
your right to object, which is enshrined in data protection law. Meta
had tried to get around GDPR, but by settling Tanya’s case they have
admitted that they need to give their users this right.
If you’re in the UK, you can either use the form on the site, or use
the mailto link to open up a new regular email from your own account
pre-populated with the opt out text. This is a win not just because it
could mean less money for a
transnational criminal organization and more money staying in the
UK, but also because it’s going to mean better products and services for
the people who do it.
Generate tracking data that is hard to link to you
Set an opt out while doing the surveilled activity
Send an opt out or Right to Delete after doing the
surveilled activity
Having access to this new tool doesn’t mean not to do the others.
Even if I could figure out how to use the Meta apps in a way that’s
totally safe for me, it’s still a win to switch away because it helps
build network effects for the alternatives and more safety for other
people. So even if you do this opt out, it’s also a good idea to do the
other effective
privacy tips.
The personalization of ads on Facebook helps vendors of crappy,
misrepresented goods match their products to the shoppers who are most
likely to fall for their bullshit. Yes, you can follow the advice in
articles like Don’t
Get Scammed! Tips For Spotting AI-Generated Fake Products Online on
Bellingcat, but it’s a time-saver and an extra layer of protection not
to get the scam ad in the first place.
Why
We Need Shortwave 2.0 by Kim Andrew Elliott on Radio World.
Because Shortwave Radiogram is transmitted on a regular
amplitude-modulated shortwave transmitter, it can be received on any
shortwave radio, from inexpensive portable radios with no sideband
capability, to more elaborate communications receivers, amateur
transceivers (most of which nowadays have general coverage receivers),
and software defined radios (SDRs). (Then you need a program to
convert the encoded signal into text and/or images—or this functionality
could be built into future inexpensive radios.)
I’m still waiting for my copy of Careless
People by Sarah Wynn-Williams, so I don’t have anything more
on the content of the book than what I have seen in other reviews. The
local bookstore had a stack—more than they normally order for new
hardcovers—but I hesitated and they were gone next time I went in there.
So yes, I am a little behind on this.
But come on, people.
Does anyone seriously think that Meta execs don’t understand the
Streisand
effect?
But Big Tech’s apparent eagerness to appear in bad news makes sense
when you look at the results. Out of all the people who read and were
outraged by Careless People over the weekend, how many are
going to come in to work on Monday and delete their Meta tracking pixel
or turn off Meta CAPI? And how many people concerned about Google’s
malware, CSAM, and infringing content
problems are going to switch to inclusion lists and validated
SupplyChain objects and stop with the crappy,
often illegal ad placements that Google recommends and legit ad agencies
don’t? For Big Tech, doing crimes in an obvious way is a power move,
a credible, costly signal. If there were a Meta alternative that didn’t
do genocide,
or an honest alternative to Google search advertising, then advertising
decision-makers would have switched to them already. All these
embarrassing-looking stories are a signal: don’t waste your time looking
for an alternative to paying us. The publisher’s
page for Careless People has a Meta pixel on it.
I do have a book recommendation that might be a little easier to get
a hold of. Codes
of the Underworld by Diego Gambetta was the weekly
book recommendation on A Collection of Unmitigated
Pedantry. I’m glad to see that it is still in print, because it’s
a useful way to help understand the Big Tech companies. Actions that
might not have made sense in a company’s old create more
value than you capture days are likely to be easier to figure out
after understanding the considerations applied by other criminal
organizations.
Criminals have hard-to-satisfy communications needs, such as the need
to convey a credible threat to a victim without attracting the attention
of enforcers. This is related to the signaling
problem faced by honest advertisers, but in reverse. How can a
representative of a protection racket indicate to a small business that
they represent a true threat, and aren’t just bluffing? Gambetta digs
into a variety of signaling problems. It’s a 2009 book, so many of the
Big Tech firms were still
legit when it came out, but a lot of the communications methods from
back then apply to the companies of today.
Is there a solution? As Gambetta points out, real-life organized
crime perpetrators tend to copy from the movies, and today they’re
copying the partnership with a
friendly government subplot from The Godfather
Part II. Maybe it’s time to watch that movie again.
Related
Update 12 Apr 2025: tante/Jürgen Geuter makees a similar point, in Vulgar
Display of Power. It is a display of power: You as an artist, an
animator, an illustrator, a writer, any creative person are powerless.
We will take what we want and do what we want. Because we can.
imho AI-generated images used to
illustrate a blog post (and not specifically to discuss AI images)
usually send a louder message than the writing does. Gareth Watkins: AI:
The New Aesthetics of Fascism
Privacy-Respecting
European Tech Alternatives by Jonah Aragon. [T]he United States
certainly does not have a monopoly on the best technologies, and many of
our favorite recommended tools come from Europe and all over the world.
Tools from the European Union also generally benefit from much stronger
data protection laws, thanks to the EU’s General Data Protection
Regulation (GDPR). Related: But
how to get to that European cloud?
Please
stop externalizing your costs directly into my face by Drew DeVault.
Whether it’s cryptocurrency scammers mining with FOSS compute
resources or Google engineers too lazy to design their software properly
or Silicon Valley ripping off all the data they can get their hands on
at everyone else’s expense…
But the good news for slacker states is that doing the most work,
cranking out the most lines of code, or the most pages of PDFs, or
whatever, does not necessarily produce the best results. Given the
amount of work that other states, and juridictions like the European
Union, have already done on privacy, a slacker state can, right now, get
not just the best privacy protection but also save a lot of time and
grief for state employees and for business people in your state.
You need two laws. And we know that people are going to print them
out, so please keep them short. (Maybe do a printer ink right to refill
law next year?)
Focus on the riskiest companies with the most money and staff for
compliance—don’t put extra work on small local businesses.
Save your state’s attorney general and their staff a bunch of
time. They’re not Big
Tech’s support department. If a Big Tech company drops the ball on
user support, just suspend their surveillance license until they clean
up their act, like a problem bar and their liquor license.
You can define surveillance really briefly in the law and make
the big out-of-state companies do the work of describing their
surveillance practices in their license application.
That one is pretty easy to do as long as you focus purely on inbound
data, the surveillance part, and don’t touch anything that sounds like
speech from the company to others. And you can push most of the
work off onto Big Tech and a new surveillance licensing board. I’m sure
every state has people who would be willing to get on one of those.
Second, copy all the details from other states and
countries. The other law would be focused on maximum privacy,
minimum effort. The goal is to make a law that small business people can
comply with, without even reading it, because they already had to do
some privacy thing for somewhere else. Two parts.
Any privacy feature offered in some other jurisdiction must be
offered here, too. A company only breaks the law if someone out-of-state
gets a privacy feature that someone in-state doesn’t.
This law may be enforced by anyone except a state employee.
(Borrow the Texas S.B. 8 legal hack, to protect yourself from Big Tech
industry groups trying to block the law by starting an expensive
case.)
A small business that operates purely locally can just do their
thing. But if they already have some your California privacy
rights feature or whatever, they just turn it on for this state too.
Easier compliance project for the companies, better privacy for the
users, no enforcement effort for the state, it’s a win-win-win. After
all, state legislators don’t get paid by the page, and we each only get
one set of carpal tunnels.
Meta,
Apparently, Really Wants Everyone To Read This Book (By Trying To Ban
It) by Mike Masnick. Macmillan showed up just long enough to
point out the blazingly obvious: they never signed any agreement with
Meta and thus can’t be bound by arbitration. The arbitrator, displaying
basic common sense, had to admit they had no jurisdiction over
Macmillan.
AI
Search Has A Citation Problem by Klaudia Jaźwińska and Aisvarya
Chandrasekar. Chatbots were generally bad at declining to answer
questions they couldn’t answer accurately, offering incorrect or
speculative answers instead. (related: fix Google
Search)
How
Ukraine integrates machine vision in battlefield drones by Oleksandr
Matviienko, Bohdan Miroshnychenko & Zoriana Semenovych. In
November 2024, the government procured 3,000 FPV drones with machine
vision and targeting technologies. Reports also suggested that the
procurement would be expanded to 10,000 units.
Preparing
for the next European war by Azeem Azhar. One challenge will be
the simple rate of innovation in the actual battlefield. Drone warfare
in Ukraine has shown iteration cycles measuring weeks not years. So any
systems procured today need to be future-proofed for those
dynamics.
Thread
by Trent TelenkoThe logistical facts are that the FM-MAG machine
gun, the 60 mm & 81mm mortars, LAWS, Javelins, any infantry crew
served weapon you care to name are all going to be most to fully
replaced with drones and drone operators, because of the logistical
leverage drones represent on the battlefield.
Long-range
drone strikes weakening Russia’s combat ability, senior Ukrainian
commander says by Deborah Haynes. Some of the drones are remotely
piloted, others work via autopilot. Russia’s war has forced Ukraine to
use technology and innovation to fight back against its far more
powerful foe. It has accelerated the use of autonomous machines in an
irreversible transformation of the warzone that everyone is watching and
learning from. Brigadier Shchygol said: Right now, Ukraine’s
battlefield experience is essentially a manual for the
world.
Ukraine
Drives Next Gen Robotic Warfare by Mick Ryan. Another more
interesting trend has arisen which will force policy makers and military
strategists to undertake an even more careful analysis of Ukraine war
trends, and how these trends apply in other theatres, particularly the
Pacific. This trend, robotic teaming, has emerged over the past year
with the advent on drone-on-drone combat in the air and on the ground.
In particular, several recent combat actions in Ukraine provide insights
that need to be studied and translated for their employment in the
massive ocean expanses, tens of thousands of kilometres of littoral,
thousands of large and small islands and at least three continents that
constitute the Pacific theatre.
DEEP
DIVE: Taiwan miltech aims to undermine Chinese components by Tim
Mak. Taiwan has learnt the central tech lesson from the war in
Ukraine: the next global conflicts will heavily feature cheap, small
drones—and in large numbers. So as an electronics and hardware component
giant—especially relative to its size and diplomatic status—it is trying
not only to develop a domestic industry, but also become an arsenal for
the free world, building drones and devices for allied militaries
worldwide.
Why
America fell behind in drones, and how to catch up again by Cat
Orman and Jason Lu. Also Building
Drones for Developers: A uniquely open architecture on the F-11
means that every part of the drone is truly built around the
[NVIDIAn] Orin [GPU]. This enables sophisticated autonomy applications
in which ML models are able to not only analyze data obtained in-flight,
but actually use that analysis to inform flight actions in real
time.
switching.software offers
Ethical, easy-to-use and privacy-conscious alternatives to well-known
software
Pro tip: avoid generative AI images in blog posts (even if your CMS
says you should have one for SEO purposes) unless you want to make a
political statement: AI:
The New Aesthetics of Fascism by Gareth Watkins
The Linux kernel project can’t use code from sanctioned countries.
Other projects need to check compliance with sanctions, too. US
Blocks Open Source ‘Help’ From These Countries by Steven J.
Vaughan-Nichols
p. 3 We do not consider legal arguments for consumer
privacy as a fundamental right or concerns about access to personal data
by malign actors or governments.
Avoiding malign actors is the big reason for restricting personalized
ads. And malign actors are numerous. The
high-profile national security threats are already in the news, but
most people will encounter miscellaneous malware, scams,
rip-offs and other lesser villainy enabled by ad personalization more
often than they have to deal with state or quasi-state adversaries.
There is no hard line between malign actors and totally legit
sellers—not only does the personalized ad business have plenty of
halfway crooks, you can find n/m-way crooks for arbitrary
values of n and m.
Ad personalization gives
a bunch of hard-to-overcome advantages to deceptive sellers.
Although scams are generally illegal and/or against advertising platform
policies, personalization makes the rules easier to evade, as we see
with some ways that
Facebook ads are optimized for deceptive advertising. Most
personalized ads aren’t clustered at the good (high-quality pair of
shoes in your size, on sale, next door!) or bad (malware pre-configured
for your system) ends of the spectrum. Advertisers at all levels of
quality and honesty are present, so any framework for thinking about ad
personalization needs to take that variability into account.
p. 3 Some privacy advocates assume, incorrectly, that
personalized marketing based on granular consumer data is automatically
harmful…
Treating personalized advertising as harmful by default is not an
assumption, but a useful heuristic based on both theoretical models and
real-world experience. personally, I don’t pay
attention to your ad if it’s personalized to me—it’s as credible as a
cold call. But I might pay attention to your ad if it’s run in a place
where the editors of sites that cover your industry would see it, or
your mom would. Yes, it is possible for professors to imagine a
hypothetical world in which personalization is beneficial, but that only
works if you make the unrealistic simplifying assumption that all
sellers are honest and that the only impact of personalization is to
show people ads that are more or less well matched to them. The
theoretical arguments in favor of personalized advertising break down as
soon as you level up your economic model to consider the presence of
both honest and deceptive advertisers in a market.
See Gardete and Bart, Tailored
Cheap Talk: The Effects of Privacy Policy On Ad Content and Market
Outcomes. Our research suggests that another peril of sharing
very high quality targeting information with advertisers is that ad
content may become less credible and persuasive to consumers. An
advertising medium that allows for personalization is incapable of
conveying as much information from an honest seller to a potential buyer
as an advertising medium that does not support personalization.
Mustri et al., in Behavioral
Advertising and Consumer Welfare, find that products found in
behaviorally targeted ads are likely to be associated with
lower quality vendors and higher product prices compared to competing
alternatives found among search results.
p. 8 Which Consumers Care Most About Privacy, and Do
Privacy Policies Unintentionally Favor the Privileged?
Lots of studies show that, basically, some people really want
cross-context personalized advertising, some people don’t, and for the
largest group in the middle, it depends how you ask. (references at the 30-40-30 rule). But the
difference in consumer preferences is not about privilege, it’s
about information level. See Turow et. al, Americans
Reject Tailored Advertising and Three Activities That Enable It.
That study includes a survey of privacy preferences before and after
informing the participants about data practices—and people were
more likely to say they do not want tailored advertising
after getting the additional information.
In the Censuswide study Who’s
In the Know: The Privacy Pulse Report, the experienced
advertisers surveyed in the USA (people with 5 or more years of ad
experience) were more likely than average to use an ad blocker (66% >
52%), and privacy is now the number one reason for people to use one. It
is reasonable for policy-makers to consider the preferences of
better-informed people—which is already a thing in fields such as
transportation safety and public health.
p. 11 Poorer consumers live in data deserts (Tucker
2023), causing algorithmic exclusion due to missing or fragmented data.
This exclusion thwarts marketing outreach and may deprive them of
offers, exacerbating data deserts and marginalization.
Instead of speculating about this problem, personalized advertising
proponents who are concerned about some people not being tracked enough
can already look at other good examples of possibly under-surveilled
consumers. Early adopters of privacy tools and preferences are helpfully
acting as the experimental group for a study that the surveillance
business hasn’t yet run. If people on whom less data is collected
are getting fewer win-win offers, then the privacy early adopters should
have worse consumer outcomes than people who leave the personalization
turned on. For example, Apple iOS users with App Tracking Transparency
(ATT) set to allow tracking should be reporting higher satisfaction and
doing fewer returns and chargebacks. So far, this does not seem to be
happening. (For a related result, see Bian et al., Consumer Surveillance and
Financial Fraud. Consumers who deliberately placed themselves in a
data desert by changing ATT to disallow tracking reported less
fraud.) Click
this to buy better stuff and be happier
And there’s little evidence to suggest that if a personalized ad
system knows someone to be poor, that they’ll receive more of the kind
of legit, well-matched offers that are targeted to the more affluent.
Poor people tend to receive more predatory finance and other deceptive
offers, so may be better off on average with ads less well matched to
their situation.
p. 13 More broadly, without cross-site/app identity,
consumers enjoy less free content
This depends on how you measure content and how you define
enjoy.The
Kircher and Foerderer paper states that, although children’s games
for Android got fewer updates on average after a targeted advertising
policy change by Google,
Only exceptionally well-rated and demanded games experienced more
feature updates, which could be interpreted as a sign of opportunity due
to better monetization potential or weakened competition. However,
considering that we observed these effects only for games in the highest
decile of app quality and demand and given that the median user rating
of a game is 4.1 of 5, our findings suggest widespread game
abandonment.
By Sturgeon’s
Law, a policy change that benefits the top 10% of games but not the
bottom 90% (which, in total, account for a small fraction of total
installs and an even smaller fraction of gameplay) is a win for the
users.
We find that more targeting increases competition and reduces the
websites’ profits, but yet in equilibrium websites choose maximum
targeting as they cannot credibly commit to low targeting. A privacy
protection policy can be beneficial for both consumers and websites.
When both personalized and non-personalized ad impressions are
available in the same market, the personalized impressions tend to go
for about
double the non-personalized. But it doesn’t work to artificially
turn off some data collection for a fraction of ad impressions, observe
that revenue for those impressions is lower (compared to impressions
with the data that are still available), and then extrapolate the
revenue difference to a market in which no impressions have the data
available.
It is also important to consider the impact of extremely low-quality
and/or illegal content in the personalized advertising market. Much of
the economic role of ad personalization is not to match the right ad to
the right user but to monetize a higher-value user on lower-value
content. The
surveillance economy is more like the commodification economy.
Surveillance advertising companies are willing to pursue content
commodification even to the point of taking big reputational risks from
feeding ad money to the worst people on the Internet (Hiding
in Plain Sight: The Ad-Supported Piracy Ring Driving Over a Billion
Monthly Visits - deepsee.io, Senators
Decry Adtech Failures as Ads Appear On CSAM Site). If advertising
intermediaries were more limited in their ability to put a good ad on a
bad site using user tracking, the higher-quality content sites would
enjoy significantly increased market power.
p. 14 Restrictions to limit the effectiveness of digital
advertising would likely disproportionately disadvantage small
businesses, since nine out of ten predominantly use digital advertising,
especially on Meta
Are small businesses really better off in the surveillance
advertising era? Although personalized Big Tech advertising is the main
ad medium available to small businesses today, there is clearly some
survivorship bias going on here. The Kerrigan
and Keating paper states that, While entrepreneurship has
rebounded since the Great Recession and its aftermath, startup activity
remains weak by historical standards. This period of time overlaps
with the golden age of personalized advertising, after widespread
adoption of smartphones but before Apple’s ATT, the EU’s GDPR, and
California’s CCPA. If personalized advertising is so good for small
businesses, where are the extra small businesses enabled by it? We
should have seen a small business boom in the second half of the 2010s,
after most
people in the USA got smartphones but before CCPA and ATT.
Jakob Nielsen may have provided the best explanation in 2006’s Search
Engines as Leeches on the Web, which likely applies not just to
search, but to other auction-based ad placements like social media
advertising. An auction-based advertising platform like those operated
by Google and Meta is able to dynamically adjust its advertising rates
to capture all of the expected incremental profits from the customers
acquired through it.
Part of the missing small business effect may also be caused by
platform concentration. If, instead of an advertising duopoly, small
businesses had more options for advertising, the power balance between
platform (rentier) and small business (entrepreneur) might shift more
toward the latter. See also Crawford et al., The
antitrust orthodoxy is blind to real data harms. Policy makers might
choose to prioritize pro-competition privacy legislation such as surveillance
licensing for the largest, riskiest platforms in order to address
competition concerns in parallel with privacy ones.
p. 15 Since PETs are costly for firms to implement,
forward-looking regulation should consider how to incentivize PET
adoption and innovation further.
In a section about how so-called privacy-enhancing
technologies (PETs) have equal perceived privacy violation and
bigger competition issues than conventional personalization, why
recommend incentivizing PETs? The works cited would better support a
recommendation to have a more detailed or informative consent experience
for PETs than for cookie-based tracking. Because PETs
obfuscate real-world privacy problems such as fraud and algorithmic
discrimination, it would be more appropriate to require additional
transparency, and possibly licensing, for PETs.
PETs, despite their mathematical appeal to many at Big Tech firms,
have a
long list of problems when applied to the real world. The
creeped-out attitude of users toward PETs is worth paying attention to,
as people who grow up in market economies generally develop good
instincts about information in markets—just like people who grow up
playing ball games can get good at catching a ball without consciously
doing calculus. Policymakers should pay more attention to user
perceptions—which are based on real-world market activity—than to
mathematical claims about developers’ PET projects. PETs should be
considered from the point of view of regulators investigating
discrimination and fraud complaints, which are often difficult
to spot on large platforms. Because PETs have the effect of
shredding the evidence of platform misdeeds, enabling the existing
problems of adtech, just in a harder-to-observe way, they need more
scrutiny, not incentivization.
Coming soon: a useful large-scale experiment
Policymakers may soon be able to learn from what could be the
greatest experiment on the impact of ad personalization ever
conducted.
If Meta is required to offer Facebook users in the European Union a
meaningfully de-personalized ad experience (and not just the less
personalized ads option that still allows for personalization
using fraud risk factors like age, gender, and location) then there will
be a chance to measure what happens when users can choose personalized
or de-personalized ads on a service that is otherwise the same.
I put these links and notes together to help myself out when someone
drops a link to the Dubé et al. paper into an Internet argument, and put
them up here in the hope that they will help others. Hardly anyone will
read all the literature in this field, but a lot of the most interesting
research is still found in corners of the library that Big Tech isn’t
actively calling attention to.
Thanks to Fengyang Lin for reviewing a draft of this post.
First, I know that pretty much everyone is (understandably) freaking
out about stuff that is getting worse, but I just wanted to share some
good news in the form of an old-fashioned open-source success story. I’m
a fairly boring person and developed most of my software habits in the
late 1990s and early 2000s, so it’s pretty rare that I actually hit a
bug.
A helpful developer, Jan Kratochvil, wrote a fix and put in a
pull request.
A bot made test packages and commented with instructions for me
on how to test the fix.
I commented that the new version works for me
The fix just went into Fedora. Pretty damn slick.
This is a great improvement over how this kind of thing used to work.
I hardly had to do anything. These kids today don’t know how good they
have it.
story number 2: why support the Linux desktop?
Amazon
Chime is shutting down. Did anyone use it? I get invited to a lot of
video conferences, and I never got invited to an Amazon Chime meeting.
Even though Amazon.com is normally really good at SaaS, this one didn’t
take off. What happened?
It looks like Amazon Chime was an interesting example of Nassim
Nicholas Taleb’s intransigent
minority effect.
The system
requirements for Amazon Chime look pretty reasonable, right? Should
get 95% of the client systems out there. The number of desktop Linux
users is pretty small. But if you have 20 meetings a week, at 95%
compatibility you’re going to average a compatibility issue every week.
Even worse, the people you most want to make a good first impression on
are the people whose client platform you’re least likely to know.
And if you do IT support for a company with 100 people organizing
meetings, Amazon Chime is going to cause way too many support issues to
put up with. Taleb uses the examples of kosher and halal food—only a
small fraction of the population will only eat kosher or halal, but when
planning food for a large group, the most practical choice is to satisfy
the minority.
The minority rule will show us how it all it takes is a small number
of intolerant virtuous people with skin in the game, in the form of
courage, for society to function properly.
Anyway, something to keep in mind in the future for anyone
considering moving the support desktop Linux or support
Firefox tickets to backlog. None of the successul video conferencing
platforms give me any grief for my Linux/Firefox/privacy nerdery
client-side setup.
Bonus links
Liam Proven and Thomas Claburn cover the latest web browser
surveillance drama: Mozilla
flamed by Firefox fans after reneging on promises to not sell their
dataMozilla doesn’t sell data about you (in the way that most
people think about selling data), and we don’t buy data about
you, he said. We changed our language because some jurisdictions
define sell more broadly than most people would usually
understand that word. (Don’t forget to turn
off advertising features in Firefox.)
David Roberts interviews Mustafa Amjad and Waqas Moosa about Pakistan’s solar
boom. What has prompted this explosion of distributed solar is
some combination of punishingly high prices for grid power and solar
panels getting very, very, very cheap. A glut of Chinese overcapacity
means that the price of panels in Pakistan has gone from 24 cents a watt
to 10 cents a watt in just the past year or two. Distributed solar is
breaking over Pakistan like a tidal wave, despite utilities and a grid
that do not seem entirely prepared for it.
AI
and Esoteric Fascism by Baldur Bjarnason. When I first began to
look into Large Language Models (LLMs) and Diffusion Model back in 2022,
it quickly became obvious that much of the rhetoric around LLMs was…
weird. Or, if we’re being plain-spoken, much of what the executives and
engineers at the organisations making these systems were saying was
outright weirdo cult shit…
It
is no longer safe to move our governments and societies to US clouds
by Bret Hubert. With all sorts of magic legal spells like
DPIAs and DTIAs, organizations attempt to justify
transferring our data and processes to the US. There is a whole industry
that has been aiding and abetting this process for years. People also
fool themselves that special keys and “servers in the EU” will get you
“a safe space” within the American cloud. It won’t.
This came out in 2020 but worth re-reading today: Puncturing
the Paradox: Group Cohesion and the Generational Myth by Harry
Guild. The highest group cohesion by profession is in Marketing.
This is advertising’s biggest problem in a single chart. This is the
monoculture. How can we possibly understand, represent and sell to an
entire country when we exist in such a bubble? We like to style
ourselves as free thinkers, mavericks and crazies, but the grim truth is
that we’re a more insular profession than farming and boast more
conformists than the military.
Tech
continues to be political by Miriam Eric Suzanne. Maybe we should
consider the beliefs and assumptions that have been built into a
technology before we embrace it? But we often prefer to treat each new
toy as as an abstract and unmotivated opportunity. If only the good
people like ourselves would get involved early, we can surely teach
everyone else to use it ethically!
Important reminder at Sauropod Vertebra Picture of the Week. If
you believe in “Artificial Intelligence”, take five minutes to ask it
about stuff you know wellBecause LLMs get catastrophically wrong
answers on topics I know well, I do not trust them at all on topics I
don’t already know. And if you do trust them, I urge you to spend five
minutes asking your favourite one about something you know in
detail. (This is a big part of the reason I don’t use LLMs for
search or research. A lot of the material isn’t just wrong, but reflects
old, wrong, but often repeated assumptions that you need to know a lot
about a field to know not to apply.)
Wendy Davis covers Meta
Sued Over Discriminatory College Ads. A civil rights group has
sued Meta Platforms over ad algorithms that allegedly discriminate by
disproportionately serving ads for for-profit colleges to Black users
and ads for public colleges to white users. Situations like this are
a big part of the reason why people should stop
putting privacy-enhancing advertising technologies in web
browsers—they mainly obfuscate discrimination and fraud.)
Cities
Can Cost Effectively Start Their Own Utilities Now by Kevin Burke.
Most PG&E ratepayers don’t understand how much higher the rates
they pay are than what it actually costs PG&E to generate and
transmit the electricity to their house. When I looked into this
recently I was shocked. The average PG&E electricity charge now
starts at 40 cents per kilowatt hour and goes up from there. Silicon
Valley Power, Santa Clara’s utility company, is getting power to
customers for 17 cents per kilowatt hour. Sacramento’s utility company
charges about the same.
Three
years on, Europe looks to Ukraine for the future of defense tech by
Mike Butcher. But in order to invest in the right technology, Europe
will have to look to Ukraine, because that is where future wars are
being fought right now. TechCrunch recently put a call out for Ukrainian
dual-use and defense tech startups to update us on what they are working
on. Below is what they sent us, in their own words. Related: Ukrainian
Drones Flew 500 Miles And, In A Single Strike, Damaged 5% Of Russia’s
Oil Refining Capacity by David Axe. (Drones are getting longer
ranges and better autonomy, fast. Fossil fuel infrastructure is not
getting any better protected or faster to repair. In the near future,
you’re only going to be in the oil or gas business if nobody who is good
at ML and model airplanes has a strong objection to you being in the oil
or gas business.)