“404 Media previously reported Cox Media Group (CMG) was advertising a service that claimed to target ads based on what potential customers said near device microphones. Now, here is the pitch deck CMG sent to prospective companies. Google has kicked CMG off its Partner Program in response…”
They are certainly very creepy but I doubt companies like Google or Meta even need this kind of data from a third party. If they truly wanted to have mic access, they could for a long time, and it would have been known. The reality is it is too expensive and risky to run this kind of spying, and I don’t think the benefit is worth the risk to them. To me this screams “SCAMMERS”.
If they truly wanted to have mic access, they could for a long time
agreed
and it would have been known
are you sure?
The reality is it is too expensive
imo this commonly repeated view has never been substantiated.
we’ve yet to see a technical explanation for why it’s “impossible/too expensive” which addresses the modern realities of efficient voice codecs, even rudimentary signal processing and modern speech-to-text network models.
and risky
how so? previously invasive features are simply written off as “a bug”. they barely even need to issue some b̶r̶i̶b̶e̶s̶ fines (typical corporate solution to getting caught), that is the level we’re currently at:
“whoops it was a bug, we’ll switch it off”
“whoops another update switched it on again” (if caught, months/years later)
“whoops some other opt-in surveillance switched itself on again, just another bug ¯_(ツ)_/¯”
in my experience when reading past the “they’re not listening” headlines, and into the actual technical reports, noone has been able to conclusively rule it out. if you know of conclusive documentation, please post.
then there’s the “they have enough data already” argument. which is entirely without foundation, as we all know very well: nothing is ever enough for these pathologically greedy entities. ‘enough’ simply isn’t in their vocabulary. we all know this already.
[i didn’t downvote you btw]
I simply think that until now (maybe they will start tomorrow), the PR and lawsuit risk of listening to people is too high, for the benefit they would get out of it. Much simpler metrics are enough for them to get a very good profile of the user. Voice data isn’t like in the test scenarios where the person will repeat 45x the word cat food, people talk about the weather and about gas prices which is pretty useless for creating an ad profile if you ask me. But the scary part is now with AI models and on device AI everything, local processing of the mic data into topics that then get sent to their servers is more concerning is not much more feasible.
And for the lawsuits I am not sure they could write it off as a bug everywhere other than the us and Canada because there are actually normal laws in most other countries
the PR and lawsuit risk
what risk? facebook & others conducted illegal human experiments. this is an enormous crime and was widely reported yet all fb had to do from a pr perspective was apologise.
as we all know, fb even interfered with with the electoral process of arguably the world’s most powerful nation, and all they had to do was some rebranding to meta and it’s business as usual. this is exactly how powerful these organisations are. go up against a global superpower & all you need to do is change your business name??? they don’t face justice the same way anyone else would, therefore we cannot assess the risk for them as we would another entity - and they know it.
So, while i personally disagree for above reasons, I can accept in your opinion they wouldn’t take the legal risk.
simpler metrics are enough
when has ‘enough’ ever satisfied these entities? we merely need to observe the rate of evolution of various surveillance methods, online, in our devices, in shopping centers to see ‘enough’ is never enough. its always increasing, and at an alarming rate.
local processing of the mic data into topics that then get sent to their servers is more concerning is not much more feasible
sorry i didn’t quite understand, are you saying its not feasible or it is feasible? from the way the sentence started i thought you were going to say it could be, but then you said ‘not much more feasible’?
Voice data isn’t
voice conversations are near-universally prized in surveillance & intelligence. There hasn’t been any convincing argument for any generalised exception to that.
I am not sure they could write it off as a bug
it’s already been written off as a bug. i didn’t follow that story indefinitely but i’m not aware of even a modest fine being paid in relation to the above story. if it can accidentally transcribe and send your conversations to your contact list without your knowledge or consent (literally already happened - with impunity(?)), they can 1000% “accidentally” send it to some ‘debug’ server somewhere.
Are they actually doing it? It ofc remains to be seen. Imo the fallout if it was revealed would roughly look like this
- A few people would say “no shit”
- Most people would parrot the “ive done nothing wrong so i don’t care” line.
- A few powerless people would be upset.
Google already has a fleet of “Hello Google” enabled devices that do listen all the time. Some phones surely also support always-on for this. My TV supports it. Users are already deliberately enabling this. There is no need for shady tactics.
The consensus a few years ago in /r/privacy was that it’s too expensive and risky for smartphones to transmit audio data to their HQ, bandwidth constraints, processing power and capabilities considered.
Now… with higher specs and advances and optimizations in AI for audio transcription, would it be feasible to do all that spying but locally on the device itself? The device would transmit ‘daily reports’ after processing.