Here's a title you can reuse freely for the next decade or so.
(Startup/public/private equity owned) company <IOT device>'s collect data you don't want them collecting, use it for profit to your detriment, and didn't bother securing any of it because they don't care.
Each time it happens, it needs to be news to name and shame the companies. Unfortunately, once you've bought the product, it's game over for privacy. So this info needs to be explicitly available for each product/company so that when future buyers are researching, they might be able to stumble upon these articles.
Product reviewers need to explicitly state that the cameras/mics/whatevs of devices have been used for nefarious purposes other than what is advertised on the box.
But we should not just sweep everything under the rug because a couple of nerds "knows about it" because there's a heck of a lot more people that do not.
Name and shame doesn't work in that it doesn't stop the next guy, or even the current guy. It does at least make the information available to those that care. If you don't care, great. Continue to live with your head in the sand. If you do care, at least the information is available for you to make an informed decision.
If we do nothing because of the "it doesn't work" in a manner you think fitting, then we'll make no progress. It's yet another example of a choice between doing anything versus doing nothing because the perfect answer isn't available.
I do not use Microsoft products. I do not use Google products in my personal life. Others do not use Apple products. So for some people, it absolutely does work. I don't shop at WalMart, and am damn near Amazon free too.
A new sucker is born every minute. If the only time the name and shame is mentioned is when it happens, then those new borns will potentially never hear about it.
Why do they preach to the choir? Because that's how you get them to sing.
Is it really completely legal? That would surprise me. And, of course, is it is, it shouldn't be.
I mean, in the end, it's just how you frame it, so it surely must be a viable class action lawsuit. When a teenager playfully hacks into someone's completely unprotected IoT anyone could walk into, he is breaking the law for some reason. When your business is not doing anything "wrong", but it provides a technical service targeted to businesses who actually produce malware and do harmful stuff like that, you can end up in prison for a lifetime. So surely there must be a way to frame this kind of thing as criminal activity too.
Tech progress at its finest. I stick to my 90s-made fridge, samey vintage washing machine, non-smart vacuum and non-smart microwave. All solved and sturdy appliances. Cheers.
Oh the interesting part is “our AI could not interpret images of common objects at unusual angles”.
Now that’s fascinating - why not? Is computer vision just boring pattern recognition and really does not have “concepts” underlying it - if so 90% of the AI hype is false?
There are cases where AI can recognise gender on an X-ray when humans can't, find tumors that experienced doctor's can't. This must mean that human doctors looking at Xrays use just boring pattern recognition and AI has actual concepts of what it's seeing.
But does it really? Or is it more observant than a human doctor and more thorough, but only at the limited task of deciding if this X-ray looks like the million other X-rays of a male abdomen versus the million X-rays of a female abdomen.
I assume counting the number of ribs is not what is meant …
It's simple math. If the correlation between gender and sex is 0.99 then if a method can determine your sex with say a 90% accuracy then it can determine your gender with an 89% accuracy (very roughly). The difference is negligible.
What ? Stable diffusion doesn't have an underlying understanding that humans typically have two arms, two hands and five fingers per hand gathered from vast sea of training data ? That's a bold statement.
IIRR it’s a debate as to the difference between 99% of the time
It predicts the next pixel will be fleshy and the pixel next to it is background this making something that looks fingery (and so when presented with
An odd angle that 99% drops crazily” or that somehow there is a executive function that has evolved that gets a concept of finger with movement, musculature etc
It’s the “somehow evolved” part that is where I have my concerns.
Predictive ability based on billions images, sounds good. Executive function - how does that work? But at some point we are playing “what is consciousness” games.
Would love to hear more rigourous thought than mine - any links gratefully received:-)
I actually agree with you. I was a bit sarcastic. If I understand correctly there isn't a fundamental difference when it comes to text output vs pixel data output in this context. If so then it suddenly sounds much more of a stretch (intuitively) to claim that somehow stable diffusion understands the real world (like people claim to be the case with language models).
Just to clarify: The photos and audio collection isn't related to the mentioned security flaws. These are two separate issues.
> Ecovacs robot vacuums, which have been found to suffer from critical cybersecurity flaws...
> An Ecovacs spokesperson confirmed the company uses the data collected as part of its product improvement program to train its AI models.
At least I know I'm right to avoid anything with a camera on it. You're not crazy if they're after you. I also try to avoid chinese products, but we all know that's not completely possible anymore.
Is this news anymore?
Here's a title you can reuse freely for the next decade or so.
(Startup/public/private equity owned) company <IOT device>'s collect data you don't want them collecting, use it for profit to your detriment, and didn't bother securing any of it because they don't care.
Each time it happens, it needs to be news to name and shame the companies. Unfortunately, once you've bought the product, it's game over for privacy. So this info needs to be explicitly available for each product/company so that when future buyers are researching, they might be able to stumble upon these articles.
Product reviewers need to explicitly state that the cameras/mics/whatevs of devices have been used for nefarious purposes other than what is advertised on the box.
But we should not just sweep everything under the rug because a couple of nerds "knows about it" because there's a heck of a lot more people that do not.
Name and shame doesn't work. What does work is prison terms for ceos.
*Prison terms for the board of directors.
In the event of incidents that result in mass deaths (oil spills, etc), tried as if mass murderers for the board.
CEOs are too often just scapegoats for the evils of a board.
I mean I've seen it go both ways, but sure, the board should know what the ceo is doing and vice versa.
Name and shame doesn't work in that it doesn't stop the next guy, or even the current guy. It does at least make the information available to those that care. If you don't care, great. Continue to live with your head in the sand. If you do care, at least the information is available for you to make an informed decision.
If we do nothing because of the "it doesn't work" in a manner you think fitting, then we'll make no progress. It's yet another example of a choice between doing anything versus doing nothing because the perfect answer isn't available.
> What does work is prison terms for ceos.
Could work. But, unfortunately those CEOs make the law.
> Each time it happens, it needs to be news to name and shame the companies.
Was this ever a problem for Microsoft, Google or Apple ? /s
I do not use Microsoft products. I do not use Google products in my personal life. Others do not use Apple products. So for some people, it absolutely does work. I don't shop at WalMart, and am damn near Amazon free too.
A new sucker is born every minute. If the only time the name and shame is mentioned is when it happens, then those new borns will potentially never hear about it.
Why do they preach to the choir? Because that's how you get them to sing.
And what repercussions does a company like that have to fear? None.
Legislation worldwide needs to catch up with tech badly.
Is it really completely legal? That would surprise me. And, of course, is it is, it shouldn't be.
I mean, in the end, it's just how you frame it, so it surely must be a viable class action lawsuit. When a teenager playfully hacks into someone's completely unprotected IoT anyone could walk into, he is breaking the law for some reason. When your business is not doing anything "wrong", but it provides a technical service targeted to businesses who actually produce malware and do harmful stuff like that, you can end up in prison for a lifetime. So surely there must be a way to frame this kind of thing as criminal activity too.
This has been an on-going issue for decades, I believe. Will this ever happen? Do people in control want it to happen?
Tech progress at its finest. I stick to my 90s-made fridge, samey vintage washing machine, non-smart vacuum and non-smart microwave. All solved and sturdy appliances. Cheers.
But, it's not power efficient. You want a great uppercase letter closer to the beginning of ABC :D.
Oh the interesting part is “our AI could not interpret images of common objects at unusual angles”.
Now that’s fascinating - why not? Is computer vision just boring pattern recognition and really does not have “concepts” underlying it - if so 90% of the AI hype is false?
There must be several phds in that at least :-)
There are cases where AI can recognise gender on an X-ray when humans can't, find tumors that experienced doctor's can't. This must mean that human doctors looking at Xrays use just boring pattern recognition and AI has actual concepts of what it's seeing.
But does it really? Or is it more observant than a human doctor and more thorough, but only at the limited task of deciding if this X-ray looks like the million other X-rays of a male abdomen versus the million X-rays of a female abdomen.
I assume counting the number of ribs is not what is meant …
That is actually pretty cool, but I believe you meant to say "biological sex" instead of "gender". :P
I have no clue how an AI may find the gender (which is in the mind) of someone through x-rays alone.
the mathematical correlation between the two is so high as to be negligible
What does this mean? I do not think anyone could determine my gender based on x-rays alone. My biological sex, however, definitely.
It's simple math. If the correlation between gender and sex is 0.99 then if a method can determine your sex with say a 90% accuracy then it can determine your gender with an 89% accuracy (very roughly). The difference is negligible.
What ? Stable diffusion doesn't have an underlying understanding that humans typically have two arms, two hands and five fingers per hand gathered from vast sea of training data ? That's a bold statement.
I think the issue is “understanding”
IIRR it’s a debate as to the difference between 99% of the time It predicts the next pixel will be fleshy and the pixel next to it is background this making something that looks fingery (and so when presented with An odd angle that 99% drops crazily” or that somehow there is a executive function that has evolved that gets a concept of finger with movement, musculature etc
It’s the “somehow evolved” part that is where I have my concerns.
Predictive ability based on billions images, sounds good. Executive function - how does that work? But at some point we are playing “what is consciousness” games.
Would love to hear more rigourous thought than mine - any links gratefully received:-)
I actually agree with you. I was a bit sarcastic. If I understand correctly there isn't a fundamental difference when it comes to text output vs pixel data output in this context. If so then it suddenly sounds much more of a stretch (intuitively) to claim that somehow stable diffusion understands the real world (like people claim to be the case with language models).
> and five fingers per hand
In my experience it's more like three to six. But your argument's still valid. There is a concept
Just to clarify: The photos and audio collection isn't related to the mentioned security flaws. These are two separate issues.
> Ecovacs robot vacuums, which have been found to suffer from critical cybersecurity flaws... > An Ecovacs spokesperson confirmed the company uses the data collected as part of its product improvement program to train its AI models.
I have one, it comes with a sticker over the camera. I just left it on.
https://dontvacuum.me/
Privacy respecting vacuum robots? Why not Valetudo?
https://valetudo.cloud/
That’s what Valetudo uses.
aside: W1XM is a cool callsign. I'm sure it elicits the satellite radio company, but to me it's https://en.wikipedia.org/wiki/XM_(file_format)
also it's whisky one xray mike, which rolls off the tongue
At least I know I'm right to avoid anything with a camera on it. You're not crazy if they're after you. I also try to avoid chinese products, but we all know that's not completely possible anymore.