Warning: mysqli::__construct(): (HY000/1203): User howardkn already has more than 'max_user_connections' active connections in D:\Inetpub\vhosts\howardknight.net\al.howardknight.net\includes\artfuncs.php on line 21
Failed to connect to MySQL: (1203) User howardkn already has more than 'max_user_connections' active connections
Warning: mysqli::query(): Couldn't fetch mysqli in D:\Inetpub\vhosts\howardknight.net\al.howardknight.net\index.php on line 66
Article <v89b9o$ms11$3@dont-email.me>
Deutsch   English   Français   Italiano  
<v89b9o$ms11$3@dont-email.me>

View for Bookmarking (what is this?)
Look up another Usenet article

Path: ...!eternal-september.org!feeder3.eternal-september.org!news.eternal-september.org!.POSTED!not-for-mail
From: Alan <nuh-uh@nope.com>
Newsgroups: misc.phone.mobile.iphone,alt.privacy
Subject: Re: Apple accused of underreporting suspected CSAM on its platforms
Date: Mon, 29 Jul 2024 17:14:48 -0700
Organization: A noiseless patient Spider
Lines: 242
Message-ID: <v89b9o$ms11$3@dont-email.me>
References: <v7mup4$7vpf$1@solani.org> <lg8ea1Fa94U1@mid.individual.net>
 <xn0oonlp4azqw16000@reader443.eternal-september.org>
 <lga2k1F7uk8U1@mid.individual.net>
 <xn0oonrftb7hazk002@reader443.eternal-september.org>
 <v7olut$19iie$1@dont-email.me> <lga8vfF8qq0U3@mid.individual.net>
 <v7q9vj$1l9co$1@dont-email.me>
 <v7qn3b$2hg0$1@nnrp.usenet.blueworldhosting.com>
 <lgcm26Fbd6U4@mid.individual.net>
 <v7rfko$18sp$1@nnrp.usenet.blueworldhosting.com>
 <lgda8oF3c6aU3@mid.individual.net>
 <v7tiln$2g2b$1@nnrp.usenet.blueworldhosting.com>
 <v7u2gu$2co3l$1@dont-email.me> <lgfmvgFee44U1@mid.individual.net>
 <v7uduv$11n1$1@nnrp.usenet.blueworldhosting.com>
 <v80bgk$2s7ns$1@dont-email.me>
 <v83f93$31ac$1@nnrp.usenet.blueworldhosting.com>
 <v860er$1kea$1@dont-email.me> <v86141$j$1@nnrp.usenet.blueworldhosting.com>
 <v87e4d$ckpm$1@dont-email.me>
 <v87u43$1kl6$1@nnrp.usenet.blueworldhosting.com>
 <v88npo$j33d$8@dont-email.me> <v8943a$lrfd$1@dont-email.me>
 <v89858$jup0$14@dont-email.me> <v89b1n$mtmj$1@dont-email.me>
MIME-Version: 1.0
Content-Type: text/plain; charset=UTF-8; format=flowed
Content-Transfer-Encoding: 8bit
Injection-Date: Tue, 30 Jul 2024 02:14:49 +0200 (CEST)
Injection-Info: dont-email.me; posting-host="54396cd3675252b2a65935bdbb96af3f";
	logging-data="749601"; mail-complaints-to="abuse@eternal-september.org";	posting-account="U2FsdGVkX187NqTnuZbifZ1MHiAXfQUtC7/E8U1iYKY="
User-Agent: Mozilla Thunderbird
Cancel-Lock: sha1:cfUM/6HxxrKax01EFBTwpViLdSY=
Content-Language: en-CA
In-Reply-To: <v89b1n$mtmj$1@dont-email.me>
Bytes: 12246

On 2024-07-29 17:10, Chips Loral wrote:
> Alan wrote:
>> On 2024-07-29 15:11, Chips Loral wrote:
>>> Alan wrote:
>>>> On 2024-07-29 04:23, Andrew wrote:
>>>>> Chris wrote on Mon, 29 Jul 2024 06:50:53 -0000 (UTC) :
>>>>>
>>>>>>> You not comprehending the difference between zero percent of 
>>>>>>> Apple reports
>>>>>>> versus zero total convictions is how I know you zealots own 
>>>>>>> subnormal IQs.
>>>>>>
>>>>>> Not at all. My position hasn't changed. You, however, have had 
>>>>>> about three
>>>>>> different positions on this thread and keep getting confused which 
>>>>>> one
>>>>>> you're arguing for. lol.
>>>>>
>>>>> Au contraire
>>>>>
>>>>> Because I only think logically, my rather sensible position has never
>>>>> changed, Chris, and the fact you "think" it has changed is simply 
>>>>> that you
>>>>> don't know the difference between the percentage of convictions 
>>>>> based on
>>>>> the number of reports, and the total number of convictions.
>>>>>
>>>>> When you figure out that those two things are different, then (and 
>>>>> only
>>>>> then) will you realize I've maintained the same position throughout.
>>>>>
>>>>> Specifically....
>>>>>
>>>>> a. If the Apple reporting rate is low, and yet if their conviction
>>>>>     rate is high (based on the number of reports), then they are NOT
>>>>>     underreporting images.
>>>>
>>>> Apple's reporting rate is ZERO, because they're not doing scanning 
>>>> of images of any kind.
>>>
>>> After getting caught.
>>>
>>> You can't seem to get ANYTHING right, Mac-troll:
>>>
>>> https://www.wired.com/story/apple-photo-scanning-csam-communication-safety-messages/
>>>
>>> In August 2021, Apple announced a plan to scan photos that users 
>>> stored in iCloud for child sexual abuse material (CSAM). The tool was 
>>> meant to be privacy-preserving and allow the company to flag 
>>> potentially problematic and abusive content without revealing 
>>> anything else. But the initiative was controversial, and it soon drew 
>>> widespread criticism from privacy and security researchers and 
>>> digital rights groups who were concerned that the surveillance 
>>> capability itself could be abused to undermine the privacy and 
>>> security of iCloud users around the world. At the beginning of 
>>> September 2021, Apple said it would pause the rollout of the feature 
>>> to “collect input and make improvements before releasing these 
>>> critically important child safety features.” In other words, a launch 
>>> was still coming.
>>>
>>> Parents and caregivers can opt into the protections through family 
>>> iCloud accounts. The features work in Siri, Apple’s Spotlight search, 
>>> and Safari Search to warn if someone is looking at or searching for 
>>> child sexual abuse materials and provide resources on the spot to 
>>> report the content and seek help.
>>>
>>> https://sneak.berlin/20230115/macos-scans-your-local-files-now/
>>>
>>> Preface: I don’t use iCloud. I don’t use an Apple ID. I don’t use the 
>>> Mac App Store. I don’t store photos in the macOS “Photos” 
>>> application, even locally. I never opted in to Apple network services 
>>> of any kind - I use macOS software on Apple hardware.
>>>
>>> Today, I was browsing some local images in a subfolder of my 
>>> Documents folder, some HEIC files taken with an iPhone and copied to 
>>> the Mac using the Image Capture program (used for dumping photos from 
>>> an iOS device attached with an USB cable).
>>>
>>> I use a program called Little Snitch which alerts me to network 
>>> traffic attempted by the programs I use. I have all network access 
>>> denied for a lot of Apple OS-level apps because I’m not interested in 
>>> transmitting any of my data whatsoever to Apple over the network - 
>>> mostly because Apple turns over customer data on over 30,000 
>>> customers per year to US federal police without any search warrant 
>>> per Apple’s own self-published transparency report. I’m good without 
>>> any of that nonsense, thank you.
>>>
>>> Imagine my surprise when browsing these images in the Finder, Little 
>>> Snitch told me that macOS is now connecting to Apple APIs via a 
>>> program named mediaanalysisd (Media Analysis Daemon - a background 
>>> process for analyzing media files).
>>>
>>> ...
>>>
>>>
>>> Integrate this data and remember it: macOS now contains network-based 
>>> spyware even with all Apple services disabled. It cannot be disabled 
>>> via controls within the OS: you must used third party network 
>>> filtering software (or external devices) to prevent it.
>>>
>>> This was observed on the current version of macOS, macOS Ventura 13.1.
>>>
>>
>> 'A recent thread on Twitter raised concerns that the macOS process 
>> mediaanalysisd, which scans local photos, was secretly sending the 
>> results to an Apple server. This claim was made by a cybersecurity 
>> researcher named Jeffrey Paul. However, after conducting a thorough 
>> analysis of the process, it has been determined that this is not the 
>> case.'
>>
> 
> 
> Bullshit.
> 
> https://www.majorgeeks.com/content/page/stop_apple_scanning_iphone_photos.html
> 
> Apple’s new iPhone photo-scanning feature is a very controversial thing. 
> You might want to consider the only current option to stop Apple from 
> scanning your photos.
> 
> Apple's new photo-scanning feature will scan photos stored in iCloud to 
> see whether they match known Child Sexual Abuse Material (CSAM). The 
> problem with this, like many others, is that we often have hundreds of 
> photos of our children and grandchildren, and who knows how good or bad 
> the new software scanning technology is? Apple claims false positives 
> are one trillion to one, and there is an appeals process in place. That 
> said, one mistake from this AI, just one, could have an innocent person 
> sent to jail and their lives destroyed.
> 
> Apple has many other features as part of these upgrades to protect 
> children, and we like them all, but photo-scanning sounds like a problem 
> waiting to happen.
> 
> Here are all of the "features" that come with anti-CSAM, expected to 
> roll out with iOS 15 in the fall of 2021.
> 
> Messages: The Messages app will use on-device machine learning to warn 
> children and parents about sensitive content.
> 
> iCloud Photos: Before an image is stored in iCloud Photos, an on-device 
> matching process is performed for that image against the known CSAM hashes.
> 
> Siri and Search: Siri and Search will provide additional resources to 
> help children and parents stay safe online and get help with unsafe 
> situations.
> 
> Now that you understand how anti-CSAM works, the only way to avoid 
> having your photos scanned by this system is to disable iCloud Photos. 
> Your photos are scanned when you automatically upload your photos to the 
> cloud, so the only current way to avoid having them scanned is not to 
> upload them.
> 
> This adds an interesting problem. The majority of iPhone users use 
> iCloud to back up their photos (and everything else). If you disable 
> iCloud, you will need to back up your photos manually. If you have a PC 
> or Mac, you can always copy them to your computer and back them up. You 
> can also consider using another cloud service for backups.
> 
> Let's talk about disabling iCloud and also removing any photos you 
> already have uploaded. You will have 30 days to recover your photos if 
========== REMAINDER OF ARTICLE TRUNCATED ==========