Technology
Apple Just Gave 1 Billion iPhone Users A Reason To Stay
Apple’s latest shocking update is definitely a reason to buy the new iPhone 13 and upgrade to iOS 15 soon. But with the sudden end of Apple’s war with Google and Facebook, the future size of your iPhone is really in jeopardy.
Now you know that Apple has delayed adding iCloud Photos CSAM Filter and iMessage Photo Filter to iOS 15. The setback has become inevitable as the consequences of Apple’s recession have begun. The release of the iPhone 13 was very dirty and Apple could not shake the negative news. When the stage lights go out, and when the release of a shiny apple begins, we don’t expect it.
Apple is the simple victim of its own success here. We pay a premium for our iPhones because we rely on superlatives – privacy, security, control. Apple’s decision to enable customer page tracking has sparked genuine questions about whether users should be with the brand, indicating how important the issue has become.
It’s a disaster for Apple – while listening to user feedback on pressing pause may be PR, the truth is there are bigger problems that can’t be solved without completely changing their plans. Worse for Apple, the company is lagging behind its main competitors this year after its own inconvenient returns.
Three privacy shocks in a few months – at least we can take a rest. The privacy lobby respects its capabilities and aligns its voice. When this is not the case, organized campaigns have made a physical change.
We first saw Facebook’s insistence that two billion WhatsApp users accept the new data sharing service terms or lose access to the app. That decision, which was as badly touted as Apple’s CSAM update, sparked control threats and complaints, a viral user boom that propelled Telegram and Signal into the limelight, and a blank slate in WhatsApp PR recovery. .
After secretly adding millions of Chrome users to a poorly designed FLoC test, we backed up to Google. Again, the privacy lobby was furious. Google said everything was fine before the investigation quietly ended and even more quietly acknowledged that the privacy concerns were justified. Google is back. The FLoC V1 has been killed and the company says it will take time to think of next steps.
Now apple. Google told the FLoC that “it’s clear that more time is needed in the ecosystem to do this right.”
So, Apple now puts the company on Facebook and Google — along with the world’s leading data harvesters — in an instant privacy backdrop. It’s really not good. And like its competitors, Apple has to ask some serious questions about how it went wrong.
But the one billion iPhone users now need to watch carefully, as other feedback makes a big difference — what Apple does next will shape the future of the iPhone.
Facebook wants to monetize WhatsApp – it runs the world’s biggest ambassador to generate revenue and revenue, no matter what the humorous, privacy-based social media posts suggest. WhatsApp’s end-to-end encryption may be a philosophy when operated privately, but it’s a marketing USB owned by Facebook.
Similarly, Google requires some kind of user tracking in Chrome. A company cannot sell targeted ads if it is unable to promote a data-based ecosystem along with mechanisms to buy and sell ads and measure their success. That’s why Google suspended tracking cookies, and why you should quit Chrome.
Device Filter is not required to run on Apple iPhones. Yes, it may have to scale up its CSAM screening beforehand, but it can do it with cloud screening, which is the industry standard with some added Apple ingenuity. It promotes almost no resistance compared to its program – others do the same, and its servers must use technology to keep such mess out of law enforcement or otherwise.
Knowing that Apple has already screened some iCloud Mail to CSAM, there will be no real response, no feedback, users will be fine. Steve Jobs said in 2010, “Privacy means that people know in clear English what they are signing for. What does it mean.” Ask them. Ask them every time. Stop listening to them, tell them they are tired. You ask them… “Apple, remember.
No need to run a machine learning page on the Apple iMessage page, which provides alerts to minors who send or receive sexual images. This is an unimaginable idea at every level and should be left to collect dust on the shelf.
Therefore, when Apple “releases critical child protection features”, they must include content, complete and early innovations, or encourage this regression again. The company never sells users with control over the idea of privacy and software surveillance that lurks behind their home screens.
Apple’s plans for CSAM scanning and iMessage conversion are technically complex and full of twists. Many complaints are false. Yes, comparing hashes of watch list images with user images can lead to false positives, and yes, there is a remote risk that computers could be misused to compromise users’ photo albums. But of the many concerns, the four issues are the most real and will continue to be a serious concern as long as Apple makes sure it wants to split or reorganize its plans.
First, the threat that governments could pressure Apple to comply with local laws or accept a settlement from these organizations in order to release local lawmakers. Apple Can’t Leave It Because of Its Record in China Hosting iCloud and App Data