The blog post "Let's talk about AI and end-to-end encryption" by Matthew Green on cryptographyengineering.com delves into the complex relationship between artificial intelligence and end-to-end encryption (E2EE), exploring the perceived conflict between allowing AI access to user data for training and maintaining the privacy guarantees provided by E2EE. The author begins by acknowledging the increasing calls to allow AI models access to encrypted data, driven by the desire to leverage this data for training more powerful and capable AI systems. This desire stems from the inherent limitations of training AI on solely public data, which often results in less accurate and less useful models compared to those trained on a broader dataset, including private user data.
Green meticulously dissects several proposed solutions to this dilemma, outlining their technical intricacies and inherent limitations. He starts by examining the concept of training AI models directly on encrypted data, a technically challenging feat that, while theoretically possible in limited contexts, remains largely impractical and computationally expensive for the scale required by modern AI development. He elaborates on the nuances of homomorphic encryption and secure multi-party computation, explaining why these techniques, while promising, are not currently viable solutions for practical, large-scale AI training on encrypted datasets.
The post then transitions into discussing proposals involving client-side scanning, often framed as a means to detect illegal content, such as child sexual abuse material (CSAM). Green details how these proposals, while potentially well-intentioned, fundamentally undermine the core principles of end-to-end encryption, effectively creating backdoors that could be exploited by malicious actors or governments. He meticulously outlines the technical mechanisms by which client-side scanning operates, highlighting the potential for false positives, abuse, and the erosion of trust in secure communication systems. He emphasizes that introducing any form of client-side scanning necessitates a shift away from true end-to-end encryption, transforming it into something closer to client-to-server encryption with client-side pre-decryption scanning, thereby compromising the very essence of E2EE's privacy guarantees.
Furthermore, Green underscores the slippery slope argument, cautioning against the potential for expanding the scope of such scanning beyond CSAM to encompass other types of content deemed undesirable by governing bodies. This expansion, he argues, could lead to censorship and surveillance, significantly impacting freedom of expression and privacy. The author concludes by reiterating the importance of preserving end-to-end encryption as a crucial tool for protecting privacy and security in the digital age. He emphasizes that the perceived tension between AI advancement and E2EE necessitates careful consideration and a nuanced approach that prioritizes user privacy and security without stifling innovation. He suggests that focusing on alternative approaches, such as federated learning and differential privacy, may offer more promising avenues for developing robust AI models without compromising the integrity of end-to-end encrypted communication.
The WatchTowr Labs blog post, entitled "Backdooring Your Backdoors – Another $20 Domain, More Governments," details a disconcerting discovery of further exploitation of vulnerable internet infrastructure by nation-state actors. The researchers meticulously describe a newly uncovered campaign employing a compromised domain, acquired for a nominal fee of $20 USD, to facilitate malicious activities against high-value targets within governmental and diplomatic circles. This domain, deceptively registered to mimic legitimate entities, acts as a command-and-control (C2) server, orchestrating the deployment and operation of sophisticated malware.
This revelation builds upon WatchTowr's previous investigation into similar malicious infrastructure, suggesting a broader, ongoing operation. The blog post elaborates on the technical intricacies of the attack, highlighting the strategic use of seemingly innocuous internet resources to mask malicious intent. The researchers delve into the domain registration details, tracing the obfuscated registration path to uncover links suggestive of government-backed operations.
Furthermore, the post emphasizes the expanding scope of these activities, implicating a growing number of nation-state actors engaging in this type of cyber espionage. It paints a picture of a complex digital battlefield where governments leverage readily available, low-cost tools to infiltrate secure networks and exfiltrate sensitive information. The seemingly insignificant cost of the domain registration underscores the ease with which malicious actors can establish a foothold within critical infrastructure.
The researchers at WatchTowr Labs meticulously dissect the technical characteristics of the malware employed, illustrating its advanced capabilities designed to evade traditional security measures. They detail the methods used to establish persistent access, conceal communications, and exfiltrate data from compromised systems. This comprehensive analysis sheds light on the sophistication of these attacks and the considerable resources dedicated to their execution.
Ultimately, the blog post serves as a stark reminder of the escalating threat posed by state-sponsored cyber espionage. It highlights the vulnerability of even seemingly secure systems to these sophisticated attacks and underscores the need for constant vigilance and robust security measures to mitigate the risks posed by these increasingly prevalent and sophisticated cyber campaigns. The researchers' detailed analysis contributes significantly to the understanding of these evolving threats, providing valuable insights for security professionals and policymakers alike.
The Hacker News post "Backdooring Your Backdoors – Another $20 Domain, More Governments" (linking to an article about governments exploiting vulnerabilities in commercially available surveillance tech) generated a moderate discussion with several compelling points raised.
Several commenters focused on the inherent irony and dangers of governments utilizing exploits in already ethically questionable surveillance tools. One commenter highlighted the "turf war" aspect, noting that intelligence agencies likely want these vulnerabilities to exist to exploit them, creating a conflict with law enforcement who might prefer secure tools for their investigations. This creates a complex situation where fixing vulnerabilities could be detrimental to national security interests (as perceived by intelligence agencies).
Another commenter pointed out the concerning implications for trust and verification in digital spaces. If governments are actively exploiting these backdoors, it raises questions about the integrity of digital evidence gathered through such means. How can we be certain evidence hasn't been tampered with, especially in politically sensitive cases? This commenter also touched upon the potential for "false flag" operations, where one nation could plant evidence via these backdoors to implicate another.
The discussion also delved into the economics and practicalities of this type of exploit. One commenter questioned why governments would bother purchasing commercial spyware with existing backdoors when they likely have the capability to develop their own. The responses to this suggested that commercial solutions might offer a quicker, cheaper, and less legally complicated route, particularly for smaller nations or for specific, targeted operations. The "plausible deniability" aspect of using commercial software was also mentioned.
Some skepticism was expressed about the WatchTowr Labs article itself, with one commenter noting a lack of technical depth and questioning the overall newsworthiness. However, others argued that the implications of the article, even without deep technical analysis, were significant enough to warrant discussion.
Finally, a few comments touched on the broader ethical implications of the surveillance industry and the chilling effect such practices have on free speech and privacy. One commenter expressed concern about the normalization of these types of surveillance tools and the erosion of privacy rights.
The blog post "Homomorphic Encryption in iOS 18" by Bastian Bohm details the introduction of homomorphic encryption capabilities within Apple's iOS 18 operating system, specifically focusing on the newly available APIs for performing calculations on encrypted data without requiring decryption. The author expresses excitement about this development, highlighting the potential for enhanced privacy and security in various applications.
The post begins by explaining the concept of homomorphic encryption, emphasizing its ability to process encrypted information directly, thus preserving the confidentiality of sensitive data. It distinguishes between Fully Homomorphic Encryption (FHE), which supports arbitrary computations, and Partially Homomorphic Encryption (PHE), which is limited to specific operations like addition or multiplication. The post clarifies that iOS 18 implements PHE, specifically focusing on additive homomorphic encryption.
The core of the post revolves around the newly introduced SecKeyEncryptedData
class and its associated methods. The author provides a concise code example demonstrating how to create encrypted integers using this class and how to perform homomorphic addition on these encrypted values. The resulting sum remains encrypted, and only the holder of the decryption key can reveal its true value. The author meticulously breaks down the code snippet, explaining the role of each function and parameter. For instance, the post elucidates the process of generating a public key specifically designated for encrypted data operations and how this key is subsequently used to encrypt integer values. It also explains the significance of the perform
method in executing homomorphic operations on these encrypted integers.
Furthermore, the post discusses the underlying cryptographic scheme employed by Apple, revealing that it leverages a variant of the Paillier cryptosystem. This choice is deemed suitable for integer additions and is acknowledged for its established security properties. The post also touches upon the practical limitations of PHE, specifically noting the inability to perform other operations like multiplication or comparison directly on the encrypted data without decryption.
Finally, the author speculates on the potential applications of this technology within the Apple ecosystem. The example given is privacy-preserving data collection, suggesting how homomorphic encryption could enable the aggregation of user statistics without compromising individual data privacy. This could be useful for applications like collecting usage metrics or accumulating health data while ensuring that the individual contributions remain confidential. The author concludes with an optimistic outlook on the future implications of homomorphic encryption within the iOS environment and expresses anticipation for further advancements in this field.
The Hacker News post titled "Homomorphic encryption in iOS 18" spawned a modest discussion with a handful of comments focusing on the practicalities and limitations of the technology, rather than the announcement itself. No one expressed outright excitement or skepticism about the announcement, instead offering pragmatic observations.
One commenter pointed out that the homomorphic encryption being utilized is limited to integer addition and multiplication, and thus isn't fully homomorphic encryption (FHE) in the broader, more powerful sense. They clarified that true FHE allows arbitrary computation on encrypted data, which is not what Apple is implementing. This comment served as an important clarification to distinguish the specific type of homomorphic encryption being employed.
Another user expanded on this by mentioning that the specific technique used is called "additive homomorphic encryption" and likely leverages the Paillier cryptosystem. This added technical depth to the discussion, providing a potential underlying mechanism for Apple's implementation. They then speculated about its use case, suggesting it could be applied to scenarios like federated learning or aggregated metrics collection.
A subsequent comment explored the performance limitations of homomorphic encryption. The commenter noted the significant computational overhead associated with these techniques, which makes them unsuitable for many real-time or performance-sensitive applications. This comment highlighted the trade-offs involved in using homomorphic encryption, emphasizing that while it offers enhanced privacy, it comes at the cost of performance.
Finally, one commenter linked to a related project called "Concrete," further adding context to the types of operations and optimizations possible within the homomorphic encryption space. This provides an avenue for those interested in learning more about practical implementations and advancements in the field.
Overall, the comments section offers a concise and informed discussion focusing on the technical nuances of Apple's implementation rather than broad speculation or hype. They provide valuable context and clarification regarding the specific type of homomorphic encryption being used and its inherent limitations.
The Home Assistant blog post entitled "The era of open voice assistants" heralds a significant paradigm shift in the realm of voice-controlled smart home technology. It proclaims the dawn of a new age where users are no longer beholden to the closed ecosystems and proprietary technologies of commercially available voice assistants like Alexa or Google Assistant. This burgeoning era is characterized by the empowerment of users to retain complete control over their data and personalize their voice interaction experiences to an unprecedented degree. The post meticulously details the introduction of Home Assistant's groundbreaking "Voice Preview Edition," a revolutionary system designed to facilitate local, on-device voice processing, thereby eliminating the need to transmit sensitive voice data to external servers.
This localized processing model addresses growing privacy concerns surrounding commercially available voice assistants, which often transmit user utterances to remote servers for analysis and processing. By keeping the entire voice interaction process within the confines of the user's local network, Home Assistant's Voice Preview Edition ensures that private conversations remain private and are not subject to potential data breaches or unauthorized access by third-party entities.
The blog post further elaborates on the technical underpinnings of this new voice assistant system, emphasizing its reliance on open-source technologies and the flexibility it offers for customization. Users are afforded the ability to tailor the system's functionality to their specific needs and preferences, selecting from a variety of speech-to-text engines and wake word detectors. This granular level of control stands in stark contrast to the restricted customization options offered by commercially available solutions.
Moreover, the post highlights the collaborative nature of the project, inviting community participation in refining and expanding the capabilities of the Voice Preview Edition. This open development approach fosters innovation and ensures that the system evolves to meet the diverse requirements of the Home Assistant user base. The post underscores the significance of this community-driven development model in shaping the future of open-source voice assistants. Finally, the announcement stresses the preview nature of this release, acknowledging that the system is still under active development and encouraging users to provide feedback and contribute to its ongoing improvement. The implication is that this preview release represents not just a new feature, but a fundamental shift in how users can interact with their smart homes, paving the way for a future where privacy and user control are paramount.
The Hacker News post titled "The era of open voice assistants," linking to a Home Assistant blog post about their new voice assistant, generated a moderate amount of discussion with a generally positive tone towards the project.
Several commenters expressed enthusiasm for a truly open-source voice assistant, contrasting it with the privacy concerns and limitations of proprietary offerings like Siri, Alexa, and Google Assistant. The ability to self-host and control data was highlighted as a significant advantage. One commenter specifically mentioned the potential for integrating with other self-hosted services, furthering the appeal for users already invested in the open-source ecosystem.
A few comments delved into the technical aspects, discussing the challenges of speech recognition and natural language processing, and praising Home Assistant's approach of leveraging existing open-source projects like Whisper and Rhasspy. The modularity and flexibility of the system were seen as positives, allowing users to tailor the voice assistant to their specific needs and hardware.
Concerns were also raised. One commenter questioned the practicality of on-device processing for resource-intensive tasks like speech recognition, especially on lower-powered devices. Another pointed out the potential difficulty of achieving the same level of polish and functionality as commercially available voice assistants. The reliance on cloud services for certain features, even in a self-hosted setup, was also mentioned as a potential drawback.
Some commenters shared their experiences with existing open-source voice assistant projects, comparing them to Home Assistant's new offering. Others expressed interest in contributing to the project or experimenting with it in their own smart home setups.
Overall, the comments reflect a cautious optimism about the potential of Home Assistant's open-source voice assistant, acknowledging the challenges while appreciating the move towards greater privacy and control in the voice assistant space.
Summary of Comments ( 98 )
https://news.ycombinator.com/item?id=42734478
Hacker News users discussed the feasibility and implications of client-side scanning for CSAM in end-to-end encrypted systems. Some commenters expressed skepticism about the technical challenges and potential for false positives, highlighting the difficulty of distinguishing between illegal content and legitimate material like educational resources or artwork. Others debated the privacy implications and potential for abuse by governments or malicious actors. The "slippery slope" argument was raised, with concerns that seemingly narrow use cases for client-side scanning could expand to encompass other types of content. The discussion also touched on the limitations of hashing as a detection method and the possibility of adversarial attacks designed to circumvent these systems. Several commenters expressed strong opposition to client-side scanning, arguing that it fundamentally undermines the purpose of end-to-end encryption.
The Hacker News post "Let's talk about AI and end-to-end encryption" has generated a robust discussion with several compelling comments. Many commenters grapple with the inherent tension between the benefits of AI-powered features and the preservation of end-to-end encryption (E2EE).
One recurring theme is the practicality and potential misuse of client-side scanning. Some commenters express skepticism about the feasibility of truly secure client-side scanning, arguing that any client-side processing inherently weakens E2EE and creates vulnerabilities for malicious actors or governments to exploit. They also voice concerns about the potential for function creep, where systems designed for specific purposes (like detecting CSAM) could be expanded to encompass broader surveillance. The chilling effect on free speech and privacy is a significant concern.
Several comments discuss the potential for alternative approaches, such as federated learning, where AI models are trained on decentralized data without compromising individual privacy. This is presented as a potential avenue for leveraging the benefits of AI without sacrificing E2EE. However, the technical challenges and potential limitations of federated learning in this context are also acknowledged.
The "slippery slope" argument is prominent, with commenters expressing worry that any compromise to E2EE, even for seemingly noble purposes, sets a dangerous precedent. They argue that once the principle of E2EE is weakened, it becomes increasingly difficult to resist further encroachments on privacy.
Some commenters take a more pragmatic stance, suggesting that the debate isn't necessarily about absolute E2EE versus no E2EE, but rather about finding a balance that allows for some beneficial AI features while mitigating the risks. They suggest exploring technical solutions that could potentially offer a degree of compromise, though skepticism about the feasibility of such solutions remains prevalent.
The ethical implications of using AI to scan personal communications are also a significant point of discussion. Commenters raise concerns about false positives, the potential for bias in AI algorithms, and the lack of transparency and accountability in automated surveillance systems. The potential for abuse and the erosion of trust are recurring themes.
Finally, several commenters express a strong defense of E2EE as a fundamental right, emphasizing its crucial role in protecting privacy and security in an increasingly digital world. They argue that any attempt to weaken E2EE, regardless of the intended purpose, represents a serious threat to individual liberties.