Score: 0

Evaluating Apple Intelligence's Writing Tools for Privacy Against Large Language Model-Based Inference Attacks: Insights from Early Datasets

Published: June 4, 2025 | arXiv ID: 2506.03870v1

By: Mohd. Farhan Israk Soumik, Syed Mhamudul Hasan, Abdur R. Shahid

Potential Business Impact:

Makes writing tools hide your feelings from computers.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

The misuse of Large Language Models (LLMs) to infer emotions from text for malicious purposes, known as emotion inference attacks, poses a significant threat to user privacy. In this paper, we investigate the potential of Apple Intelligence's writing tools, integrated across iPhone, iPad, and MacBook, to mitigate these risks through text modifications such as rewriting and tone adjustment. By developing early novel datasets specifically for this purpose, we empirically assess how different text modifications influence LLM-based detection. This capability suggests strong potential for Apple Intelligence's writing tools as privacy-preserving mechanisms. Our findings lay the groundwork for future adaptive rewriting systems capable of dynamically neutralizing sensitive emotional content to enhance user privacy. To the best of our knowledge, this research provides the first empirical analysis of Apple Intelligence's text-modification tools within a privacy-preservation context with the broader goal of developing on-device, user-centric privacy-preserving mechanisms to protect against LLMs-based advanced inference attacks on deployed systems.

Country of Origin
🇺🇸 United States

Page Count
29 pages

Category
Computer Science:
Machine Learning (CS)