All News

Apple Users Can Claim Settlement from $95 Million Siri Privacy Case

Apple has opened claims for a $95 million settlement related to a 2019 lawsuit over Siri inadvertently recording private conversations. Eligible US users who experienced unintended Siri activations between 2014 and 2024 can submit claims for up to $100 across five devices. Applications close July 2, 2025.

Published May 8, 2025 at 08:08 AM EDT in Cybersecurity

Apple users in the United States now have the opportunity to claim compensation from a $95 million settlement related to privacy concerns with Siri, Apple’s voice assistant. This settlement stems from a 2019 class action lawsuit alleging that Apple recorded private conversations without user consent when Siri was unintentionally activated.

Between September 17, 2014, and December 31, 2024, some Apple devices reportedly captured private conversations due to unintended Siri activations. These recordings were allegedly shared with third-party contractors for quality control, raising significant privacy concerns.

In response, Apple issued a formal apology and committed to no longer retaining user recordings. Despite denying allegations that it used this data for targeted advertising, Apple agreed in January 2025 to pay $95 million to affected users to settle the lawsuit.

Eligible users can submit claims through a dedicated website until July 2, 2025. Claimants may apply for up to five Siri-enabled devices, including iPhone, iPad, Apple Watch, Mac, HomePod, iPod touch, and Apple TV. Each device’s claim is capped at $20, with a maximum payout of $100 per user.

Applicants must swear under oath that Siri was unintentionally activated on each device. Notifications are being sent to users who received Claim Identification and Confirmation Codes, but anyone who believes they are eligible can submit a claim regardless of prior notification.

Broader Implications for Privacy and Voice Assistants

This settlement highlights growing concerns about privacy in voice-activated technologies. As voice assistants become more integrated into daily life, ensuring user consent and data protection is paramount. Companies must implement transparent policies and robust safeguards to maintain consumer trust.

For developers and businesses, this case underscores the importance of privacy-by-design principles in AI and voice assistant development. Proactively addressing potential privacy risks can prevent costly legal challenges and enhance brand reputation.

QuarkyByte’s expertise in cybersecurity and privacy compliance offers valuable insights into navigating these challenges. By leveraging our analysis, organizations can develop strategies that align with evolving regulations and user expectations in voice technology.

Keep Reading

View All
The Future of Business is AI

AI Tools Built for Agencies That Move Fast.

QuarkyByte offers in-depth analysis of privacy and security settlements like Apple’s Siri case. Explore how our insights help businesses strengthen data protection and navigate compliance challenges in voice assistant technologies. Discover actionable strategies to safeguard user privacy and build trust in your products.