Monday, June 16, 2025

Differential privateness on belief graphs

Differential privateness (DP) is a mathematically rigorous and extensively studied privateness framework that ensures the output of a randomized algorithm stays statistically indistinguishable even when the info of a single person modifications. This framework has been extensively studied in each concept and follow, with many functions in analytics and machine studying (e.g., 1, 2, 3, 4, 5, 6, 7).

The 2 primary fashions of DP are the central mannequin and the native mannequin. Within the central mannequin, a trusted curator has entry to uncooked information and is answerable for producing an output that’s differentially non-public. The native mannequin requires that each one messages despatched from a person’s machine are themselves differentially non-public, eradicating the necessity for a trusted curator. Whereas the native mannequin is interesting attributable to its minimal belief necessities, it typically comes with considerably greater utility degradation in comparison with the central mannequin.

In real-world data-sharing eventualities, customers typically place various ranges of belief in others, relying on their relationships. For example, somebody would possibly really feel comfy sharing their location information with household or shut mates however would hesitate to permit strangers to entry the identical info. This asymmetry aligns with philosophical views of privateness as management over private info, the place people specify with whom they’re keen to share their information. Such nuanced privateness preferences spotlight the necessity for frameworks that transcend the binary belief assumptions of present differentially non-public fashions, accommodating extra lifelike belief dynamics in privacy-preserving techniques.

In “Differential Privateness on Belief Graphs”, printed on the Improvements in Theoretical Laptop Science Convention (ITCS 2025), we use a belief graph to mannequin relationships, the place the vertices signify customers, and related vertices belief one another (see under). We discover how you can apply DP to those belief graphs, guaranteeing that the privateness assure applies to messages shared between a person (or their trusted neighbors) and everybody else they don’t belief. Particularly, the distribution of messages exchanged by every person u or considered one of their neighbors with some other person not trusted by u needs to be statistically indistinguishable if the enter held by u modifications, which we name belief graph DP (TGDP).

Related Articles

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Latest Articles