ArXiv TLDR

Quotient Semivalues for False-Name-Resistant Data Attribution

🐦 Tweet
2605.07663

Florian A. D. Burnat, Brittany I. Davidson

cs.GTcs.CRcs.LG

TLDR

This paper introduces quotient semivalues to prevent false-name manipulation in ML data attribution, significantly reducing Sybil attack gains.

Key contributions

  • Formalizes false-name manipulation, where contributors inflate data value through pseudonymous identities or duplication.
  • Proposes quotient semivalues, computing data values over evidence-backed clusters to absorb within-cluster duplication.
  • Proves exact Shapley-fair attribution is incompatible with unrestricted false-name-proofness on reported identities.
  • Reduces Sybil attack manipulation gains from 1.74 to 0.96, near honest levels, using quotient semivalues.

Why it matters

Data attribution is crucial for fair compensation and auditing in ML, but current methods are vulnerable to strategic manipulation. This work provides a robust mechanism to ensure fair data valuation even when contributors attempt to inflate their share through false identities or data duplication. It significantly improves the integrity of data markets.

Original Abstract

Data valuation methods allocate payments and audit training data's contribution to machine-learning pipelines; however, they often assume passive contributors. In reality, contributors can split datasets across pseudonymous identities, duplicate high-value examples, create near-duplicates, or launder synthetic variants to inflate their share. We formalize this as false-name manipulation in ML data attribution. Our main construction is the quotient semivalue mechanism: compute Shapley-, Banzhaf-, or Beta-style values over evidence-backed attribution clusters instead of raw identities, using a canonical-representative operator to absorb within-cluster duplication. We prove an impossibility: on a fixed monotone data-value game, exact Shapley-fair attribution over reported identities is incompatible with unrestricted false-name-proofness, even on binary-valued instances, and characterize the split-gain of a general semivalue on a unanimity counter-example. The mechanism is exactly false-name-proof under two structural conditions: false-name-neutral within-cluster allocation and quotient-stable manipulations. Under imperfect provenance, when these conditions hold approximately, manipulation gain and fairness loss are bounded by three measurable quantities: escaped-cluster mass, value-estimation error, and clustering distance. We instantiate the mechanisms in DataMarket-Gym, a benchmark for attribution under strategic provider attacks. On synthetic classification tasks, quotient semivalues with example-level evidence reduce manipulation gain on duplicate and near-duplicate Sybil attacks from $1.74$ under baseline Shapley to $0.96$, near the honest level. The cosine-threshold and (false-merge, false-split) rate sweeps trace the corresponding fairness--Sybil frontier.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.