These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
This position paper argues that achieving robustness, privacy, and efficiency
simultaneously in machine learning systems is infeasible under prevailing
threat models. The tension between these goals arises not from algorithmic
shortcomings but from structural limitations imposed by worst-case adversarial
assumptions. We advocate for a systematic research agenda aimed at formalizing
the robustness-privacy-efficiency trilemma, exploring how principled
relaxations of threat models can unlock better trade-offs, and designing
benchmarks that expose rather than obscure the compromises made. By shifting
focus from aspirational universal guarantees to context-aware system design,
the machine learning community can build models that are truly appropriate for
real-world deployment.