We propose and analyze algorithms to solve a range of learning tasks under
user-level differential privacy constraints. Rather than guaranteeing only the
privacy of individual samples, user-level DP protects a user's entire
contribution ($m \ge 1$ samples), providing more stringent but