Skip to Main Content
HBS Home
  • About
  • Academic Programs
  • Alumni
  • Faculty & Research
  • Baker Library
  • Giving
  • Harvard Business Review
  • Initiatives
  • News
  • Recruit
  • Map / Directions
Faculty & Research
  • Faculty
  • Research
  • Featured Topics
  • Academic Units
  • …→
  • Harvard Business School→
  • Faculty & Research→
Publications
Publications
  • 2025
  • Working Paper
  • HBS Working Paper Series

Enhancing Treatment Effect Prediction on Privacy-Protected Data: An Honest Post-Processing Approach

By: Ta-Wei Huang and Eva Ascarza
  • Format:Print
  • | Language:English
  • | Pages:91
ShareBar

Abstract

As firms increasingly rely on customer data for personalization, concerns over privacy and regulatory compliance have grown. Local Differential Privacy (LDP) offers strong individual-level protection by injecting noise into data before collection. While effective for safeguarding privacy, LDP can severely degrade data utility. This paper examines its impact on an important marketing practice: personalized interventions through Conditional Average Treatment Effect (CATE) prediction. We show that LDP introduces model-dependent bias and variance distortions in CATE estimates, which vary across covariates and can impair treatment prioritization, reduce predictive accuracy, and lead to inefficient resource allocation. Existing correction methods from the measurement error literature are often inadequate in high-dimensional, privacy-constrained settings where flexible models are used. To address these challenges, we propose a model-agnostic post-processing approach that refines CATE predictions while preserving LDP guarantees. Our method leverages an unbiased but noisy proxy for the true CATE and applies an iterative boosting procedure to better align predictions with this proxy. A major challenge is the high approximation error of the proxy variable due to LDP noise, which can lead to overfitting in standard post-processing. To mitigate this, we introduce an honest algorithm that combines sample splitting with a novel subgroup cross-learning technique to prevent overfitting. Through simulation studies and real-world marketing applications, we show that our method significantly improves both CATE prediction and treatment prioritization under LDP protection. Our findings offer a practical solution for organizations seeking to balance personalization and privacy, positioning post-processing as a powerful tool in navigating the privacy–utility trade-off.

Keywords

Targeted Intervention; Conditional Average Treatment Effect Estimation; Differential Privacy; Honest Estimation; Post-processing; Analytics and Data Science; Consumer Behavior; Marketing

Citation

Huang, Ta-Wei, and Eva Ascarza. "Enhancing Treatment Effect Prediction on Privacy-Protected Data: An Honest Post-Processing Approach." Harvard Business School Working Paper, No. 24-034, December 2023. (Revised March 2025.)
  • SSRN
  • Read Now

About The Author

Eva Ascarza

Marketing
→More Publications

More from the Authors

    • April 2025
    • Faculty Research

    Tabby: Winning Customers' Digital Wallets

    By: Eva Ascarza
    • March 2025
    • Faculty Research

    Unintended Consequences of Algorithmic Personalization

    By: Ayelet Israeli and Eva Ascarza
    • 2025
    • Faculty Research

    Dynamic Personalization with Multiple Customer Signals: Multi-Response State Representation in Reinforcement Learning

    By: Liangzong Ma, Ta-Wei Huang, Eva Ascarza and Ayelet Israeli
More from the Authors
  • Tabby: Winning Customers' Digital Wallets By: Eva Ascarza
  • Unintended Consequences of Algorithmic Personalization By: Ayelet Israeli and Eva Ascarza
  • Dynamic Personalization with Multiple Customer Signals: Multi-Response State Representation in Reinforcement Learning By: Liangzong Ma, Ta-Wei Huang, Eva Ascarza and Ayelet Israeli
ǁ
Campus Map
Harvard Business School
Soldiers Field
Boston, MA 02163
→Map & Directions
→More Contact Information
  • Make a Gift
  • Site Map
  • Jobs
  • Harvard University
  • Trademarks
  • Policies
  • Accessibility
  • Digital Accessibility
Copyright © President & Fellows of Harvard College.