Goodhart's Law and Machine Learning: A Structural Perspective

Social Science Research Network(2021)

引用 1|浏览0
暂无评分
摘要
We develop a structural framework illustrating how penalized regression algorithms affect Goodhart bias when training data is clean but covariates are manipulated at cost by future agents facing prediction models. With quadratic manipulation costs, bias is proportional to sum-of-squared slopes, micro-founding Ridge. Lasso is micro-founded in the limit under increasingly steep cost functions. However, standard penalization is inappropriate if costs depend upon percentage rather than absolute manipulation. Nevertheless, with known costs of either form, the following algorithm is proven manipulation-proof: Within training data, evaluate candidate coefficient vectors at their respective incentive-compatible manipulation configuration. Moreover, we obtain analytical expressions for the resulting coefficient adjustments: slopes (intercept) shift downward if costs depend upon percentage (absolute) manipulation. Statisticians ignoring agent borne manipulation costs select socially suboptimal penalization, resulting in socially excessive, and futile, manipulation. Model averaging, especially over Lasso or ensemble estimators, reduces manipulation costs significantly. Standard cross-validation fails to detect Goodhart bias.
更多
查看译文
关键词
machine learning,law,structural perspective
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要