Score contribution per author:
α: calibrated so average coauthorship-adjusted count equals average raw count
This paper studies the quality of portfolio performance tests based on out-of-sample returns. By disentangling the components of the out-of-sample performance, we show that the observed differences are driven largely by the differences in estimation risk. Our Monte Carlo study reveals that the puzzling empirical findings of inferior performances of theoretically superior strategies result mainly from the low power of these tests. Thus, our results provide an explanation as to why the null hypothesis of equal performance of the simple equally-weighted portfolio compared to many theoretically-superior alternative strategies cannot be rejected in many out-of-sample horse races. Our findings turn out to be robust with respect to different designs and the implementation strategies of the tests.