The Sex-Bias Myth in Medicine

Though it is commonly believed that American health-care delivery and research benefit men at the expense of women, the truth appears to be exactly the opposite