By Wesley J. Smith, First Things - Until recently, healthcare was not culturally controversial. Medicine was seen as primarily concerned with extending lives, curing diseases, healing injuries, palliating symptoms, birthing babies, and promoting wellness—and hence, as a sphere in which people of all political and social beliefs were generally able to get along.