Oh my God, American health care/health insurance is so depressing to read about. As a Canadian (and I realize that as soon as I type those words, someone out there is going PFFFFT at their computer screen), I just don't understand why and how Americans live with this level of expensive-ass, unsocialized health care. I feel like folks should be marching in the street for it. I go to Emerg, flash a card, and get care. Same with a clinic. Same with a doctor. Same with my yearly checkups and referral appointments. Some of my mental health care practitioners are covered. I don't get a bill, I don't pay down the road. It is shocking and shameful that Americans don't have the same.