Authors: Barbara Ehrenreich and Deirdre English | PDF file size: 817 KB
Women have always been healers. They were the unlicensed doctors and anatomists of western history. They were abortionists, nurses and counsellors. They were pharmacists, cultivating healing herbs and exchanging the secrets of their uses. They were midwives, travelling from home to home and village to village. For centuries women were doctors without degrees, barred from books and lectures, learning from each other, and passing on experience from neighbour to neighbour and mother to daughter. They were called “wise women” by the people, witches or charlatans by the authorities. Medicine is part of our heritage as women, our history, our birthright.
Today, however, health care is the property of male professionals. Ninety-three percent of the doctors in the US are men; and almost all the top directors and administrators of health institutions. Women are still in the overall majority — 70 percent of health workers are women — but we have been incorporated as workers into an industry where the bosses are men. We are no longer independent practitioners, known by our own names, for our own work. We are, for the most part, institutional fixtures, filling faceless job slots: clerk, dietary aide, technician, maid….
“To know our history is to begin to see how
to take up the struggle again!”