In a world filled with fear and confusion about diagnoses and medication, why are we not taught that the body can heal itself? Why are nature's remedies routinely ignored or ridiculed by the medical establishment? Could there be a motive to keep us sick?