Does organic mean healthier?
Posted at: 09/04/2012 6:58 AM
A new study suggests organic food doesn't necessarily mean healthier.
The popularity of organic food and drinks has skyrocketed but does the label mean it's better for you? Researchers at Stanford University found that conventional and organic food have the same vitamin content.
So while taste and concerns about pesticides may still influence grocery shoppers, regular fruits, veggies and milk may be just as healthy.