Our food (including fruits & vegetables) have been experimented on so much that there are endless possibilities on what could have been put into our meals. The "going organic" movement has definitely helped businesses sell their products, but just because a piece of meat has the "Organic" stamp doesn't mean that they're healthier than vegetables, let alone better for your health.
In the United States, many people believe that anything under the USDA (United States Department of Agriculture) is healthy for their body. However, the definition of the word "organic" has changed several times; I even wrote a blog post describing the changes HERE. But back to organic being linked to great health: A poisonous berry could be organic, but the effects will still harm you. Same goes with meat. A piece of steak could be organic, but that doesn't mean that you're living a healthier lifestyle.
The only difference between organic meat and the meat that is commonly sold in markets is that 1.) Organic meat comes from livestock that is not injected with hormones, 2.) These animals have wide-open spaces to move around in, and 3.) Their diet consists of healthier vegetables and grains. Yet, a person will benefit much more from eating the same meals that are given to organic livestock than killing and eating the actual animal.
Long before hormones and poor diets were given to animals though, people would still develop heart disease, diabetes, cancer, and other illnesses. But that wasn't very common because people stilled lived agricultural lifestyles which included more fruits, vegetables, legumes, grains, and less meat and dairy. However, evidence on organic meat being unhealthy for people can be traced as far back as the 1700s, even earlier! High-class societies were the only people who could afford lavish meals with chicken, pork, steak, veil, fish, and heavy, meaty foods. The effects that these meals left on the human body were very noticeable.
Artist: Jim Gillray (1757 - 1815)
So, are organic meats better than organic vegetables? Definitely not. There have been "studies" proving that organic fruits and vegetables are no different than the ones with hormones and pesticides, but I've had my own organic garden before and I've seen, tasted, and felt the difference. Those studies are just run by paid scientists and doctors from meat, dairy, and poultry businesses to put healthy and beneficial foods out of the market.
Consider this vegetarian myth DEBUNKED! Like seeing vegetarian myths get solved? Stay tuned for next Fridays vegetarian myth! And feel free to share this post as well.
Want to read past vegetarian myths? Click on the links below:
Hope that everyone has a great Friday night!