Before we even get into this post, let’s define what feminism is and make sure we are all on the same page: fem·i·nism – noun: the theory of the political, economic, and social equality of the sexes. I used to see feminism as a dirty word. It was for all these liberal women who didn’t realize that…