Growing up seems easier for men, maybe because their rites of passage are clearer. They perform acts of bravery on the battlefield or show they're men through physical labor or by making money. For women, it's more confusing. We have no rites of passage. Do we become women when a man first makes love to us? If so, why do we refer to it as a loss of virginity? Doesn't the word 'loss' imply that we are better off before? I abhor the idea that we become women only through the physical act of a man. No, I think we become women when we learn what is important in our lives, when we learn to give and to take with a loving heart.