Because I'm really getting sick of minorities pushing away their own damn people for someone who is white. I don't care if you're into someone white, but if you can't find the beauty in your own people, what the hell does say about you? Has Hollywood fucked you up that much? Also, many of you have been embracing homosexuality. Back in the day, black people used to kick the shit out of fags, now you're growing more accepting of it?