Is it normal, it makes me happy to see white people die on television?
Whenever I see white people on television die, I become very happy. Their smug know-it-all faces, their horrible arrogant beliefs, their whole culture upsets me. Is this normal. Minorities are the first to die in most movies and television shows, and whenever I see whitey die, it brightens my day.