Why don't we hear anything about non-black minorities in the media? Why don't they tell us about white privilege and systematic racism? I've got an impression blacks is the only ethnic group that finds it difficult to integrate into American society.
They're conditioned from birth to believe that they're different, that people are out to get them, that standards must be lowered for them, that they can depend on certain people to make excuses for them, that those same people will largely ignore any bad behavior.
Inflict that on them over multiple generations and it becomes ingrained in their culture.
That's a pretty fucked up thing to do to people, but that's where we are right now.
.