Of course there is, just like blacks disproportionately murder, rob, and rape whites and Asians. Just like many hispanics and blacks don't like each other. Just like blacks are VERY racist towards whites. etc etc. It's human nature and it isn't going to change.
But are blacks somehow oppressed? No, they're given more advantages all things being equal. A black with similar qualifications will always get more scholarships, get into a better school, or get the promotion. The only thing holding them back is their own self destructive culture. Every other minority does fine here.
Name ONE place in the world where blacks have built a place you'd want to raise your family? From Oakland to Liberia, Detroit to South Africa, there is human misery anywhere their culture is dominant. Does that mean they're all bad? Of course not. But here they have a systematic culture of drug abuse, violence, child abandonment, and thug worship that will always keep them down. No one has ever improved their lot in life by blaming their problems on others. If it's racist to point out the truth then so be it. I'd rather be "racist" than live in denial of the obvious. Enabling others to continue living self destructive lifestyles is disgraceful, childish, and foolish.