It would seem that almost all college programs include some kind of history course that tries to lure you into a certain way of thinking. I'm taking a genocide class, and the professor keeps on pushing the idea that "race is a myth". I simply won't except this. She (the professor) says that race hasn't been scientifically proven. Well, there has to be some reason why black people, usually, have higher blood pressure than whites and why blacks murder had a much higher rate than whites. Scientists even proved blacks have tinier brains! I wrote a paper in response to a reading we did on this very subject, she admitted my piece was quite good, but I sure I did not sway her. My professor is Hispanic by the way. She speaks intelligently, but talking about this kind of thing sounds crazy!
Why is it that all schools want to push everybody into the foolish mindset that we're all the same. We are not! There's a reason why we have religions, nationalities and cultures! I am not the same as black murdering beasts.
What they are teaching in this class is a very liberal concept, just like what is taught in the majority of American and European colleges. Why must we all get along? I'm willing to, but blacks don't seem too willing, juding by all the whites they murder. I don't want to get along with these animals if that's the case.