I would like to know why ALL democrats,left wingers,liberals whatever the hell you want to call yourself are so insistent on the US becoming a non white majority nation....you do realize this country was founded,fought for and built BY and FOR whites of good moral character and without white men and women the things we ALL enjoy we would not have and the high standard of living we have will only continue to erode the less white it becomes. Yet you democrats,liberals,left wingers seem to WANT that why is that? Some stupid sense of white guilt? You think the US deserves to be less white because its so far more advanced? what is it? The more I read from and about democrats,liberals and left wingers beliefs and comments on things the more I am convinced its a mental disease. No other explanation for it. So I am truly interested in knowing WHY.