Quote:
Originally Posted by Guest
Wow, you either did not grow up in America or you were so spoiled that you do not appreciate our great country and the greatest benefit to the world. Or, maybe you are just having a bad day when you wrote this.
|
The "Constitution" didn't do all these things...WHITE MEN did. Just as they did throughout Europe.
America "became the greatest" after Germany lost the title...and Great Britain, and France, and Spain, Portugal, The Netherlands, and Italy...before that. The Greeks, Romans, and Egyptians...ALL white people creating great empires and changing the world.
What did negros do? The American Indians? They had no wheel, no written language, they were CAVEMEN when we found them. Blacks aren't even the same species, how can you compare them?
America became great when others fell and we were the only one's left standing after WWII. And then...in the 1960s...we began our decline. The growing of the minorities...like cattle...grown and warehoused in the inner cities. They just became the majority...God help us. The decay has started in the cities and is working it's way out.
Disagree?