Re-colonize africa
Ever since the colonial powers left, Africa has gone downhill. The only country with anything approaching a civil society was South Africa (coincidently, the only white-governed nation in Africa) but with the rise of Mandela and his gang of thugs, even South Africa has fallen. It won't be popular to say, but Africa should be re-colonized for the benefit of all. Blacks will be appaled at the idea that they can't govern themselves, yet Africans already depend on the billions of dollars given to them each year from the west. No black-ruled nation has ever effectively governed itself. Call me a racist if you like, but search the history books. Not one has ever come even close to being a successful state.