Western nations have dominated much of the world for nearly five centuries. During this era, European nations – and later, the United States – gained power and spread the influence of Western civilization over much of the globe. China and Japan were forcibly opened to European and American trade. Africa, India, and much of Asia were overwhelmed and carved up ...