Sunday, May 8, 2011

How did American Imperialism effect Japan After WW2?

Question:




Answer:


America basically westernize Japan

No comments:

Post a Comment