Answer :
This is technically false. Although the end of World War II brought with it a complete rearrangement of Europe, and far more interdependence between European nations than before the war, there were still small cravings of imperialism from several European states.
I believe the answer is false, World War II did not cause the end to European imperialism.