World War II cost far more money than any other war in American history,[39] but it boosted the economy by providing capital investment and jobs, while bringing many women into the labor market. Among the major combatants, the United States was the only nation to become richer—indeed, ...