States' rights is the political doctrine that emphasizes the powers and rights of individual states over the federal government. This concept played a pivotal role in the pre-Civil War era, as many Southern states believed they had the right to govern themselves, especially regarding issues like slavery and tariffs. The debate over states' rights became a significant factor contributing to tensions between the North and South, ultimately leading to the outbreak of the Civil War.