American slavery wasn’t just a white man’s business—new research shows how white women profited, too

As the United States continues to confront the realities and legacy of slavery, Americans continue to challenge myths about the country’s history. One enduring myth is that slavery was a largely male endeavor—that for the most part, the buying, selling, trading and profiting from enslavement were carried out by white men alone.

This post was originally published on this site

Skip The Dishes Referral Code

More In Finance