The history of women in the United States encompasses the lived experiences and contributions of women throughout American history. The earliest women living in what is now the United States were Native Americans. European women arrived in the 17th century and brought with them European culture and values. During the 19th century, women were primarily restricted to domestic roles in keeping ...
Source