Explanation: The reform to the American education system that began in the early nineteenth century was women's colleges. Around this time, more educational opportunities for women began to emerge with the founding of the first colleges that women could also attend.