20 Movies That Represent American Culture

Movies have always been one of America’s most revealing self-portraits, capturing how the nation sees itself and how it wants to be seen. From studio-era Westerns mythologizing frontier freedom to postwar dramas wrestling with conformity, film has functioned as a...