I’ve said this before, I’ll say it again; I really, really dislike Westerns.
For starters, I hate the West (never mind that I live there; that’s a different post); it’s the most ugly and unphotogenic region captured on film. I see no reason to revel in a landscape that’s mostly barren and dead.
But more importantly, the movies themselves are generally repugnant; they mostly celebrate the ugly side of American history while ignoring the complexity of Manifest Destiny. Indians are always pure evil, and Americans are always the victims of violence, and the historical whitewashing only serves to piss me off.