Fifty years ago, critic Lionel Trilling declared that "in the United States at this time, liberalism is not only the dominant but even the sole intellectual tradition." Today, however, even most Democrats avoid calling themselves liberal. What happened to the liberal tradition in the second half of the twentieth century? What does liberalism stand for at the beginning of the twenty-first century? Can liberals reclaim their once-dominant position in American politics, or is their ideology history?