The majority of us—even those of us who oppose racism and embrace the idea that the U.S. is a nation of immigrants—hold the belief that America = white.