Experts blame Hollywood for promoting a false history of cowboy culture. The cowboy culture of Hollywood is not the cowboy culture of America -- the thousands of Black, Hispanic and Indigenous cowboys ...