Actor Michael Jai White has decried the decline of masculinity in America. On an episode of The Joe Rogan Experience, White and Rogan discussed their belief that masculinity is no longer important in ...