are Japanese the only ones worthy of life?
blacks:
steal shit
shoot each other
whites:
colonize other countries
wave war against each other
arabs:
rape women
behead babies
shoot each other
meanwhile all that Japanese are doing is drawing cute anime girls and being the greatest country in the world. everything about their culture is about manners, humbleness, love, health, and behaving like a human in general. it seems like they're the only ones that don't live like animals, like us - just to eat and sleep and insult each other. are they the true master race?