I'm so fucking sick of it. Every tv show has a family of niggers or faggots walking around like it's normal, and there isn't a single decent Christian-family show on mainstream TV anymore. It's like God doesn't even exist in popular culture. Niggers in the streets looting and shouting about some other nigger dying because he was high as a kite on whatever drugs he was on.
You've got all these backwards beliefs being forced down our throats over and over from all sides. What happened to my America? When did it get this bad? Can Trump do anything about this?
You can't be angry over this my dude. Are they sending hit teams to your house to kill you? No. They only make propaganda and you are just mad that none of the propaganda is for you anymore. You don't need TV shows to live. You pay the Netflix sub, you watch the ads... this whole system you are complaining about is only propped up by you and everyone else who is too attached to walk away from it.