Leeky Forums I want your opinion on this matter, I know I won't get a "real" answer anywhere else.
My dad owned a pet store growing up, I spent a lot of time there as a child and worked there in my teens. We had a backroom with all the exotic critters like sneks, lizards, tarantulas, scorpions. Every time a black person walked back there, and I mean every. Fucking. Time. Like clockwork: "AWWWW HEEELLLLLL NAW!"
I recently got a dog, I was walking in a park and most white people would smile and greet my dog "hey pooch! That's a beautiful dog you've got there." Or something along those lines. Every nigger who crossed my path would pull their children away "get that dog away from me!"
The other day, I was at school and had my dog in the car and a basketball-American asked me for a ride home, "sure, why not? You're not scared of dogs are you?"
"They're my biggest fear"
Then he has the gall to ask me to tie up the dog in the back of the car. Like he gets to dictate the terms of my generosity. (Don't worry Leeky Forums I was CCing a glock.)
Anyways, I just wanted to get y'alls take on this phenomenon. Personally, I'm leaning towards some hardwired animalistic instinct combined with a general lack of empathy. Thoughts?