Hey all. It's been a dream of mine of sorts to live in Florida. Something about palm trees, beaches, and the ocean that make me happy. Also, the prospect of not having to deal with winter anymore.
Anyway, I was wondering if it's as bad as it's been portrayed here. I'm lucky enough right now to be in northern Minnesota, an exclusively white town with no crime to speak of. I'll have my HVAC degree in a year and was planning on putting it to use in FLA but all the stories and newslinks with the rampant nig crime have got me thinking. For the most part it's too cold for the nigs up here. One thing I've learned in my life is that nigs, hate, cold!
Any thoughts from the Floridians of the bunch? Thanks!
"Is there a moron factory working double shift to produce you stream of idiots?"--The Almighty Master Supreme Mayo Alex Linder