All of what you said was true. Scientists try and find "original sources" for data, but they're usually analyzing measurements that have been altered in some way before they receive them... usually in the measurement hardware itself. One can't really avoid it completely. The flip side is that noise can actually be helpful, especially in situations where the signal itself is too weak to reach some detectable threshold on its own.
I generally try hard to remove bias/noise from data before analyzing it, but I also add noise back in later on... e.g. jittering data for plots, or putting a noise feature into a model to establish a threshold for useless features. I'm also a bit miffed that I had to discover this all on my own... Noise is useful and interesting! In business-oriented analysis, noise can be even more important, since changes in "data noise" can indicate important changes in business patterns, or problems with how they're collecting data (companies seldom share this information with analysts beforehand). I'd go so far as to say that understanding the noise in a dataset is the primary concern for most of my analyses these days.
I generally try hard to remove bias/noise from data before analyzing it, but I also add noise back in later on... e.g. jittering data for plots, or putting a noise feature into a model to establish a threshold for useless features. I'm also a bit miffed that I had to discover this all on my own... Noise is useful and interesting! In business-oriented analysis, noise can be even more important, since changes in "data noise" can indicate important changes in business patterns, or problems with how they're collecting data (companies seldom share this information with analysts beforehand). I'd go so far as to say that understanding the noise in a dataset is the primary concern for most of my analyses these days.
Here's a good book with more pop-sci noise examples: https://www.amazon.com/Noise-Bart-Kosko/dp/0670034959