I filter out any user agents that are invalid, but there's no way to see which are real or faked. The access logs include the useragent of every single site visitor - not only errors/bad actors.
Hm! It could just be a parsing error. The useragent in question is like this: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/108.0.0.0 Safari/537.36 Edg/108.0.1462.41
> most scraping tasks require either desktop or mobile useragents and not both together
Why? Do some sites serve completely different content? Or it's simply markup differences? Have never done much scraping and I'd expect the viewport size to be the decisive factor these days, not the user agent. But, again, I don't know much about that.
I would accept this argument if the sample was unbiased but noisy. In this case it's extremely biased but (potentially) low in noise.
If people from Uganda aren't part of the target audience of this site, we won't get Ugandan user agents even if they happen to be a fair chunk of web users worldwide (certainly more than in my small but high-tech country.)
Thanks! And yep, fair comment, and I had noticed this as well even more so in last week's list. I have been thinking about how I could adjust the numbers in some way to counteract this or add another data source.