20240311-tricking_the_web.txt
I've recently been visiting a site and, as usual, the various browsers
I use provide various different user experiences. The site is coded
poorly, but I like it anyway as a resource. The in-site search engine
has GET vs. POST errors. There are also errors in images that are,
judging by the nomenclature, generated by request, which probably
wastes a ton of electricity and processor power.

When you use a site and drive up traffic or introduce "strange"
traffic, sometimes the webmaster's response is to blacklist you
thinking you're part of a botnet or something. Sometimes they
blacklist the UA string. I simply change the UA string and everything
works again (usually), which is the reason for the title.

In my experience, this is most often employed by news sites, The
reason? Their sites are perfect for TUI/JS-less consumption and that's
somewhat antithetical to online adware. The most ridiculous thing is
most of the time they will block off articles that are more than 24
hours old, despite the fact that they are viewable by anyone in that
tiny time frame.

It's just sad when a legitimate user has to trick and lie just to view
content freely available on the Web. Web 1.0 or 1.5 was much better
than the current massive dump truck of bytes and dynamic fat code the
current Web is.