Efforts by servers hosting websites to counteract bots vary. Servers may choose to outline rules on the behaviour of internet bots by implementing a robots.txt file: this file is simply text stating the rules governing a bot's behaviour on that server. Any bot that does not follow these rules when interacting with (or 'spidering') any server should, in theory, be denied access to, or removed from, the affected website. If the only rule implementation by a server is a posted text file with no associated program/software/app, then adhering to those rules is entirely voluntary – in reality there is no way to enforce those rules, or even to ensure that a bot's creator or implementer acknowledges, or even reads, the robots.txt file contents. Some bots are "good" – e.g. search engine spiders – while others can be used to launch malicious and harsh attacks, most notably, in political campaigns.[2]
Bots are also used to buy up good seats for concerts, particularly by ticket brokers who resell the tickets.[12] Bots are employed against entertainment event-ticketing sites. The bots are used by ticket brokers to unfairly obtain the best seats for themselves while depriving the general public of also having a chance to obtain the good seats. The bot runs through the purchase process and obtains better seats by pulling as many seats back as it can.

At Facebook’s F8 Developers Conference, Messenger Bots were announced. These bots are being developed by media corporations and retailers alike and very quickly so which raises the question as to what a Messenger bot is and how it’s useful to so many different types of companies. Even more important to know is what these bots mean for the average user, whether or not they will always be safe or can they present a potential threat if they are developed by anyone with malicious intent. Here’s a the answer to all that and more.