Bots. Nick Monaco

Bots - Nick Monaco


Скачать книгу
popular notion of the bot – that of the heavily argumentative troll account or inauthentic social media account operated by a human – is technically not necessarily an automated bot at all: these accounts are more correctly called “sock puppet” or “troll” accounts (Gorwa & Guilbeault, 2018). (Note that the meaning of the terms bot, troll, and sock puppet may differ significantly from language to language. For instance, in Polish, many speakers use “bot” and “troll” interchangeably to indicate a manipulative online social media account, whether automated or manually controlled (Gorwa, 2017a).)14

      We highlight these misuses and ambiguities in order to help the reader clearly understand what the term “bot” may mean when encountered in the wild. In this book, when we use the term bot, we will always be referring to a program that is partially or fully automated.

      Finally, there are a range of bot characteristics that can be used to describe a bot’s behavior or evaluate its intentions (Maus, 2017).

       Transparency – does the bot clearly state that it is an automated agent, or does it attempt to hide its automation, playing itself off as human?

       Degree of automation – is the bot automated all of the time? Do some of its actions only occur with human intervention? Can a human operate the bot while it is also performing other operations autonomously? (These questions all relate to the relative “cyborg-ness” of the bot.)

       Coordination with other bots – does this bot operate as part of a botnet or with other deceptive human users?

       Interaction and passivity – does this bot interact with or engage with human users in any way (likes, retweets, shares, conversation, etc.)? Are other users aware that the bot is present in the online environment? Does it silently surveil or collect data on other users or websites?

       Intent – what is the goal of this bot’s behavior?

       Politicization – is this bot engaged in political or social messaging?

      We have only briefly touched on the history of technology and the internet here. Any account of internet history is necessarily incomplete; it not only highlights or neglects the importance of individual activists, contributors, developers, policymakers, companies depending on when it is told, but it also changes so rapidly that it is difficult to capture accurately. Websites and platforms come and go. Services like Geocities, Alta Vista, and AskJeeves that were well known and widely used in the 1990s are virtually unknown today among most young internet users. Similarly, the social media platforms that have dominated internet usage since the mid 2000s may disappear. The internet’s future landscape may be defined by different services and companies, ones that are inconceivable today. As one technology historian puts it, “the history of computing […] can never be written” (Ceruzzi, 2012, p. ix). In a sense, the history of the internet is always in beta development: a product that is always taking shape, never in its final form.

      1  1 Chatbot has now become the most common term for conversational agents, but “chatterbot,” a term coined by famous bot builder Michael Mauldin in the 1990s, was then a common term for the same phenomenon (Deryugina, 2010; Leonard, 1997, p. 4).

      2  2 ELIZA’s pattern matching and substitution methods work in the same way that “regular expressions” or “RegEx” work today. RegEx is a formal language for pattern-matching in human language. It is often used to help computers search for and detect pre-defined patterns in human language (words, sentences, phone numbers, URLs, etc.) as a pre-processing step for modern natural language processing programs (Jurafsky & Martin, 2018, pp. 9–10).

      3  3 The fundamental importance of bots as a sense-making and infrastructural part of the internet is one of the primary reasons why laws or regulations advocating a blanket ban on bots would destroy the modern web as we know it.

      4  4 Of course, modern computer networks and the internet use dozens of protocols in addition to HTTP, such as the transmission control protocol (TCP), internet protocol (IP), and simple mail transfer protocol (SMTP) – all used millions of times every day. All of these small parts are necessary cogs that make up the machinery of the modern internet (Frystyk, 1994; Shuler, 2002).

      5  5 The first graphical MUDs did not begin to appear until the mid 1980s (Castronova, 2005, p. 55).

      6  6 Early web indexing bots were also called by other names, including “wanderers,” “worms,” “fish,” “walkers,” “knowbots,” and “web robots,” among others (Gudivada et al., 1997).

      7  7 While HTML is the primary language that web developers use to build webpages, other languages, such as CSS and JavaScript, provide very important secondary functions for websites and are essential building blocks for modern websites.

      8  8 Today, the ubiquity of web-indexing crawler bots on the World Wide Web are one aspect of what distinguishes the everyday internet from what is known as the “dark web.” In contrast, the dark web is an alternative form of the internet, which requires additional software, protocols, and technical knowledge to access – Tor, I2P, Freenet, ZeroNet, GNUnet, are but a few of the possible tools that can be used to access the dark web (Gehl, 2018). While security enthusiasts, researchers, and cybersecurity firms can build crawler bots to explore the dark web, large-scale centralized search engines like Google do not exist on the dark web. Navigation of the dark web is therefore mainly conducted through word-of-mouth within small communities, or on smaller scale search engines that resemble the “web directories” of the early internet. Much of the dark web is made of sites that are not indexed by crawler bots at all. Much of the activity that takes place on the dark web is meant to be clandestine (such as online crime, illegal marketplaces, and censorship circumvention websites and tools). The dark web also allows users and publishers to remain anonymous online.

      9  9 Martijn Koster, one of the most prominent bot developers and bot thinkers in the 1990s, also built a database of known crawlers (or “Web Robots,” in the parlance of the times) that is still online (Koster, n.d.).

      10 10 The “deep web” is also a concept worth noting. Deep web sites are sites that require special permissions (such as a password) to access and cannot be read or seen without that access. For example, while facebook.com itself is indexable and readable from all search engines, particular Facebook users’ profiles and posts may not be visible in search engine results due to individual privacy settings. So, while facebook.com itself is part of the clear web, unviewable and unindexed profiles would qualify as being part of the deep web.

      11 11 The Robot Exclusion Standard is also known as the


Скачать книгу