Abstract
In spite of significant research efforts, argument technologies do not seem poised to scale up as much as most commentators would hope or even predict. In this paper, I discuss what obstacles bar the way to more widespread success of argument technologies and venture some suggestions on how to circumvent such difficulties: doing so will require a significant shift in how this research area is typically understood and practiced. I begin by exploring a much broader yet closely related question: To what extent are people natively good at arguing? This issue has always been central to philosophical reflection and it has become even more urgent nowadays, with the explosion of persuasive technologies and unprecedented opportunities for large-scale social influence. The answer hinges on what aspect of argumentation is taken under consideration: evidence suggests that people are relatively bad at analyzing the structure of arguments, especially when these are presented out of context and in abstract terms; in contrast, data show that even laymen tend to excel in the interactive practice of argumentation, in particular when motivation is high and something significant is at stake. Unfortunately, current argument technologies are more closely tailored to the former type of activity than to the latter, which is the main reason behind their relative lack of success with the general public. Changing this state of affair will require a commitment to ecological argument technologies: that is, technologies designed to support real-time, engaging and meaningful argumentative interactions performed by laypeople in their ordinary life, instead of catering to the highly specific needs of a minority of niche users (typically, argumentation scholars).
Similar content being viewed by others
Notes
“Argument”, retrieved December 28, 2015, from http://www.merriam-webster.com/dictionary/argument
CMNA, Computational Models of Natural Arguments, since 2001 (http://www.cmna.info/); ArgMAS, Argumentation in Multi-Agent Systems, since 2004 (http://www.mit.edu/~irahwan/argmas/); UM4Motivation, User Models for Motivational Systems, in 2011 and 2012 (http://cgi.csc.liv.ac.uk/~floriana/UM4Motivation2/Home.html)
COMMA, Computational Model of Arguments, since 2006 (http://www.comma-conf.org/)
Just to mention a few: ASPIC, Argumentation Service Platform with Integrated Components (http://cordis.europa.eu/ist/kct/aspic_synopsis.htm); DAM, Dialectical Argumentation Machines (http://www.arg-tech.org/index.php/projects/dialectical-argumentation-machines/); DYNARG, The Dynamics of Argumentation (http://icr.uni.lu/dynarg/DYNARG/Home.html); and the recent EPSRC large grant on Argument Mining (http://gow.epsrc.ac.uk/NGBOViewGrant.aspx?GrantRef=EP/N014871/1)
The EU COST action Agreement Technologies (http://www.agreement-technologies.eu/)
AT, International Conference on Agreement Technologies, since 2012 (http://www.agreement-technologies.eu/resources/at-conference-series)
The notion of ecological argument technologies detailed in this paper refers to the broader concept of ecological rationality, as developed by Gigerenzer and Selten (2001) and Gigerenzer et al. (2011). The ties between the present proposal and this line of work will be further discussed later on: for now, let us just say that an argumentative technology, in order to be ecological in the required sense, ought to be designed to match the argumentative skills human users developed in response to the ecological pressures they face in everyday life.
Non-ecological argument technologies may still be very valuable for dedicated purposes, e.g., education, and may even provide guidance on how to design more productive and sustainable platforms and tools. What they cannot do, however, is to scale up, as long as they fail to meet the ecological challenge. This is the claim articulated in this paper, which of course does not deny the potential usefulness of argument technologies also in more restricted domains.
Both the universality and the incorrigibility claims are tied to the tradition of classifying fallacies at a high level of abstraction, as idealized inference patterns—a tradition criticized by van Eemeren and Grootendorst (1995), among others. In contrast, situated learning theory (Lave and Wenger 1991) has provided evidence that otherwise common errors occur less frequently (and are more varied) in concrete problem-solving situations, as opposed to abstractly defined problems. An echo of this line of work can be found in recent attempts to articulate more nuanced versions of fallacy theory, which will be discussed in Section 2.2 (see also Boudry et al. 2015; Paglieri 2016).
The key documents to consult, in order to get a sense of the dominant institutional view on critical thinking education in Italy, are the current guidelines on the curriculum of Italian kindergartens and primary and secondary schools (“Indicazioni nazionali per il curricolo della scuola dell’infanzia e del primo ciclo di istruzione,” http://www.indicazioninazionali.it/documenti_Indicazioni_nazionali/indicazioni_nazionali_infanzia_primo_ciclo.pdf), issued by the Italian Ministry of Education and Research (MIUR) in September 2012. In that document, critical thinking is identified as a priority across all disciplines, in spite of the utter lack of (i) dedicated critical thinking training for teachers and (ii) resources to support the proclaimed shift towards more rigorous critical thinking education in Italian schools.
Source: http://www.admissionstestingservice.org/images/47832-tsa-test-specification.pdf (last consulted on December 27, 2015)
Source: http://www.admissionstestingservice.org/for-test-takers/thinking-skills-assessment/tsa-cambridge/about-tsa-cambridge/ (last consulted on December 27, 2015)
Source: http://www.merton.ox.ac.uk/admissions-feedback-economics-and-management (last consulted on December 27, 2015)
Given all these differences across groups, in terms of test materials, sample size, methodology, context, etc., running statistical analyses on these data would be rather uninformative. Thus, statistical details will be kept to a minimum and provided only for those comparisons that are methodologically appropriate. For the same reason, these findings are meant here only to illustrate the difficulties that various sorts of people encounter in dealing with abstract argument analysis.
Of course, the fact that we are forced to admit the existence of non-fallacious fallacies is just another indication of the theoretical inadequacy of the standard notion of a fallacy: a category designed to capture erroneous reasoning ends up including in its extension also acceptable forms of inference. But what to do about the problematic status of fallacy theory is a topic for another day (for discussion, see Woods 2013; Paglieri 2016).
These results should not be confused with the claim that people are “good with probabilities,” since there is ample evidence they are not—witness the garden variety of well-known biases of probabilistic reasoning, e.g., gambler’s fallacy (Tversky and Kahneman 1971), conjunction fallacy (Tversky and Kahneman 1983), and base rate neglect (Tversky and Kahneman 1982). What these findings show, instead, is that people track the quality of argument based on the same factors that would be relevant if Bayesian update was used: the observed similarity is at the level of the outcome, with no claim being made on a corresponding similarity of mechanisms—indeed, people are certainly not doing Bayesian computation as part of their explicit reasoning. The fact that we get the outcome right when evaluating each other’s arguments, while we do not in solitary reasoning tasks of the sort used by Tversky and Kahneman, provides further support to the argumentative theory of reasoning.
Similar considerations apply also to (and have received much more attention in) instructional design in computer-supported collaborative learning, as documented by scripting studies; for a discussion of several applications to argument education, see Weinberger et al. (2007).
While this may seem reminiscent of Facebook likes, it is actually not—or, more precisely, it aims to achieve the same level of intuitive appeal but for entirely different purposes. What prevents Facebook likes from being used as argumentative indicators is their ambiguity of meaning: by liking a post, a comment, a photo, or anything else, a Facebook user may express a variety of communicative intentions—approval of the contents, support, or solidarity for the author of the contents, approval for the fact that the author of the post decided to make it public, hilarity prompted by the contents or their posting, and more. Argument technologies should strive to provide users with tagging options that are as appealing as Facebook likes but with a much better defined semantics.
Interestingly, expecting users to dwell too much on argument structure may be not only unrealistic but even undesirable. A recent study on using Facebook for learning purposes (Tsovaltzi et al. 2015b) showed that participants who had carefully prepared their own arguments on a given topic were less likely to interact productively with other SNS users on that topic, compared to people without that sort of individual preparation—yet another finding fully compatible with the argumentative theory of reasoning. That led the authors to conclude that “directly interacting with the support of argumentation scripts and without long individual preparation and reflection may be preferable to carefully preparing arguments before joining discussion in a SNS” (Tsovaltzi et al. 2015b, p. 588).
Crucially, complex argument annotation systems are problematic also for experts, not only for laypeople. When using a simplified version of Walton’s argumentation scheme taxonomy (including 14 schemes out of 60), with coders trained on argumentation and four iterations of the annotation procedure, the final level of inter-coder agreement can still be as low as 0.48, measured by Cohen’s kappa (Schneider et al. 2013). This figure raises to more acceptable levels only when simpler annotation schemes are adopted (Schneider et al. 2012b), and it is further improved by isolating ex post the most reliable sub-groups of annotators (Peldszus and Stede 2013b).
References
Amgoud, L., & Prade, H. (2009). Using arguments for making and explaining decisions. Artificial Intelligence, 173(3–4), 413–436.
Antoci, A., Sabatini, F., & Sodini, M. (2015). Online and offline social participation and social poverty traps. Journal of Mathematical Sociology. forthcoming.
Baroni, P., & Giacomin, M. (2007). On principle-based evaluation of extension-based argumentation semantics. Artificial Intelligence, 171(10), 675–700.
Bench-Capon, T., & Dunne, P. (2007). Argumentation in artificial intelligence. Artificial Intelligence, 171(10), 619–641.
Besnar, P., & Hunter, A. (2001). A logic-based theory of deductive arguments. Artificial Intelligence, 128(1–2), 203–235.
Bex, F., Lawrence, J., Snaith, M., & Reed, C. (2013). Implementing the argument web. Communications of the ACM, 56(10), 66–73.
Bex, F., Snaith, M., Lawrence, J., & Reed, C. (2014). ArguBlogging: an application for the argument web. Web Semantics: Science, Services and Agents on the World Wide Web, 25, 9–15.
Boudry, M., Pigliucci, M., & Paglieri, F. (2015). The fake, the flimsy, and the fallacious: demarcating arguments in real life. Argumentation, 29(4), 431–456.
Buckingham Shum, S. (2008). Cohere: towards web 2.0 argumentation. In P. Besnard, S. Doutre, & A. Hunter (Eds.), Computational models of argument: proceedings of COMMA 2008 (pp. 97–108). Amsterdam: IOS Press.
Buckingham Shum, S., Selvin, A., Sierhuis, M., Conklin, J., Haley, C., & Nuseibeh, B. (2006). Hypermedia support for argumentation-based rationale. In A. Dutoit, R. McCall, I. Mistrík, & B. Paech (Eds.), Rationale management in software engineering (pp. 111–132). Berlin: Springer.
Butterworth, J., & Thwaites, G. (2013). Thinking skills: critical thinking and problem solving (2nd ed.). Cambridge: Cambridge University Press.
Cabrio, E., & Villata, S. (2013). A natural language bipolar argumentation approach to support users in online debate interactions. Argument & Computation, 4(3), 209–230.
Caminada, M., & Amgoud, L. (2007). On the evaluation of argumentation formalisms. Artificial Intelligence, 171(5–6), 286–310.
Carbogim, D., Robertson, D., & Lee, J. (2000). Argument-based applications to knowledge engineering. The Knowledge Engineering Review, 15(2), 119–149.
Chang, C. F., Miller, A., & Ghose, A. (2010). Mixed-initiative argumentation: group decision support in medicine. In P. Kostkova (Ed.), Electronic healthcare: proceedings of eHealth 2009 (pp. 43–50). Berlin: Springer.
Chesnevar, C., McGinnis, J., Modgil, S., Rahwan, I., Reed, C., Simari, G., South, M., Vreeswijk, G., & Willmott, S. (2006). Towards an argument interchange format. Knowledge Engineering Review, 21(4), 293–316.
Collins, P., Hahn, U., von Gerber, Y., & Olsson, E. (2015). The bi-directional relationship between source characteristics and message content. In D. Noelle, R. Dale, A. Warlaumont, J. Yoshimi, T. Matlock, C. Jennings & P. Maglio (Eds.), Proceedings of the 37th Annual Meeting of the Cognitive Science Society (pp. 423–428). Austin, TX: Cognitive Science Society.
Conklin, J., Selvin, A., Buckingham Shum, S., & Sierhuis, M. (2001). Facilitated hypertext for collective sensemaking: 15 years on from gIBIS. In K. Grønbæk, H. Davis &Y. Douglas (Eds.), Hypertext’01: Proceedings of the 12th ACM Conference on Hypertext and Hypermedia (pp. 123–124). New York: ACM.
Corner, A., & Hahn, U. (2012). Normative theories of argumentation: are some norms better than others? Synthese, 190(16), 3579–3610.
de Moor, A., & Aakhus, M. (2006). Argumentation support: from technologies to tools. Communications of the ACM, 49(3), 93–98.
Duggan, M. (2014). Online harassment. Washington: Pew Research Internet Project.
Dung, P. (1995). On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming, and n-person games. Artificial Intelligence, 77, 321–357.
Dung, P., Kowalski, R., & Toni, F. (2006). Dialectic proof procedures for assumption-based, admissible argumentation. Artificial Intelligence, 170(2), 114–159.
Ellison, N., Steinfield, C., & Lampe, C. (2007). The benefits of facebook friends: social capital and college students’ use of online social network sites. Journal of Computer-Mediated Communication, 12, 114–1168.
Ennis, R. (1989). Critical thinking and subject specificity: clarification and needed research. Educational Researcher, 18(3), 4–10.
Ennis, R. (1993). Critical thinking assessment. Theory Into Practice, 32(3), 179–186.
Facione, P. (Ed.) (1990). Critical thinking: a statement of expert consensus for purposes of educational assessment and instruction. American Philosophical Association: ERIC document ED 315–423.
Finocchiaro, M. (1981). Fallacies and the evaluation of reasoning. American Philosophical Quarterly, 18(1), 13–22.
Fiske, S., & Taylor, S. (1984). Social cognition. Reading: Addison-Wesley.
Floridi, L. (2009). Logical fallacies as informational shortcuts. Synthese, 167, 317–325.
Fogg, B. J. (2003). Persuasive technology: using computers to change what we think and do. San Francisco: Morgan Kaufmann.
Frederick, S. (2005). Cognitive reflection and decision making. Journal of Economic Perspectives, 19(4), 25–42.
Gabbriellini, S., & Torroni, P. (2012). Large-scale agreements via microdebates. In S. Ossowski, G. Vouros & F. Toni (Eds.), AT 2012: Proceedings of the 1st International Conference on Agreement Technologies (pp. 366–377). Tilburg: CEUR-WS.org.
Garcia, A., & Simari, G. (2014). Defeasible logic programming: DeLP-servers, contextual queries, and explanations for answers. Argument & Computation, 5(1), 63–88.
Garrison, D. R., Anderson, T., & Archer, W. (2001). Critical thinking, cognitive presence, and computer conferencing in distance education. American Journal of Distance Education, 15(1), 7–23.
Gigerenzer, G., & Selten, R. (Eds.). (2001). Bounded rationality: the adaptive toolbox. Cambridge: The MIT Press.
Gigerenzer, G., Hertwig, R., & Pachur, T. (Eds.). (2011). Heuristics: the foundations of adaptive behavior. New York: Oxford University Press.
Goodwin, J. (1998). Forms of authority and the real ad verecundiam. Argumentation, 12(2), 267–280.
Gordon, T. (2010). An overview of the Carneades argumentation support system. In C. Tindale & C. Reed (Eds.), Dialectics, dialogue and argumentation. An examination of Douglas Walton’s theories of reasoning (pp. 145–156). London: College Publications.
Govier, T. (1987). Problems in argument analysis and evaluation. Dordrecht: Foris.
Groarke, L. (2009). What’s wrong with the California critical thinking skills test? CT testing and accountability. In J. Sobocan & L. Groarke (Eds.), Critical thinking education and assessment: can higher order thinking be tested? (pp. 35–54). London: The Althouse Press.
Habermas, J. (1984). The theory of communicative action. Boston: Beacon.
Hahn, U., & Oaksford, M. (2007). The rationality of informal argumentation: a Bayesian approach to reasoning fallacies. Psychological Review, 114, 704–732.
Hahn, U., Harris, A. J. L., & Corner, A. (2009). Argument content and argument source: an exploration. Informal Logic, 29(4), 337–367.
Hahn, U., Oaksford, M., & Harris, A. J. L. (2012). Testimony and argument: a Bayesian perspective. In F. Zenker (Ed.), Bayesian argumentation (pp. 15–38). Dordrecht: Springer.
Hahn, U., Oaksford, M., & Harris, A. J. L. (2013). Rational inference, rational argument. Argument & Computation, 4, 21–35.
Hamblin, C. (1970). Fallacies. London: Methuen.
Harris, A. J. L., Hsu, A. S., & Madsen, J. K. (2012). Because Hitler did it! Quantitative tests of Bayesian argumentation using ad hominem. Thinking and Reasoning, 18(3), 311–343.
Hintikka, J. (1987). The fallacy of fallacies. Argumentation, 1(3), 211–238.
Hitchcock, D. (1995). Do the fallacies have a place in the teaching of reasoning skills or critical thinking? In H. V. Hansen & R. C. Pinto (Eds.), Fallacies: classical and contemporary readings (pp. 319–327). University Park: Penn State University Press.
Hitchcock, D. (2006). Informal logic and the concept of argument. In D. Jacquette (Ed.), Philosophy of logic (Handbook of the philosophy of science, Vol. Volume 5, pp. 101–129). Amsterdam: Elsevier.
Hitchcock, D. (2007). Why there is no argumentum ad hominem fallacy. In F. H. van Eemeren & B. Garssen (Eds.), Proceedings of the Sixth Conference of the International Society for the Study of Argumentation (Volume 1, pp. 615–620). Amsterdam: Sic Sat.
Introne, J., & Iandoli, L. (2014). Improving decision-making performance through argumentation: an argument-based decision support system to compute with evidence. Decision Support Systems, 64, 79–89.
Janier, M., Lawrence, J., & Reed, C. (2014). OVA+: an argument analysis interface. In S. Parsons, N. Oren, C. Reed & F. Cerutti (Eds.), Computational Models of Argument: Proceedings of COMMA 2014 (pp. 463–464). Amsterdam: IOS Press
Johnson, R., & Blair, A. (1977). Logical self-defense. Toronto: McGraw-Hill Ryerson.
Kahneman, D. (2011). Thinking, fast and slow. New York: Farrar, Straus and Giroux.
Karacapilidis, N., & Papadias, D. (2001). Computer supported argumentation and collaborative decision making: the HERMES system. Information Systems, 26(4), 259–277.
Karunatillake, N., Jennings, N., Rahwan, I., & McBurney, P. (2009). Dialogue games that agents play within a society. Artificial Intelligence, 173(9–10), 935–981.
Kirschner, P. (2015). Facebook as learning platform: argumentation superhighway or dead-end street? Computers in Human Behavior, 53, 621–625.
Kirschner, P., Buckingham Shum, S., & Carr, C. (Eds.). (2003). Visualizing argumentation. Software tools for collaborative and educational sense-making. Berlin: Springer.
Klein, M. (2012). Enabling large-scale deliberation using attention-mediation metrics. Journal of Computer-Supported Cooperative Work, 21(4), 449–473.
Klein, M., & Convertino, G. (2014). An embarrassment of riches. Communications of the ACM, 57(11), 40–42.
Kuhn, D. (1991). The skills of arguments. Cambridge: Cambridge University Press.
Kunda, Z. (1990). The case for motivated reasoning. Psychological Bulletin, 108, 480–498.
Laughlin, P., & Ellis, A. (1986). Demonstrability and social combination processes on mathematical intellective tasks. Journal of Experimental Social Psychology, 22, 177–189.
Lave, J., & Wenger, E. (1991). Situated learning: legitimate peripheral participation. Cambridge: Cambridge University Press.
Lawrence, J., Bex, F., & Reed, C. (2012). Dialogues on the argument web: mixed initiative argumentation with Arvina. In B. Verheij, S. Szeider & S. Woltran (Eds.), Computational Models of Argument: Proceedings of COMMA 2012 (pp. 513–514). Amsterdam: IOS Press.
Lawrence, J., Reed, C., Allen, C., McAlister, S., Ravenscroft, A., & Bourget, D. (2014). Mining arguments from 19th century philosophical texts using topic based modelling. In N. Green, K. Ashley, D. Litman, C. Reed & V. Walke (Eds.), Proceedings of the First Workshop on Argumentation Mining (pp. 79–87). Stroudsburg, PA: ACL
Levesque, H. J. (1986). Making believers out of computers. Artificial Intelligence, 30(1), 81–108.
Levi, D. S. (1999). The fallacy of treating the ad baculum as a fallacy. Informal Logic, 19(2–3), 145–159.
Lindsay, B. (2009). Creating “the Wikipedia of pros and cons”. In D. Riehle & A. Bruckman (Eds.), WikiSym’09: Proceedings of the 5th International Symposium on Wikis and Open Collaboration (n. 36). New York: ACM.
Mackenzie, P. T. (1980). Ad hominem and ad verecundiam. Informal Logic, 3(3), 9–11.
Massey, G. (1981). The fallacy behind fallacies. Midwest Studies In Philosophy, 6(1), 489–500.
McPeck, J. (1990). Critical thinking and subject specificity: a reply to Ennis. Educational Researcher, 19(4), 10–12.
Mercier, H. (2010). The social origins of folk epistemology. Review of Philosophy and Psychology, 1(4), 499–514.
Mercier, H. (2013). Our pigheaded core: how we became smarter to be influenced by other people. In K. Sterelny, R. Joyce, B. Calcott, & B. Fraser (Eds.), Cooperation and its evolution (pp. 373–398). Cambridge: MIT Press.
Mercier, H., & Sperber, D. (2009). Intuitive and reflective inferences. In J. S. B. T. Evans & K. Frankish (Eds.), In two minds: dual processes and beyond (pp. 149–170). New York: Oxford University Press.
Mercier, H., & Sperber, D. (2011). Why do humans reason? Arguments for an argumentative theory. Behavioral and Brain Sciences, 34(2), 57–74.
Mercier, H., Trouche, E., Yama, H., Heintz, C., & Girotto, V. (2015). Experts and laymen grossly underestimate the benefits of argumentation for reasoning. Thinking and Reasoning, 21(3), 341–355.
Mizrahi, M. (2010). Take my advice—I am not following it: ad hominem arguments as legitimate rebuttals to appeals to authority. Informal Logic, 30(4), 435–456.
Mochales Palau, R., & Moens, M.-F. (2009). Argumentation mining: the detection, classification and structure of arguments in text. In P. Casanovas & C. Hafner (Eds.), Proceedings of the 12th International Conference on Artificial intelligence and Law (pp. 98–107). New York: ACM.
Mochales Palau, R., & Moens, M.-F. (2011). Argumentation mining. Artificial Intelligence and Law, 19(1), 1–22.
Modgil, S. (2009). Reasoning about preferences in argumentation frameworks. Artificial Intelligence, 173(9–10), 901–934.
Modgil, S., & Caminada, M. (2009). Proof theories and algorithms for abstract argumentation frameworks. In I. Rahwan & G. Simari (Eds.), Argumentation in artificial intelligence (pp. 105–129). Berlin: Springer.
Modgil, S., & Prakken, H. (2014). The ASPIC+ framework for structured argumentation: a tutorial. Argument and Computation, 5(1), 31–62.
Modgil, S., Toni, F., Bex, F., Bratko, I., Chesñevar, C., Dvorák, W., & Woltran, S. (2013). The added value of argumentation. In S. Ossowski (Ed.), Agreement technologies (pp. 357–403). Berlin: Springer.
Moens, M.-F., Boiy, E., Mochales Palau, R., & Reed, C. (2007). Automatic detection of arguments in legal texts. In A. Gardner & R. Winkels (Eds.), Proceedings of the 11th International Conference on Artificial intelligence and Law (pp. 225–230). New York: ACM.
Morge, M. (2008). The hedgehog and the fox. An argumentation-based decision support system. In I. Rahwan, S. Parsons & C. Reed (Eds.), Argumentation in Multi-Agent Systems: Proceedings of ArgMAS 2007 (pp. 114–131). Berlin: Springer.
Moshman, D., & Geil, M. (1998). Collaborative reasoning: evidence for collective rationality. Thinking and Reasoning, 4(3), 231–248.
Nickerson, R. (1998). Confirmation bias: a ubiquitous phenomena in many guises. Review of General Psychology, 2, 175–220.
O’Keefe, D. (1977). Two concepts of argument. Journal of the American Forensic Society, 13, 121–128.
Ossowski, S. (Ed.). (2012). Agreement technologies. Berlin: Springer.
Paglieri, F. (2016). Don’t worry, be gappy! On the unproblematic gappiness of fallacies. In F. Paglieri, L. Bonelli, & S. Felletti (Eds.), The psychology of argument: cognitive approaches to argumentation and persuasion (pp. 153–172). London: College Publications.
Paglieri, F., & Castelfranchi, C. (2010). Why argue? Towards a cost–benefit analysis of argumentation. Argument and Computation, 1(1), 71–91.
Peldszus, A., & Stede, M. (2013a). From argument diagrams to argumentation mining in texts: a survey. International Journal of Cognitive Informatics and Natural Intelligence, 7(1), 1–31.
Peldszus, A., & Stede, M. (2013b). Ranking the annotators: An agreement study on argumentation structure. In S. Dipper, M. Liakata & A. Pareja-Lora (Eds.), Proceedings of the 7th Linguistic Annotation Workshop & Interoperability with Discourse (pp. 196–204). Stroudsburg, PA: ACL.
Perkins, D., Farady, M., & Bushey, B. (1991). Everyday reasoning and the roots of intelligence. In J. Voss, D. Perkins, & J. Segal (Eds.), Informal reasoning and education (pp. 83–105). Hillsdale: Lawrence Erlbaum Associates.
Possin, K. (2008). A field guide to critical-thinking assessment. Teaching Philosophy, 31(3), 201–228.
Prakken, H. (2010). An abstract framework for argumentation with structured arguments. Argument and Computation, 1(2), 93–124.
Rahwan, I. (2008). Mass argumentation and the semantic web. Web Semantics: Science, Services and Agents on the World Wide Web, 6(1), 29–37.
Rahwan, I., & McBurney, P. (2007). Argumentation technology. IEEE Intelligent Systems, 22(6), 21–23.
Rahwan, I., & Simari, G. (Eds.). (2009). Argumentation in artificial intelligence. Berlin: Springer.
Rahwan, I., Ramchurn, S., Jennings, N., McBurney, P., Parsons, S., & Sonenberg, L. (2004). Argumentation-based negotiation. The Knowledge Engineering Review, 18(4), 343–375.
Rahwan, I., Zablith, F., & Reed, C. (2007). Laying the foundations for a world wide argument web. Artificial Intelligence, 171(10–15), 897–921.
Rahwan, I., Banihashemi, B., Reed, C., Walton, D., & Abdallah, S. (2011). Representing and classifying arguments on the semantic web. The Knowledge Engineering Review, 26(4), 487–511.
Rahwan, I., Krasnoshtan, D., Shariff, A., & Bonnefon, J.-F. (2014). Analytical reasoning task reveals limits of social learning in networks. Journal of the Royal Society, Interface, 11(93), 20131211.
Rainie, L., Lenhart, A., & Smith, A. (2012). The tone of life on social networking sites. Washington: Pew Internet Research Center.
Reed, C., & Norman, T. (Eds.). (2004). Argumentation machines. Berlin: Springer.
Reed, C., & Rowe, G. (2004). Araucaria: software for argument analysis, diagramming and representation. International Journal on Artificial Intelligence Tools, 13(4), 961–980.
Reed, C., & Walton, D. (2003). Argumentation schemes in argument-as-process and argument-as-product. in J. A. Blair, D. Farr, H. Hansen, R. Johnson and C. Tindale (Eds.), Informal Logic @ 25: Proceedings of the 5th OSSA Conference. Windsor, Ontario: OSSA.
Reed, C., Wells, S., Snaith, M., Budzynska, K., & Lawrence, J. (2011). Using an argument ontology to develop pedagogical tool suites. In P. Blackburn, H. van Ditmarsch, M. Manzano, & F. Soler-Toscano (Eds.), Tools for teaching logic: proceedings of TICTTL 2011 (pp. 207–214). Berlin: Springer.
Rowe, G., Macagno, F., Reed, C., & Walton, D. (2006). Araucaria as a tool for diagramming arguments in teaching and studying philosophy. Teaching Philosophy, 29(2), 111–124.
Sà, W., West, R., & Stanovich, K. (1999). The domain specificity and generality of belief bias: searching for a generalizable critical thinking skill. Journal of Educational Psychology, 91(3), 497–510.
Sabatini, F., & Sarracino, F. (2014). Online networks and subjective well-being. ArXiv, 1408, 3550.
Scheuer, O., Loll, F., Pinkwart, N., & McLaren, B. (2010). Computer-supported argumentation: a review of the state of the art. International Journal of Computer-Supported Collaborative Learning, 5(1), 43–102.
Schneider, J. (2014). An informatics perspective on argumentation mining. In E. Cabrio, S. Villata, & A. Wyner (Eds.), Proceedings of the workshop on frontiers and connections between argumentation theory and natural language processing (pp. 1–4). Aachen: CEUR-WS.org.
Schneider, J., Groza, T., & Passant, A. (2012a). A review of argumentation for the Social Semantic Web. Semantic Web-Interoperability, Usability, Applicability, 4(2), 159–218.
Schneider, J., Passant, A., & Decker, S. (2012b). Deletion discussions in Wikipedia: decision factors and outcomes. In C. Lampe (Ed.), WikiSym2012: Proceedings of the 8th Annual International Symposium on Wikis and Open Collaboration (n. 17). New York: ACM.
Schneider, J., Samp, K., Passant, A., & Decker, S. (2013). Arguments about deletion: how experience improves the acceptability of arguments in ad-hoc online task groups. In A. Bruckman, S. Counts, C. Lampe & L. Terveen (Eds.), CSCW2013: Proceedings of the 2013 Conference on Computer Supported Cooperative Work (pp. 1069–1080). New York: ACM.
Scriven, M. (1987). Fallacies of statistical substitution. Argumentation, 1, 333–349.
Simon, H. A. (1956). Rational choice and the structure of the environment. Psychological Review, 63, 129–138.
Sperber, D., Clément, F., Heintz, C., Mascaro, O., Mercier, H., Origgi, G., & Wilson, D. (2010). Epistemic vigilance. Mind & Language, 25(4), 359–393.
Stanovich, K., & West, R. (2007). Natural myside bias is independent of cognitive ability. Thinking and Reasoning, 13(3), 225–247.
Steinfield, C., Ellison, N., & Lampe, C. (2008). Social capital, self-esteem, and use of online social network sites: a longitudinal analysis. Journal of Applied Developmental Psychology, 29, 434–445.
Stone, M. (2012). Denying the antecedent: its effective use in argumentation. Informal Logic, 32(3), 327–356.
Surowiecki, J. (2004). The wisdom of crowds. New York: Doubleday.
Toplak, M., West, R., & Stanovich, K. (2011). The cognitive reflection test as a predictor of performance on heuristics-and-biases tasks. Memory & Cognition, 39, 1275–1289.
Toulmin, S. (1958). The uses of argument. Cambridge: Cambridge University Press.
Towne, W. B., & Herbsleb, J. (2012). Design considerations for online deliberation systems. Journal of Information Technology & Politics, 9(1), 97–115.
Trouche, E., Sander, E., & Mercier, H. (2014). Arguments, more than confidence, explain the good performance of reasoning groups. Journal of Experimental Psychology. General, 143(5), 1958–1971.
Trouche, E., Johansson, P., Hall, L., & Mercier, H. (in press). The selective laziness of reasoning. Cognitive Science, doi: 10.1111/cogs.12303
Tsovaltzi, D., Greenhow, C., & Asterhan, C. (2015a). When friends argue: learning from and through social network site discussions. Computers in Human Behavior, 53, 567–569.
Tsovaltzi, D., Judele, R., Puhl, T., & Weinberger, A. (2015b). Scripts, individual preparation and group awareness support in the service of learning in Facebook: how does CSCL compare to social networking sites? Computers in Human Behavior, 53, 577–592.
Tversky, A., & Kahneman, D. (1971). Belief in the law of small numbers. Psychological Bulletin, 76(2), 105–110.
Tversky, A., & Kahneman, D. (1982). Evidential impact of base rates. In D. Kahneman, P. Slovic, & A. Tversky (Eds.), Judgment under uncertainty: heuristics and biases (pp. 153–160). Cambridge: Cambridge University Press.
Tversky, A., & Kahneman, D. (1983). Extensional versus intuitive reasoning: the conjunction fallacy in probability judgment. Psychological Review, 90(4), 293–315.
van Eemeren, F., & Grootendorst, R. (1992). Relevance reviewed: the case of argumentum ad hominem. Argumentation, 6(2), 14–159.
van Eemeren, F., & Grootendorst, R. (1995). The pragma-dialectical approach to fallacies. In H. V. Hansen & R. C. Pinto (Eds.), Fallacies: classical and contemporary readings (pp. 130–144). University Park: Penn State University Press.
Walton, D. (1992). Nonfallacious arguments from ignorance. American Philosophical Quarterly, 29(4), 381–387.
Walton, D. (1996). Argumentation schemes for presumptive reasoning. Mahwah: Lawrence Erlbaum Associates.
Walton, D. (1997). Appeal to expert opinion: arguments from authority. University Park: The Pennsylvania State University Press.
Walton, D. (1998). Ad hominem arguments. Tuscaloosa: The University of Alabama Press.
Walton, D. (1999). The appeal to ignorance, or argumentum ad ignorantiam. Argumentation, 13(4), 367–377.
Walton, D. (2000). Scare tactics: arguments that appeal to fear and threats. Dordrecht: Kluwer.
Walton, D., & Godden, D. M. (2007). Informal logic and the dialectical approach to argument. In H. Hansen & R. Pinto (Eds.), Reason reclaimed (pp. 3–17). Newport News: Vale Press.
Walton, D., & Gordon, T. (2012). The Carneades model of argument invention. Pragmatics & Cognition, 20(1), 1–31.
Walton, D., Reed, C., & Macagno, F. (2008). Argumentation schemes. Cambridge: Cambridge University Press.
Wason, P. C. (1966). Reasoning. In B. Foss (Ed.), New horizons in psychology: I (pp. 106–137). Harmandsworth: Penguin.
Weinberger, A., Stegmann, K., Fischer, F., & Mandl, H. (2007). Scripting argumentative knowledge construction in computer-supported learning environments. In F. Fischer, I. Kollar, H. Mandl, & J. M. Haake (Eds.), Scripting computer-supported collaborative learning (pp. 191–211). Berlin: Springer.
Woods, J. (1998). Argumentum ad baculum. Argumentation, 12(4), 493–504.
Woods, J. (2013). Errors of reasoning. Naturalizing the logic of inference. London: College Publications.
Woods, J., & Walton, D. (1974). Argumentum ad verecundiam. Philosophy and Rhetoric, 7(3), 135–153.
Woods, J., & Walton, D. (1978). The fallacy of ‘ad ignorantiam’. Dialectica, 32(2), 87–99.
Yaglikci, N., & Torroni, P. (2014). Microdebates app for Android: a tool for participating in argumentative online debates using a handheld device. In A. Andreou & G. A. Papadopoulos (Eds.), Proceedings of ICTAI 2014: IEEE 26th International Conference on Tools with Artificial Intelligence (pp. 792–799). Los Alamitos, CA: IEEE.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Paglieri, F. A Plea for Ecological Argument Technologies. Philos. Technol. 30, 209–238 (2017). https://doi.org/10.1007/s13347-016-0222-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13347-016-0222-6