References: [1] Afantenos, S., Peldszus, A., Stede, M. (2018). Comparing decoding mechanisms for parsing argumentative structures. Argument and Computation, 9, 177–192.
[2] Becker, M., Staniek, M., Nastase, V., Frank, A. (2017). Enriching Argumentative Texts with Implicit Knowledge. In F. Frasinca, A. Ittoo, L. M. Nguyen, & E. Metais (Eds.), Applications of Natural Language to Data Bases (NLDB) - Natural Language Processing and Information Systems, Lecture Notes in Computer Science. Springer. Retrieved from http://www.cl.uni-heidelberg.de/~mbecker/pdf/enriching-argumentative-texts.pdf.
[3] Botschen, T., Sorokin, D., Gurevych, I. (2018). Frame- and Entity-Based Knowledge for Common-Sense Argumentative Reasoning. In Proceedings of the 5th Workshop on Argument Mining (pp. 90–96). Retrieved from http://aclweb.org/anthology/W18-5211.
[4] Bowman, S. R., Angeli, G., Potts, C., Manning, C. D. (2015). A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP).
[5] Chen, Z., Song, W., Liu, L. (2018). TRANSRW at SemEval-2018 Task 12: Transforming Semantic Representations for Argument Reasoning Comprehension. In Proceedings of The 12th International Workshop on Semantic Evaluation (pp. 1142–1145). doi:10.18653/v1/S18-1194.
[6] Choi, H. S., Lee, H. (2018). GIST at SemEval-2018 Task 12: A network transferring inference knowledge to Argument Reasoning Comprehension task. In Proceedings of The 12th International Workshop on Semantic Evaluation (pp. 773–777). doi:10.18653/v1/S18-1122.
[7] Habernal, I., Wachsmuth, H., Gurevych, I., Stein, B. (2018). SemEval-2018 Task 12: The Argument Reasoning Comprehension Task. In Proceedings of The 12th International Workshop on Semantic Evaluation (pp. 763–772).
[8] Hochreiter, S., Schmidhuber, J. (1997). Long short-term memory. Neural computation, 9(8), 1735–1780.
[9] Hou, Y., Jochim, C. (2017). Argument Relation Classification Using a Joint Inference Model. In Proceedings of the 4th Workshop on Argument Mining (pp. 60–66).
[10] Hulpus, I., Prangnawarat, N., Hayes, C. (2015). Path-based semantic relatedness on linked data and its use to word and entity disambiguation. In International Semantic Web Conference (pp. 442–457). Springer.
[11] Kim, T., Choi, J., Lee, S. (2018). SNU_IDS at SemEval-2018 Task 12: Sentence Encoder with Contextualized Vectors for Argument Reasoning Comprehension. In Proceedings of The 12th International Workshop on Semantic Evaluation (pp. 1083–1088).
[12] Kingma, D. P., Ba, J. (2014). Adam: A Method for Stochastic Optimization. CoRR, abs/1412.6980. arXiv:1412.6980.
[13] Lin, Z., Kan, M.-Y., Ng, H. T. (2009). Recognizing Implicit Discourse Relations in the Penn Discourse Treebank. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing (pp. 343–351).
[14] Manning, C. D., Surdeanu, M., Bauer, J., Finkel, J., Bethard, S. J., McClosky, D. (2014). The Stanford CoreNLP Natural Language Processing Toolkit. In Association for Computational Linguistics (ACL) System Demonstrations (pp. 55–60).
[15] Menini, S., Tonelli, S. (2016). Agreement and Disagreement: Comparison of Points of View in the Political Domain. In COLING (pp. 2461–2470).
[16] Nguyen, H. N., Litman, D. J. (2016). Context-aware Argumentative Relation Mining. In ACL (pp. 1127–1137).
[17] Peldszus, A., Stede, M. (2013). From Argument Diagrams to Argumentation Mining in Texts: A Survey. Int. J. Cogn. Inform. Nat. Intell., 7(1), 1–31.
[18] Peldszus, A., Stede, M. (2015). Joint prediction in MST-style discourse parsing for argumentation mining. In EMNLP (pp. 938–948).
[19] Peldszus, A., Stede, M. (2016). An annotated corpus of argumentative microtexts. In Argumentation and Reasoned Action: Proceedings of the 1st European Conference on Argumentation, Lisbon 2015 / Vol. 2 (pp. 801–815). College Publications.
[20] Pennington, J., Socher, R., Manning, C. D. (2014). GloVe: Global Vectors for Word Representation. In Empirical Methods in Natural Language Processing (EMNLP) (pp. 1532–1543).
[21] Persing, I., Ng, V. (2016). End-to-End Argumentation Mining in Student Essays. In HLT-NAACL (pp. 1384–1394).
[22] Potash, P., Bhattacharya, R., Rumshisky, A. (2017). Length, Interchangeability, and
External Knowledge: Observations from Predicting Argument Convincingness. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers) (pp. 342–351).
[23] Singh, P. (2002). The Open Mind Common Sense Project. Retrieved from http://zoo.cs.yale.edu/classes/cs671/12f/12f-papers/singh-omcs-project.pdf.
[24] Speer, R., Havasi, C. (2012). Representing General Relational Knowledge in ConceptNet 5. In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC-2012) (pp. 3679–3686). European Language Resources Association (ELRA).
[25] Stab, C., Gurevych, I. (2014). Annotating Argument Components and Relations in Persuasive Essays. In COLING (pp. 1501–1510).
[26] Stab, C., Gurevych, I. (2014). Identifying Argumentative Discourse Structures in Persuasive Essays. In EMNLP (pp. 46–56).
[27] Stab, C., Gurevych, I. (2017). Parsing Argumentation Structures in Persuasive Essays. Computational Linguistics, 43, 619–659. |