Sentences with negative actors: negative strength quantified

Files: data1.xml,data3.xml,data3.xml (3 annotators) - XML validiert - 439 sentences - target: a negative cause (an actor etc.) represented by the Lemma - id: sentence number - string: the plain sentence - strength: negativity strength of the target - labels 0-3 - 0 no negative entity found (or parsi...

Full description

Bibliographic Details
Main Author: Manfred Klenner
Format: Other/Unknown Material
Language:German
Published: Zenodo 2023
Subjects:
Online Access:https://doi.org/10.5281/zenodo.7589628
Description
Summary:Files: data1.xml,data3.xml,data3.xml (3 annotators) - XML validiert - 439 sentences - target: a negative cause (an actor etc.) represented by the Lemma - id: sentence number - string: the plain sentence - strength: negativity strength of the target - labels 0-3 - 0 no negative entity found (or parsing error) - 1 slightly negative, 2 negative, 3 stronly negative - 115 out of 439 sentences with tag 0: i.e. sentences do not contain a negative actor - different reasons (see the paper below): modal, future tense etc. but also parsing errors Data source: Facebook posts of the AfD, a German right-wing party Examples: no actor here: passive voice <sent> 1 Junge 0 "Verletzt wurde auch ein 11-jähriger Junge . " </sent> stronly negative: <sent> 411 Euro 3 "Der Euro ruiniert Europa . " </sent> negative: <sent> 214 Merkel 2 "Merkel verantwortet zusätzliche 50 Milliarden Sozialkosten bis 2018 . " </sent> slightly negative: <sent> 154 Meuthen 1 "Meuthen schadet der Partei . " </sent> References: @inproceedings{nodalida, month = {Juni}, author = {Manfred Klenner and Anne G{\"o}hring and Sophia Conrad}, booktitle = {Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)}, address = {Reykjavik, Iceland}, title = {Getting Hold of Villains and other Rogues}, publisher = {Virtual Event}, pages = {435--439}, year = {2021}, language = {english}, url = {https://doi.org/10.5167/uzh-204265}, abstract = {In this paper, we introduce the first corpus specifying negative entities within sentences. We discuss indicators for their presence, namely particular verbs, but also the linguistic conditions when their prediction should be suppressed. We further show that a fine-tuned Bert-based baseline model outperforms an over-generating rule-based approach which is not aware of these further restrictions. If a perfect filter were applied, both would be on par.} }