Menü Überspringen
Kontakt
Presse
Deutsch
English
Not track
Datenverarbeitung
Suche
Anmelden
DIPF aktuell
Forschung
Infrastrukturen
Institut
Zurück
Kontakt
Presse
Deutsch
English
Not track
Datenverarbeitung
Suche
Startseite
>
Forschung
>
Publikationen
>
Publikationendatenbank
Ergebnis der Suche in der DIPF Publikationendatenbank
Ihre Abfrage:
(Schlagwörter: "Computer")
zur erweiterten Suche
Suchbegriff
Nur Open Access
Suchen
Markierungen aufheben
Alle Treffer markieren
Export
273
Inhalte gefunden
Alle Details anzeigen
Automatically assigning research methods to journal articles in the domain of social sciences
Eckle-Kohler, Judith; Nghiem, Tri-Duc; Gurevych, Iryna
Sammelbandbeitrag
| Aus: Grove, Andrew (Hrsg.): Proceedings of the 76th Annual Meeting of the Association for Information Science and Technology | Silver Spring; MD; USA: Association for Information Science and Technology | 2013
34042 Endnote
Autor*innen:
Eckle-Kohler, Judith; Nghiem, Tri-Duc; Gurevych, Iryna
Titel:
Automatically assigning research methods to journal articles in the domain of social sciences
Aus:
Grove, Andrew (Hrsg.): Proceedings of the 76th Annual Meeting of the Association for Information Science and Technology, Silver Spring; MD; USA: Association for Information Science and Technology, 2013 , S. 1-8
URL:
http://www.asis.org/asist2013/proceedings/submissions/papers/45paper.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Automatisierung; Computerlinguistik; Computerunterstütztes Verfahren; Indexierung; Metadaten; Methode; Sozialwissenschaften
Abstract:
We investigate the automatic assignment of research methods to journal articles from the domain of Social Sciences. We employ Computer Science and Computational Linguistics methodology to perform this automatic assignment of metadata. The multi-label classification system we present uses only abstracts and titles of journal articles as input. Our best system is able to assign the important research methods empirical and quantitative empirical with F-scores of 0.67 and 0.68. These research methods are in the focus of many recent manual analyses of publications databases. Our classification approach could be applied to automatically analyze large publications databases and databases of bibliographic references according to the use of empirical and quantitative empirical methods.
DIPF-Abteilung:
Informationszentrum Bildung
The impact of topic bias on quality flaw prediction in Wikipedia
Ferschke, Oliver; Gurevych, Iryna; Rittberger, Marc
Sammelbandbeitrag
| Aus: Association of Computational Linguistics (Hrsg.): 51st Annual Meeting of the Association for Computational Linguistics: Proceedings of the Conference System Demonstrations | Stroudsburg; PA: Association for Computational Linguistics | 2013
33527 Endnote
Autor*innen:
Ferschke, Oliver; Gurevych, Iryna; Rittberger, Marc
Titel:
The impact of topic bias on quality flaw prediction in Wikipedia
Aus:
Association of Computational Linguistics (Hrsg.): 51st Annual Meeting of the Association for Computational Linguistics: Proceedings of the Conference System Demonstrations, Stroudsburg; PA: Association for Computational Linguistics, 2013 , S. 721-730
URN:
urn:nbn:de:0111-dipfdocs-184570
URL:
http://www.dipfdocs.de/volltexte/2020/18457/pdf/The_impact_of_topic_bias_on_quality_flaw_prediction_in_Wikipedia_A.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Algorithmus; Computerunterstütztes Verfahren; Evaluation; Nachschlagewerk; Online; Qualität; Qualitätssicherung; Reliabilität; Soziale Software; Standard; World wide web 2.0
Abstract:
With the increasing amount of user generated reference texts in the web, automatic quality assessment has become a key challenge. However, only a small amount of annotated data is available for training quality assessment systems. Wikipedia contains a large amount of texts annotated with cleanup templates which identify quality flaws. We show that the distribution of these labels is topically biased, since they cannot be applied freely to any arbitrary article. We argue that it is necessary to consider the topical restrictions of each label in order to avoid a sampling bias that results in a skewed classifier and overly optimistic evaluation results. We factor out the topic bias by extracting reliable training instances from the revision history which have a topic distribution similar to the labeled articles. This approach better reflects the situation a classifier would face in a real-life application.
DIPF-Abteilung:
Informationszentrum Bildung
Can we hide in the web? Large scale simultaneous age and gender author profiling in social media. […]
Flekova, Lucie; Gurevych, Iryna
Sammelbandbeitrag
| Aus: Forner, Pamela; Navigli, Roberto; Tufis, Dan (Hrsg.): CLEF 2013 Labs and Workshops: Online working notes | Padua: PROMISE | 2013
34036 Endnote
Autor*innen:
Flekova, Lucie; Gurevych, Iryna
Titel:
Can we hide in the web? Large scale simultaneous age and gender author profiling in social media. Notebook for PAN at CLEF 2013
Aus:
Forner, Pamela; Navigli, Roberto; Tufis, Dan (Hrsg.): CLEF 2013 Labs and Workshops: Online working notes, Padua: PROMISE, 2013 , S. 1-11
URL:
http://ims-sites.dei.unipd.it/documents/71612/430938/CLEF2013wn-PAN-FlekovaEt2013.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Alter; Chatten <Kommunikation>; Computerlinguistik; Computerunterstütztes Verfahren; Datenanalyse; Gender; Soziale Software; Sprachanalyse; Stilistik; Web log; Zielgruppe
Abstract:
Would you target your audience differently, knowing the real age and gender of the text authors on your website forum? This paper examines hundreds of thousands of online documents, e.g. chat lines or blog posts, showing that computers are capable to address this task better than humans, without relying on content stereotypes. Pointing out that age and gender profiling are not independent problems, we approach the task as a multiclass classification problem, combining the age and gender information to define six classes. Utilizing a wide range of stylistic and content features and a large number of readability measures we demonstrate the high predictive abilities of the parts of speech, the punctuation and the amount of emotions and slang used in the text, independently of the topic discussed.
DIPF-Abteilung:
Informationszentrum Bildung
UBY - A large-scale lexical-semantic resource [Abstract]
Gurevych, Iryna; Eckle-Kohler, Judith; Hartmann, Silvana; Matuschek, Michael; Meyer, Christian M.; […]
Sammelbandbeitrag
| Aus: Theune, M. ; Nijholt, A. (Hrsg.): Book of abstracts of the 23rd Meeting of Computational Linguistics in the Netherlands (CLIN 2013) | Enschede: Universiteit Twente | 2013
33317 Endnote
Autor*innen:
Gurevych, Iryna; Eckle-Kohler, Judith; Hartmann, Silvana; Matuschek, Michael; Meyer, Christian M.; Nghiem, Tri-Duc
Titel:
UBY - A large-scale lexical-semantic resource [Abstract]
Aus:
Theune, M. ; Nijholt, A. (Hrsg.): Book of abstracts of the 23rd Meeting of Computational Linguistics in the Netherlands (CLIN 2013), Enschede: Universiteit Twente, 2013 , S. 81
URL:
http://hmi.ewi.utwente.nl/clin2013-dir/bookofabstracts.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Automatisierung; Computerlinguistik; Computerunterstütztes Verfahren; Deutsch; Englisch; Lexikon
Abstract:
We present UBY, a large-scale lexical-semantic resource combining a wide range of information from expert-constructed and collaboratively created resources for English and German. It currently contains nine resources in two languages: English WordNet, Wiktionary, Wikipedia, FrameNet and VerbNet, German Wikipedia, Wiktionary, and GermaNet, and the multilingual OmegaWiki. The main contributions of our work can be summarised as follows. First, we define a standardised format for modelling the heterogeneous information coming from the various lexical-semantic resources (LSRs) and languages included in UBY. For this purpose, we employ the ISO standard Lexical Markup Framework and Data Categories selected from ISOCat. In this way, all types of information provided by the LSRs in UBY are easily accessible on a fine-grained level. Further, this standardised format facilitates the extension of UBY with new languages and resources. This is different from previous efforts in combining LSRs which usually targeted particular applications and thus focused on aligning specific types of information only. Second, UBY contains nine pairwise sense alignments between resources. Through these alignments, we provide access to the complementary information for a word sense in different resources. For example, if one looks up a particular verb sense in UBY, one has simultaneous access to the sense in WordNet and to the corresponding sense in FrameNet. Third, UBY is freely available and we have developed an easy-to-use Java API which provides unified access to all types of information contained in UBY. This facilitates the utilization of UBY for a variety of NLP tasks.
DIPF-Abteilung:
Informationszentrum Bildung
Finding similar movements in positional data streams
Haase, Jens; Brefeld, Ulf
Sammelbandbeitrag
| Aus: ECML/PKDD (Hrsg.): Proceedings of the ECML/PKDD Workshop on Machine Learning and Data Mining for Sports Analytics (ECML/PKDD 2013) | Prag: European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases | 2013
34024 Endnote
Autor*innen:
Haase, Jens; Brefeld, Ulf
Titel:
Finding similar movements in positional data streams
Aus:
ECML/PKDD (Hrsg.): Proceedings of the ECML/PKDD Workshop on Machine Learning and Data Mining for Sports Analytics (ECML/PKDD 2013), Prag: European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases, 2013 , S. 1-9
URL:
http://www.kma.informatik.tu-darmstadt.de/fileadmin/user_upload/Group_KMA/kma_publications/paper_01.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Algorithmus; Computerspiel; Computerunterstütztes Verfahren; Daten; Datenanalyse; Evaluation; Informatik
Abstract:
In this paper, we study the problem of efficiently finding similar movements in positional data streams, given a query trajectory. Our approach is based on a translation-, rotation-, and scale-invariant representation of movements. Near- neighbours given a query trajectory are then efficiently computed using dynamic time warping and locality sensitive hashing. Empirically, we show the efficiency and accuracy of our approach on positional data streams recorded from a real soccer game.
DIPF-Abteilung:
Informationszentrum Bildung
FrameNet on the way to Babel. Creating a bilingual FrameNet using Wiktionary as interlingual […]
Hartmann, Silvana; Gurevych, Iryna
Sammelbandbeitrag
| Aus: Association of Computational Linguistics (Hrsg.): 51st Annual Meeting of the Association for Computational Linguistics: Proceedings of the Conference System Demonstrations | Stroudsburg; PA: Association for Computational Linguistics | 2013
33526 Endnote
Autor*innen:
Hartmann, Silvana; Gurevych, Iryna
Titel:
FrameNet on the way to Babel. Creating a bilingual FrameNet using Wiktionary as interlingual connection
Aus:
Association of Computational Linguistics (Hrsg.): 51st Annual Meeting of the Association for Computational Linguistics: Proceedings of the Conference System Demonstrations, Stroudsburg; PA: Association for Computational Linguistics, 2013 , S. 1363-1373
URL:
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.386.1725&rep=rep1&type=pdf#page=49pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Computerlinguistik; Evaluation; Konzeption; Lexikograhie; Lexikon; Methode; Semantik; Soziale Software; World wide web 2.0; Zweisprachigkeit
Abstract:
We present a new bilingual FrameNet lexicon for English and German. It is created through a simple, but powerful approach to construct a FrameNet in any language using Wiktionary as an interlingual representation. Our approach is based on a sense alignment of FrameNet and Wiktionary, and subsequent translation disambiguation into the target language. We perform a detailed evaluation of the created resource and a discussion of Wiktionary as an interlingual connection for the cross-language transfer of lexical-semantic resources. The created resource is publicly available at http://www.ukp.tu-darmstadt.de/fnwkde/
DIPF-Abteilung:
Informationszentrum Bildung
Headerless, quoteless, but not hopeless? Using pairwise email classification to disentangle email […]
Jamison, Emily; Gurevych, Iryna
Sammelbandbeitrag
| Aus: Angelova, Galia; Bontcheva, Kalina; Mitkov, Ruslan (Hrsg.): Proceedings of 9th Conference on Recent Advances in Natural Language Processing (RANLP 2013) | Shoumen: INCOMA Ltd. | 2013
34043 Endnote
Autor*innen:
Jamison, Emily; Gurevych, Iryna
Titel:
Headerless, quoteless, but not hopeless? Using pairwise email classification to disentangle email threads
Aus:
Angelova, Galia; Bontcheva, Kalina; Mitkov, Ruslan (Hrsg.): Proceedings of 9th Conference on Recent Advances in Natural Language Processing (RANLP 2013), Shoumen: INCOMA Ltd., 2013 , S. 327-335
URL:
https://www.ukp.tu-darmstadt.de/fileadmin/user_upload/Group_UKP/publikationen/2013/RANLP_2013_EJIG_Camera.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Computerlinguistik; Datenanalyse; E-Mail; Inhalt; Klassifikation; Semantik; Struktur; Text
Abstract:
Thread disentanglement is the task of separating out conversations whose thread structure is implicit, distorted, or lost. In this paper, we perform email thread disentanglement through pairwise classification, using text similarity measures on non-quoted texts in emails. We show that i) content text similarity metrics outperform style and structure text similarity metrics in both a class-balanced and class-imbalanced setting, and ii) although feature performance is dependent on the semantic similarity of the corpus, content features are still effective even when controlling for semantic similarity. We make available the Enron Threads Corpus, a newly-extracted corpus of 70,178 multiemail threads with emails from the Enron Email Corpus.
DIPF-Abteilung:
Informationszentrum Bildung
UKP at CrossLink2. CJK-to-English Subtasks
Kim, Jungi; Gurevych, Iryna
Sammelbandbeitrag
| Aus: Kando, Noriko; Kishida, Kazuaki (Hrsg.): Proceedings of the 10th NTCIR Conference on Evaluation of Information Access Technologies | Tokio: NTCIR | 2013
34044 Endnote
Autor*innen:
Kim, Jungi; Gurevych, Iryna
Titel:
UKP at CrossLink2. CJK-to-English Subtasks
Aus:
Kando, Noriko; Kishida, Kazuaki (Hrsg.): Proceedings of the 10th NTCIR Conference on Evaluation of Information Access Technologies, Tokio: NTCIR, 2013 , S. 57-61
URL:
http://research.nii.ac.jp/ntcir/workshop/OnlineProceedings10/pdf/NTCIR/CrossLink-2/05-NTCIR10-CROSSLINK2-KimJ.pdf
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Automatisierung; Computerlinguistik; Computerunterstütztes Verfahren; Information Retrieval; Mehrsprachigkeit; Nachschlagewerk; Online; Sprachanalyse
Abstract:
This paper describes UKP's participation in the cross-lingual link discovery task at NTCIR-10 (CrossLink2). The task addressed in our work is to find valid anchor texts from a Chinese, Japanese, and Korean (CJK) Wikipedia page and retrieve the corresponding target Wiki pages in the English language. The CrossLink framework was developed based on our previous CrossLink system that works on the opposite directions of the language pairs, i.e. discovered anchor texts from English Wikipedia pages and their corresponding targets in CJK languages. The framework consists of anchor selection, anchor ranking, anchor translation, and target discovery sub-modules. Each sub-module in the framework has been shown to work well both in monolingual settings and English to CJK language pairs. We seek to find out whether the approach that worked very well for English to CJK would still work for CJK to English. We use the same experimental settings that were used in our previous participation, and our experimental runs show that the CJK-to- English CrossLink task is a much harder task when using the same resources as the English-to-CJK one.
DIPF-Abteilung:
Informationszentrum Bildung
Leseverstehen = Hörverstehen X Dekodieren? Ein stringenter Test der Simple View of Reading bei […]
Knoepke, Julia; Richter, Tobias; Isberner, Maj-Britt; Neeb, Yvonne; Naumann, Johannes
Sammelbandbeitrag
| Aus: Redder, Angelika; Weinert, Sabine (Hrsg.): Sprachförderung und Sprachdiagnostik: Interdisziplinäre Perspektiven | Münster: Waxmann | 2013
34060 Endnote
Autor*innen:
Knoepke, Julia; Richter, Tobias; Isberner, Maj-Britt; Neeb, Yvonne; Naumann, Johannes
Titel:
Leseverstehen = Hörverstehen X Dekodieren? Ein stringenter Test der Simple View of Reading bei deutschsprachigen Grundschulkindern
Aus:
Redder, Angelika; Weinert, Sabine (Hrsg.): Sprachförderung und Sprachdiagnostik: Interdisziplinäre Perspektiven, Münster: Waxmann, 2013 , S. 256-276
Dokumenttyp:
4. Beiträge in Sammelwerken; Sammelband (keine besondere Kategorie)
Sprache:
Deutsch
Schlagwörter:
Computerunterstütztes Verfahren; Deutschland; Empirische Untersuchung; Erkennen; Grundschüler; Hörverstehen; Kognitive Prozesse; Lesekompetenz; Lesetest; Leseverstehen; Schuljahr 03; Schuljahr 04; Visuelle Wahrnehmung; Wort
Abstract:
Die Autoren prüfen im Rahmen des Projekts "Prozessbezogene Diagnostik des Lese- und Hörverstehens im Grundschulalter" systematisch, ob sich Leseverstehen als Produkt aus der Effizienz der visuellen Worterkennung und im Hörverstehen von Sprache erklären lässt, wie es der "simple view of reading" annimmt. Die Befunde, die mittels gezielt entwickelter Tests zur Erfassung der entsprechenden Teilkompetenzen erzielt wurden, sind diagnostisch weiterführend und zeigen die Bedeutung beider Kompetenzen, ohne hingegen einen simple view of reading in einfacher Weise zu stützen.
DIPF-Abteilung:
Bildungsqualität und Evaluation
SemEval-2013 Task 5: Evaluating phrasal semantics
Korkontzelos, Ioannis; Zesch, Torsten; Zanzotto, Fabio Massimo; Biemann, Chris
Sammelbandbeitrag
| Aus: Association for Computational Linguistics (Hrsg.): Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013): Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2 | Atlanta; GA: Association for Computational Linguistics | 2013
33564 Endnote
Autor*innen:
Korkontzelos, Ioannis; Zesch, Torsten; Zanzotto, Fabio Massimo; Biemann, Chris
Titel:
SemEval-2013 Task 5: Evaluating phrasal semantics
Aus:
Association for Computational Linguistics (Hrsg.): Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval 2013): Second Joint Conference on Lexical and Computational Semantics (*SEM), Volume 2, Atlanta; GA: Association for Computational Linguistics, 2013 , S. 39-47
URL:
https://aclanthology.org/S13-2007/
Dokumenttyp:
4. Beiträge in Sammelwerken; Tagungsband/Konferenzbeitrag/Proceedings
Sprache:
Englisch
Schlagwörter:
Computerlinguistik; Computerprogramm; Computerunterstütztes Verfahren; Evaluation; Mehrsprachigkeit; Semantik; Syntax; Textanalyse
Abstract (english):
This paper describes the SemEval-2013 Task 5: "Evaluating Phrasal Semantics". Its first subtask is about computing the semantic similarity of words and compositionality of phrases in a given context. The paper discusses the importance and background of these subtasks and their structure. In succession, it introduces the systems that participated and discusses evaluation results.
DIPF-Abteilung:
Informationszentrum Bildung
Markierungen aufheben
Alle Treffer markieren
Export
<
1
...
19
20
(aktuell)
Alle anzeigen
(273)