Ergebnisse für *

Es wurden 2 Ergebnisse gefunden.

Zeige Ergebnisse 1 bis 2 von 2.

Sortieren

  1. Die Herausbildung neuer Routinen der Personenreferenz am Beispiel der deutschen Weihnachts- und Neujahrsansprachen
    Erschienen: 2025
    Verlag:  Hamburg : Buske ; Mannheim : Leibniz-Institut für Deutsche Sprache (IDS) [Zweitveröffentlichung]

    The Christmas and New Year addresses of the Federal Presidents and Federal Chancellors have been an annually recurring political text formal since 1949. This makes it wellsuited for shortterm diachronic corpus analyses. In our contribution, we focus... mehr

     

    The Christmas and New Year addresses of the Federal Presidents and Federal Chancellors have been an annually recurring political text formal since 1949. This makes it wellsuited for shortterm diachronic corpus analyses. In our contribution, we focus on the use of personal nouns with a special emphasis on gender-inclusive language. Using manual annotations, we show that 11 % of all tokens in the speeches refer to persons, not including proper names. Pronouns make up 69 % of person references, personal nouns 31 %. In the case ofpersonal nouns, we see that gender-neutral nouns or neutralisations predominate (58 %), followed by masculine generics (20 %). An important research question for this article is whether the use of gender-inclusive variants in language has increased over time. We can seefrom the data that this is the case: the use of masculine generics has been declining since 1995,while pair forms have been increasing from the beginning. This is especially true for explicit pair forms like Mitbürgerinnen und Mitbürger ('fellow citizens'). Gender-inclusive language has therefore long been a part of the addresses, especially in the form of neutralisations and pair forms.

     

    Export in Literaturverwaltung   RIS-Format
      BibTeX-Format
    Quelle: BASE Fachausschnitt Germanistik
    Sprache: Deutsch
    Medientyp: Aufsatz aus einem Sammelband
    Format: Online
    DDC Klassifikation: Germanische Sprachen; Deutsch (430)
    Schlagworte: Rede; Annotation; Geschlechterforschung; Geschlechterstereotyp; Korpus; Sprachwandel
    Lizenz:

    rightsstatements.org/page/InC/1.0/ ; info:eu-repo/semantics/openAccess

  2. Human languages trade off complexity against efficiency
    Erschienen: 2025
    Verlag:  San Francisco, CA : PLOS ; Mannheim : Leibniz-Institut für Deutsche Sprache (IDS)

    From a cross-linguistic perspective, language models are interesting because they can be used as idealised language learners that learn to produce and process language by being trained on a corpus of linguistic input. In this paper, we train... mehr

     

    From a cross-linguistic perspective, language models are interesting because they can be used as idealised language learners that learn to produce and process language by being trained on a corpus of linguistic input. In this paper, we train different language models, from simple statistical models to advanced neural networks, on a database of 41 multilingual text collections comprising a wide variety of text types, which together include nearly 3 billion words across more than 6,500 documents in over 2,000 languages. We use the trained models to estimate entropy rates, a complexity measure derived from information theory. To compare entropy rates across both models and languages, we develop a quantitative approach that combines machine learning with semiparametric spatial filtering methods to account for both language- and document-specific characteristics, as well as phylogenetic and geographical language relationships. We first establish that entropy rate distributions are highly consistent across different language models, suggesting that the choice of model may have minimal impact on cross-linguistic investigations. On the basis of a much broader range of language models than in previous studies, we confirm results showing systematic differences in entropy rates, i.e. text complexity, across languages. These results challenge the long-held notion that all languages are equally complex. We then show that higher entropy rate tends to co-occur with shorter text length, and argue that this inverse relationship between complexity and length implies a compensatory mechanism whereby increased complexity is offset by increased efficiency. Finally, we introduce a multi-model multilevel inference approach to show that this complexity-efficiency trade-off is partly influenced by the social environment in which languages are used: languages spoken by larger communities tend to have higher entropy rates while using fewer symbols to encode messages.

     

    Export in Literaturverwaltung   RIS-Format
      BibTeX-Format
    Quelle: BASE Fachausschnitt Germanistik
    Sprache: Englisch
    Medientyp: Aufsatz aus einer Zeitschrift
    Format: Online
    DDC Klassifikation: Sprache (400)
    Schlagworte: Großes Sprachmodell; Informationstheorie; Maschinelles Lernen; Kontrastive Linguistik; Computerlinguistik; Korpus; Statistik; Modell
    Lizenz:

    creativecommons.org/licenses/by/4.0/ ; info:eu-repo/semantics/openAccess