Sprache im 20. Jahrhundert. Gegenwartssprache
Refine
Year of publication
Document Type
- Article (31)
- Part of a Book (31)
- Conference Proceeding (27)
- Doctoral Thesis (1)
Language
- English (90) (remove)
Keywords
- Deutsch (51)
- Englisch (18)
- Korpus <Linguistik> (16)
- Konversationsanalyse (12)
- Gesprochene Sprache (11)
- Interaktion (9)
- Verb (8)
- Bedeutung (5)
- German (5)
- Massenmedien (5)
Publicationstate
- Veröffentlichungsversion (26)
- Postprint (19)
- Zweitveröffentlichung (5)
Reviewstate
Publisher
- Benjamins (8)
- Institut für Deutsche Sprache (7)
- Elsevier (5)
- European Language Resources Association (ELRA) (4)
- Lang (3)
- Verlag für Gesprächsforschung (3)
- de Gruyter (3)
- ISCA (2)
- Niemeyer (2)
- University of Birmingham (2)
This paper outlines the generation process of a specifi computational linguistic representation termed the Multilingual Time Map, conceptually a multi-tape finit state transducer encoding linguistic data at different levels of granularity. The fi st component acquires phonological data from syllable labeled speech data, the second component define feature profiles the third component generates feature hierarchies and augments the acquired data with the define feature profiles and the fourth component displays the Multilingual Time Map as a graph.
The current paper presents a corpus containing 35 dialogues of spontaneously spoken southern German, including half an hour of articulography for 13 of the speakers. Speakers were seated in separate recording chambers, mimicking a telephone call, and recorded on individual audio channels. The corpus provides manually corrected word boundaries and automatically aligned segment boundaries. Annotations are provided in the Praat format. In addition to audio recordings, speakers filled out a detailed questionnaire, assessing among others their audio-visual consumption habits.
This paper contributes to the growing body of knowledge on current listeners' responses in talk-in-interaction. In particular, it complements earlier findings on double sayings of German JA by describing some additional prosodic-phonetic parameters and a visual feature of its realization in institutional and semi-private interaction (doctor-patient interaction, Big Brother, TV talk shows). These include pitch contour, pitch range and phonetic ending, on the one hand, and nodding on the other. The paper shows that JAJA is a truly multimodal phenomenon, with the individual features accomplishing interactional functions across sequence-organizational habitats, including re)claiming epistemic priority in an aside, making continuation relevant, agreeing/ acknowledging with reservation and aligning with the continuation of a sequence. Lack of nodding is suggested to have situational as well as misalignment reasons. On the basis of its observations, the paper also raises the question whether it is the applicability of response token variants across action and sequence types which makes them memorizable despite their variability.
This paper presents the concept of the "participant perspective" as an approach to the study of spoken language. It discusses three aspects of this concept and shows that they can offer helpful tools in spoken language research. Employing the participant perspective provides us with an alternative to many of the approaches currently in use in the study of spoken language in that it favours small-scale, qualitative research that aims to uncover categories relevant for the participants. Its results can usefully complement large-scale studies of phenomena on all linguistic dimensions of talk.
The goal of the MULI (MUltiLingual Information structure) project is to empirically analyse information structure in German and English newspaper texts. In contrast to other projects in which information structure is annotated and investigated (e.g. in the Prague Dependency Treebank, which mirrors the basic information about the topic-focus articulation of the sentence), we do not annotate theory-biased categories like topic-focus or theme-rheme. Trying to be as theory-independent as possible, we annotate those features which are relevant to information structure and on the basis of which typical patterns, co-occurrences or correlations can be determined. We distinguish between three annotation levels: syntax, discourse and prosody. The data is based on the TIGER Corpus for German and the Penn Treebank for English, since the existing information on part-of-speech and syntactic structure can be re-used for our purposes. The actual annotation of an English example sequence illustrates our choice of categories on each level. Their combination offers the possibility to investigate how information structure is realised and can be interpreted.
We present the annotation of information structure in the MULI project. To learn more about the information structuring means in prosody, syntax and discourse, theory- independent features were defined for each level. We describe the features and illustrate them on an example sentence. To investigate the interplay of features, the representation has to allow for inspecting all three layers at the same time. This is realised by a stand-off XML mark-up with the word as the basic unit. The theory-neutral XML stand-off annotation allows integrating this resource with other linguistic resources such as the Tiger Treebank for German or the Penn treebank for English.
While written corpora can be exploited without any linguistic annotations, speech corpora need at least a basic transcription to be of any use for linguistic research. The basic annotation of speech data usually consists of time-aligned orthographic transcriptions. To answer phonetic or phonological research questions, phonetic transcriptions are needed as well. However, manual annotation is very time-consuming and requires considerable skill and near-native competence. Therefore it can take years of speech corpus compilation and annotation before any analyses can be carried out. In this paper, approaches that address the transcription bottleneck of speech corpus exploitation are presented and discussed, including crowdsourcing the orthographic transcription, automatic phonetic alignment, and query-driven annotation. Currently, query-driven annotation and automatic phonetic alignment are being combined and applied in two speech research projects at the Institut für Deutsche Sprache (IDS), whereas crowdsourcing the orthographic transcription still awaits implementation.