Refine
Year of publication
- 2024 (39)
- 2023 (64)
- 2022 (80)
- 2021 (75)
- 2020 (90)
- 2019 (84)
- 2018 (77)
- 2017 (50)
- 2016 (39)
- 2015 (36)
- 2014 (45)
- 2013 (47)
- 2012 (56)
- 2011 (49)
- 2010 (55)
- 2009 (49)
- 2008 (42)
- 2007 (61)
- 2006 (45)
- 2005 (52)
- 2004 (62)
- 2003 (54)
- 2002 (45)
- 2001 (30)
- 2000 (28)
- 1999 (11)
- 1998 (6)
- 1997 (4)
- 1996 (6)
- 1995 (8)
- 1994 (3)
- 1993 (2)
- 1992 (3)
- 1991 (1)
- 1990 (2)
- 1989 (2)
- 1988 (2)
- 1987 (2)
- 1986 (3)
- 1985 (1)
- 1984 (4)
- 1983 (2)
- 1980 (3)
- 1979 (1)
- 1978 (1)
- 1976 (1)
- 1973 (1)
- 1972 (3)
- 1970 (4)
- 1969 (1)
- (31)
Document Type
- Doctoral Thesis (844)
- Article (220)
- Book (113)
- Contribution to a Periodical (108)
- Working Paper (62)
- Part of Periodical (24)
- Part of a Book (22)
- Conference Proceedings (17)
- Other (15)
- Review (10)
Language
- German (846)
- English (525)
- French (75)
- Multiple languages (15)
- Russian (1)
Has Fulltext
- yes (1462) (remove)
Keywords
- Deutschland (90)
- Luxemburg (53)
- Stress (40)
- Schule (37)
- Schüler (33)
- Modellierung (29)
- Politischer Unterricht (29)
- Demokratie (28)
- Fernerkundung (25)
- Geschichte (24)
Institute
- Psychologie (212)
- Raum- und Umweltwissenschaften (212)
- Politikwissenschaft (132)
- Universitätsbibliothek (83)
- Rechtswissenschaft (77)
- Wirtschaftswissenschaften (66)
- Mathematik (65)
- Fachbereich 4 (63)
- Medienwissenschaft (57)
- Fachbereich 6 (51)
- Geschichte, mittlere und neuere (45)
- Fachbereich 1 (38)
- Fachbereich 3 (34)
- Informatik (31)
- Germanistik (28)
- Fachbereich 2 (27)
- Kunstgeschichte (23)
- Anglistik (20)
- Soziologie (20)
- Computerlinguistik und Digital Humanities (10)
- Philosophie (10)
- Romanistik (9)
- Fachbereich 5 (7)
- Pädagogik (6)
- Allgemeine Sprach- und Literaturwissenschaft (5)
- Ethnologie (5)
- Geschichte, alte (5)
- Klassische Philologie (4)
- Sinologie (4)
- Japanologie (3)
- Archäologie (2)
- Phonetik (2)
- Servicezentrum eSciences (2)
- Forschungszentrum Europa (1)
- Institut für Rechtspolitik (1)
- Pflegewissenschaft (1)
- Slavistik (1)
- Theologische Fakultät (1)
Der vorliegende Beitrag greift die öffentliche Diskussion um den rechtspolitischen Umgang mit Hass, Hetze und Antisemitismus auf, die insbesondere nach dem Terroranschlag der Hamas am 07.10.2023 an Intensität und Dringlichkeit zugenommen hat. Dabei beleuchtet er einerseits das Straf- und Zivilrecht, legt andererseits einen besonderen Fokus auf öffentlich-rechtliche Konstellationen. Auf jedem dieser Gebiete werden Schwächen und Potenziale des Rechts und der Rechtsprechung aufgezeigt, zugleich aber auch die Grenzen staatlicher Gewalt verdeutlicht. Denn letztlich handelt es sich um ein gesellschaftliches Problem, dem – trotz aller Notwendigkeit staatlichen Handelns – in erster Linie durch Information, und erst in zweiter Linie durch das Recht begegnet werden muss.
This thesis consists of four highly related chapters examining China’s rise in the aluminium industry. The first chapter addresses the conditions that allowed China, which first entered the market in the 1950s, to rise to world leadership in aluminium production. Although China was a latecomer, its re-entry into the market after the oil crises in the 1970s was a success and led to its ascent as the world’s largest aluminium producer by 2001. With an estimated production of 40.4 million tonnes in 2022, China represented almost 60% of the global output. Chapter 1 examines the factors underlying this success, such as the decline of international aluminium cartels, the introduction of innovative technology, the US granting China the MFN tariff status, Chinese-specific factors, and supportive government policies. Chapter 2 develops a mathematical model to analyze firms’ decisions in the short term. It examines how an incumbent with outdated technology and a new entrant with access to a new type of technology make strategic decisions, including the incumbent’s decision whether to deter entry, the production choice of firms, the optimal technology adoption rate of the newcomer, and cartel formation. Chapter 3 focuses on the adoption of new technology by firms upon market entry in four scenarios: firstly, a free market Cournot competition; secondly, a situation in which the government determines technology adoption rates; thirdly, a scenario in which the government controls both technology and production; and finally, a scenario where the government dictates technology adoption rates, production levels, and also the number of market participants. Chapter 4 applies the Spencer and Brander (1983) framework to examine strategic industrial policy. The model assumes that there are two exporting firms in two different countries that sell a product to a third country. We examine how the domestic firm is influenced by government intervention, such as the provision of a fixed-cost subsidy to improve its competitiveness relative to the foreign company. Chapter 4 initially investigates a scenario where only one government offers a fixed-cost subsidy, followed by an analysis of the case when both governments simultaneously provide financial help. Taken together, these chapters provide a comprehensive analysis of the strategic, technological, and political factors contributing to China’s leadership in the global aluminium industry.
Chapter 1: The Rise of China as a Latecomer in the Global Aluminium Industry
This chapter examines China’s remarkable transformation into a global leader in the aluminium industry, a sector in which the country accounted for approximately 58.9% of worldwide production in 2022. We examine how China, a latecomer to the aluminium industry that started off with labor-intensive technology in 1953, grew into the largest aluminium producer with some of the most advanced smelters in the world. This analysis identifies and discusses several opportunities that Chinese aluminium producers took advantage of. The first set of opportunities happened during the 1970s oil crises, which softened international competition and allowed China to acquire innovative smelting technology from Japan. The second set of opportunities started at about the same time when China opened its economy in 1978. The substantial demand for aluminium in China is influenced by both external and internal factors. Externally, the US granted China’s MFN tariff status in 1980 and China entered the World Trade Organization (WTO) in 2001. Both events contributed to a surge in Chinese aluminium consumption. Internally, China’s investment-led growth model boosted further its aluminium demand. Additional factors specific to China, such as low labor costs and the abundance of coal as an energy source, offer Chinese firms competitive advantages against international players. Furthermore, another window of opportunity is due to Chinese governmental policies, including phasing out old technology, providing subsidies, and gradually opening the economy to enhance domestic competition before expanding globally. By describing these elements, the study provides insights into the dynamic interplay of external circumstances and internal strategies that contributed to the success of the Chinese aluminium industry.
Chapter 2: Technological Change and Strategic Choices for Incumbent and New Entrant
This chapter introduces an oligopoly model that includes two actors: an incumbent and a potential entrant, that compete in the same market. We assume that two participants are located in different parts of the market: the incumbent is situated in area 1, whereas the potential entrant may venture into the other region, area 2. The incumbent exists in stage zero, where it can decide whether to deter the newcomer’s entry. A new type of technology exists in period one, when the newcomer may enter the market. In the short term, the incumbent is trapped with the outdated technology, while the new entrant may choose to partially or completely adopt the latest technology. Our results suggest the following: Firstly, the incumbent only tries to deter the new entrant if a condition for entry cost is met. Secondly, the new entrant is only interested in forming a cartel with the incumbent if a function of the ratio of the variable to new technology’s fixed-cost parameters is sufficiently high. Thirdly, if the newcomer asks to form a cartel, the incumbent will always accept this request. Finally, we can obtain the optimal new technology adoption rate for the newcomer.
Chapter 3: Technological Adoption and Welfare in Cournot Oligopoly
This study examines the difference between the optimal technology adoption rates chosen by firms in a homogeneous Cournot oligopoly and that preferred by a benevolent government upon firms’ market entry. To address the question of whether the technology choices of firms and government are similar, we analyze several different scenarios, which differ in the extent of government intervention in the market. Our results suggest a relationship between the number of firms in the market and the impact of government intervention on technology adoption rates. Especially in situations with a low number of firms that are interested in entering the market, greater government influence tends to lead to higher technology adoption rates of firms. Conversely, in scenarios with a higher number of firms and a government that lacks control over the number of market players, the technology adoption rate of firms will be highest when the government plays no role.
Chapter 4: International Technological Innovation and Industrial Strategies
Supporting domestic firms when they first enter the market may be seen as a favorable policy choice by governments around the world thanks to their ability to enhance the competitive advantage of domestic firms in non-cooperative competition against foreign enterprises (infant industry protection argument). This advantage may allow domestic firms to increase their market share and generate higher profits, thereby improving domestic welfare. This chapter utilizes the Spencer and Brander (1983) framework as a theoretical foundation to elucidate the effects of fixed-cost subsidies on firms’ production levels, technological innovations, and social welfare. The analysis examines two firms in different countries, each producing a homogeneous product that is sold in a third, separate country. We first examine the Cournot-Nash equilibrium in the absence of government intervention, followed by analyzing a scenario where just one government provides a financial subsidy for its domestic firm, and finally, we consider a situation where both governments simultaneously provide financial assistance for their respective firms. Our results suggest that governments aim to maximize social welfare by providing fixed-cost subsidies to their respective firms, finding themselves in a Chicken game scenario. Regarding technology innovation, subsidies lead to an increased technological adoption rate for recipient firms, regardless of whether one or both firms in a market receive support, compared to the situation without subsidies. The technology adoption rate of the recipient firm is higher than of its rival when only the recipient firm benefits from the fixed-cost subsidy. The lowest technology adoption rate of a firm occurs when the firm does not receive a fixed-cost subsidy, but its competitor does. Furthermore, global welfare will benefit the most in case when both exporting countries grant fixed-cost subsidies, and this welfare level is higher when only one country subsidizes than when no subsidies are provided by any country.
Im Rahmen psychologischer Wissenschaftskommunikation werden Plain Language Summaries (PLS, Kerwer et al., 2021) zunehmend bedeutsamer. Es handelt sich hierbei um
zugängliche, überblicksartige Zusammenfassungen, welche das Verständnis von Lai:innen
potenziell unterstützen und ihr Vertrauen in wissenschaftliche Forschung fördern können.
Dies erscheint speziell vor dem Hintergrund der Replikationskrise (Wingen et al., 2019) sowie Fehlinformationen in Online-Kontexten (Swire-Thompson & Lazer, 2020) relevant. Die
positiven Auswirkungen zweier Effekte auf Vertrauen sowie ihre mögliche Interaktion fanden im Kontext von PLS bisher kaum Berücksichtigung: Zum einen die einfache Darstellung von Informationen (Easiness-Effekt, Scharrer et al., 2012), zum anderen ein möglichst wissenschaftlicher Stil (Scientificness-Effekt, Thomm & Bromme, 2012). Diese Dissertation hat zum Ziel, im Kontext psychologischer PLS genauere Bestandteile beider Effekte zu identifizieren und den Einfluss von Einfachheit und Wissenschaftlichkeit auf Vertrauen zu beleuchten. Dazu werden drei Artikel zu präregistrierten Online-Studien mit deutschsprachigen Stichproben vorgestellt.
Im ersten Artikel wurden in zwei Studien verschiedene Textelemente psychologischer PLS systematisch variiert. Es konnte ein signifikanter Einfluss von Fachtermini, Informationen zur
Operationalisierung, Statistiken und dem Grad an Strukturierung auf die von Lai:innen berichtete Einfachheit der PLS beobachtet werden. Darauf aufbauend wurden im zweiten Artikel vier PLS, die von Peer-Review-Arbeiten abgeleitet wurden, in ihrer Einfachheit und
Wissenschaftlichkeit variiert und Lai:innen zu ihrem Vertrauen in die Texte und Autor:innen befragt. Hier ergab sich zunächst nur ein positiver Einfluss von Wissenschaftlichkeit auf
Vertrauen, während der Easiness-Effekt entgegen der Hypothesen ausblieb. Exploratorische Analysen legten jedoch einen positiven Einfluss der von Lai:innen subjektiv wahrgenommenen Einfachheit auf ihr Vertrauen sowie eine signifikante Interaktion mit der
wahrgenommenen Wissenschaftlichkeit nahe. Diese Befunde lassen eine vermittelnde Rolle der subjektiven Wahrnehmung von Lai:innen für beide Effekte vermuten. Im letzten Artikel
wurde diese Hypothese über Mediationsanalysen geprüft. Erneut wurden zwei PLS
präsentiert und sowohl die Wissenschaftlichkeit des Textes als auch die der Autor:in manipuliert. Der Einfluss höherer Wissenschaftlichkeit auf Vertrauen wurde durch die
subjektiv von Lai:innen wahrgenommene Wissenschaftlichkeit mediiert. Zudem konnten
dimensionsübergreifende Mediationseffekte beobachtet werden.
Damit trägt diese Arbeit über bestehende Forschung hinaus zur Klärung von Rahmenbedingungen des Easiness- und Scientificness-Effektes bei. Theoretische
Implikationen zur zukünftigen Definition von Einfachheit und Wissenschaftlichkeit, sowie
praktische Konsequenzen hinsichtlich unterschiedlicher Zielgruppen von
Wissenschaftskommunikation und dem Einfluss von PLS auf die Entscheidungsbildung von
Lai:innen werden diskutiert.
This meta-scientific dissertation comprises three research articles that investigated the reproducibility of psychological research. Specifically, they focused on the reproducibility of eye-tracking research on the one hand, and studying preregistration (i.e., the practice of publishing a study protocol before data collection or analysis) as one method to increase reproducibility on the other hand.
In Article I, it was demonstrated that eye-tracking data quality is influenced by both the utilized eye-tracker and the specific task it is measuring. That is, distinct strengths and weaknesses were identified in three devices (Tobii Pro X3-120, GP3 HD, EyeLink 1000+) in an extensive test battery. Consequently, both the device and specific task should be considered when designing new studies. Meanwhile, Article II focused on the current perception of preregistration in the psychological research community and future directions for improving this practice. The survey showed that many researchers intended to preregister their research in the future and had overall positive attitudes toward preregistration. However, various obstacles were identified currently hindering preregistration, which should be addressed to increase its adoption. These findings were supplemented by Article III, which took a closer look at one preregistration-specific tool: the PRP-QUANT Template. In a simulation trial and a survey, the template demonstrated high usability and emerged as a valuable resource to support researchers in using preregistration. Future revisions of the template could help to further facilitate this open science practice.
In this dissertation, the findings of the three articles are summarized and discussed regarding their implications and potential future steps that could be implemented to improve the reproducibility of psychological research.
Today, almost every modern computing device is equipped with multicore processors capable of efficient concurrent and parallel execution of threads. This processor feature can be leveraged by concurrent programming, which is a challenge for software developers for two reasons: first, it introduces a paradigm shift that requires a new way of thinking. Second, it can lead to issues that are unique to concurrent programs due to the non-deterministic, interleaved execution of threads. Consequently, the debugging of concurrency and related performance issues is a rather difficult and often tedious task. Developers still lack on thread-aware programming tools that facilitate the understanding of concurrent programs. Ideally, these tools should be part of their daily working environment, which typically includes an Integrated Development Environment (IDE). In particular, the way source code is visually presented in traditional source-code editors does not convey much information on whether the source code is executed concurrently or in parallel in the first place.
With this dissertation, we pursue the main goal of facilitating and supporting the understanding and debugging of concurrent programs. To this end, we formulate and utilize a visualization paradigm that particularly includes the display of interactive glyph-based visualizations embedded in the source-code editor close to their corresponding artifacts (in-situ).
To facilitate the implementation of visualizations that comply with our paradigm as plugins for IDEs, we designed, implemented and evaluated a programming framework called CodeSparks. After presenting the design goals and the architecture of the framework, we demonstrate its versatility with a total of fourteen plugins realized by different developers using the CodeSparks framework (CodeSparks plugins). With focus group interviews, we empirically investigated how developers of the CodeSparks plugins experienced working with the framework. Based on the plugins, deliberate design decisions and the interview results, we discuss to what extent we achieved our design goals. We found that the framework is largely target programming-language independent and that it supports the development of plugins for a wide range of source-code-related tasks while hiding most of the details of the underlying plugin development API.
In addition, we applied our visualization paradigm to thread-related runtime data from concurrent programs to foster the awareness of source code being executed concurrently or in parallel. As a result, we developed and designed two in-situ thread visualizations, namely ThreadRadar and ThreadFork, with the latter building on the former. Both thread visualizations are based on a debugging approach, which combines statistical profiling, thread-aware runtime metrics, clustering of threads on the basis of these metrics, and finally interactive glyph-based in-situ visualizations. To address scalability issues of the ThreadRadar in terms of space required and the number of displayable thread clusters, we designed a revised thread visualization. This revision also involved the question of how many thread clusters k should be computed in the first place. To this end, we conducted experiments with the clustering of threads for artifacts from a corpus of concurrent Java programs that include real-world Java applications and concurrency bugs. We found that the maximum k on the one hand and the optimal k according to four cluster validation indices on the other hand rarely exceed three. However, occasionally thread clusterings with k > 3 are available and also optimal. Consequently, we revised both the clustering strategy and the visualization as parts of our debugging approach, which resulted in the ThreadFork visualization. Both in-situ thread visualizations, including their additional features that support the exploration of the thread data, are implemented in a tool called CodeSparks-JPT, i.e., as a CodeSparks plugin for IntelliJ IDEA.
With various empirical studies, including anecdotal usage scenarios, a usability test, web surveys, hands-on sessions, questionnaires and interviews, we investigated quality aspects of the in-situ thread visualizations and their corresponding tools. First, by a demonstration study, we illustrated the usefulness of the ThreadRadar visualization in investigating and fixing concurrency bugs and a performance bug. This was confirmed by a subsequent usability test and interview, which also provided formative feedback. Second, we investigated the interpretability and readability of the ThreadFork glyphs as well as the effectiveness of the ThreadFork visualization through anonymous web surveys. While we have found that the ThreadFork glyphs are correctly interpreted and readable, it remains unproven that the ThreadFork visualization effectively facilitates understanding the dynamic behavior of threads that concurrently executed portions of source code. Moreover, the overall usability of CodeSparks-JPT is perceived as "OK, but not acceptable" as the tool has issues with its learnability and memorability. However, all other usability aspects of CodeSparks-JPT that were examined are perceived as "above average" or "good".
Our work supports software-engineering researchers and practitioners in flexibly and swiftly developing novel glyph-based visualizations that are embedded in the source-code editor. Moreover, we provide in-situ thread visualizations that foster the awareness of source code being executed concurrently or in parallel. These in-situ thread visualizations can, for instance, be adapted, extended and used to analyze other use cases or to replicate the results. Through empirical studies, we have gradually shaped the design of the in-situ thread visualizations through data-driven decisions, and evaluated several quality aspects of the in-situ thread visualizations and the corresponding tools for their utility in understanding and debugging concurrent programs.
Representation Learning techniques play a crucial role in a wide variety of Deep Learning applications. From Language Generation to Link Prediction on Graphs, learned numerical vector representations often build the foundation for numerous downstream tasks.
In Natural Language Processing, word embeddings are contextualized and depend on their current context. This useful property reflects how words can have different meanings based on their neighboring words.
In Knowledge Graph Embedding (KGE) approaches, static vector representations are still the dominant approach. While this is sufficient for applications where the underlying Knowledge Graph (KG) mainly stores static information, it becomes a disadvantage when dynamic entity behavior needs to be modelled.
To address this issue, KGE approaches would need to model dynamic entities by incorporating situational and sequential context into the vector representations of entities. Analogous to contextualised word embeddings, this would allow entity embeddings to change depending on their history and current situational factors.
Therefore, this thesis provides a description of how to transform static KGE approaches to contextualised dynamic approaches and how the specific characteristics of different dynamic scenarios are need to be taken into consideration.
As a starting point, we conduct empirical studies that attempt to integrate sequential and situational context into static KG embeddings and investigate the limitations of the different approaches. In a second step, the identified limitations serve as guidance for developing a framework that enables KG embeddings to become truly dynamic, taking into account both the current situation and the past interactions of an entity. The two main contributions in this step are the introduction of the temporally contextualized Knowledge Graph formalism and the corresponding RETRA framework which realizes the contextualisation of entity embeddings.
Finally, we demonstrate how situational contextualisation can be realized even in static environments, where all object entities are passive at all times.
For this, we introduce a novel task that requires the combination of multiple context modalities and their integration with a KG based view on entity behavior.
This dissertation focusses on research into the personality construct of action vs. state orientation. Derived from the Personality-Systems-Interaction Theory (PSI Theory), state orientation is defined as a low ability to self-regulate emotions and associated with many adverse consequences – especially under stress. Because of the high prevalence of state orientation, it is a very important topic to investigate factors that help state-oriented people to buffer these adverse consequences. Action orientation, in contrast, is defined as a high ability to self-regulate own emotions in a very specific way: through accessing the self. The present dissertation demonstrates this theme in five studies, using a total of N = 1251 participants with a wide age range, encompassing different populations (students, non-student population (people from the coaching and therapy sector), applying different operationalisations to investigate self-access as a mediator or an outcome variable. Furthermore, it is tested whether the popular technique of mindfulness - that is advertised as a potent remedy for bringing people closer to the self -really works for everybody. The findings show that the presumed remedy is rather harmful for state-oriented individuals. Finally, an attempt to ameliorate these alienating effects, the present dissertation attempts to find theory-driven, and easy-to-apply solution how mindfulness exercises can be adapted.
Differential equations yield solutions that necessarily contain a certain amount of regularity and are based on local interactions. There are various natural phenomena that are not well described by local models. An important class of models that describe long-range interactions are the so-called nonlocal models, which are the subject of this work.
The nonlocal operators considered here are integral operators with a finite range of interaction and the resulting models can be applied to anomalous diffusion, mechanics and multiscale problems.
While the range of applications is vast, the applicability of nonlocal models can face problems such as the high computational and algorithmic complexity of fundamental tasks. One of them is the assembly of finite element discretizations of truncated, nonlocal operators.
The first contribution of this thesis is therefore an openly accessible, documented Python code which allows to compute finite element approximations for nonlocal convection-diffusion problems with truncated interaction horizon.
Another difficulty in the solution of nonlocal problems is that the discrete systems may be ill-conditioned which complicates the application of iterative solvers. Thus, the second contribution of this work is the construction and study of a domain decomposition type solver that is inspired by substructuring methods for differential equations. The numerical results are based on the abstract framework of nonlocal subdivisions which is introduced here and which can serve as a guideline for general nonlocal domain decomposition methods.
Sowohl national als auch international wird die zunehmende Digitalisierung von Prozessen gefordert. Die Heterogenität und Komplexität der dabei entstehenden Systeme erschwert die Partizipation für reguläre Nutzergruppen, welche zum Beispiel kein Expertenwissen in der Programmierung oder einen informationstechnischen Hintergrund aufweisen. Als Beispiel seien hier Smart Contracts genannt, deren Programmierung komplex ist und bei denen etwaige Fehler unmittelbar mit monetärem Verlust durch die direkte Verknüpfung der darunterliegenden Kryptowährung verbunden sind. Die vorliegende Arbeit stellt ein alternatives Protokoll für cyber-physische Verträge vor, das sich besonders gut für die menschliche Interaktion eignet und auch von regulären Nutzergruppen verstanden werden kann. Hierbei liegt der Fokus auf der Transparenz der Übereinkünfte und es wird weder eine Blockchain noch eine darauf beruhende digitale Währung verwendet. Entsprechend kann das Vertragsmodell der Arbeit als nachvollziehbare Verknüpfung zwischen zwei Parteien verstanden werden, welches die unterschiedlichen Systeme sicher miteinander verbindet und so die Selbstorganisation fördert. Diese Verbindung kann entweder computergestützt automatisch ablaufen, oder auch manuell durchgeführt werden. Im Gegensatz zu Smart Contracts können somit Prozesse Stück für Stück digitalisiert werden. Die Übereinkünfte selbst können zur Kommunikation, aber auch für rechtlich bindende Verträge genutzt werden. Die Arbeit ordnet das neue Konzept in verwandte Strömungen wie Ricardian oder Smart Contracts ein und definiert Ziele für das Protokoll, welche in Form der Referenzimplementierung umgesetzt werden. Sowohl das Protokoll als auch die Implementierung werden im Detail beschrieben und durch eine Erweiterung der Anwendung ergänzt, welche es Nutzenden in Regionen ohne direkte Internetverbindung ermöglicht, an ebenjenen Verträgen teilnehmen zu können. Weiterhin betrachtet die Evaluation die rechtlichen Rahmenbedinungen, die Übertragung des Protokolls auf Smart Contracts und die Performanz der Implementierung.
Die chinesische und westliche Forschung, die sich mit der Beziehung zwischen chinesischer Kultur und katholischer Kirche befasst, konzentriert sich in der Regel auf die katholische Kirche in China vor dem Verbot des Christentums. Die einzigartige Perspektive dieser Arbeit besteht darin, die Veränderungen in der Beziehung zwischen den beiden vom Ende der Ming-Dynastie bis zur ersten Hälfte des 20. Jahrhunderts zu untersuchen. Vor dem Verbot nährten die katholischen Missionare den konfuzianischen Gelehrten und verbanden die katholische Lehre mit dem Konfuzianismus, um ihren Einfluss in der Oberschicht der chinesischen Gesellschaft auszuüben. Nach dem Verbot achteten die katholischen Missionare nicht so sehr auf ihre Beziehung zur chinesischen Kultur wie ihre Vorgänger im 17. und 18. Jahrhundert. Einige Missionare sowie chinesische Katholiken wollten die Situation ändern und förderten gemeinsam die Gründung der Fu-Jen-Universität, die großen Wert auf die chinesische Kultur legte und die Beziehung zwischen der Katholischen Kirche und der chinesischen Kultur Anfang des 20. Jahrhunderts widerspiegeln konnte. Die Professoren der Abteilung Chinesisch und Geschichte leisteten den größten Beitrag zur Forschung der chinesischen Kultur an der Universität. Im Vergleich zu anderen wichtigen Universitäten in Peking, wo die chinesische Literatur im Fachbereich Chinesisch eine zentrale Stellung einnahm, legte die Fu-Jen-Universität mehr Wert auf die chinesische Sprache und Schriftzeichen. Anfang des 20. Jahrhunderts erlangten Frauen unter dem Einfluss der globalen feministischen Bewegung das Recht auf Hochschulbildung. Bis 1920 waren jedoch die katholischen Universitäten in Bezug auf die Hochschulbildung von Frauen Jahrzehnte hinter den protestantischen und nichtkirchlichen Universitäten zurückgefallen. Die Fu-Jen-Universität verbesserte diese Situation, indem sie nicht nur eine große Anzahl von Studentinnen annahm, sondern ihnen eine Vielzahl von Fächern einschließlich Chinesisch und Geschichte anbot. Im Allgemeinen konnte die Universität als Verbindung zwischen dem Katholizismus und der chinesischen Kultur in der ersten Hälfte des 20. Jahrhunderts angesehen werden. Sie spielte eine wichtige Rolle nicht nur bei der Erforschung und Verbreitung der chinesischen Kultur, sondern auch bei der Ausweitung des Einflusses der katholischen Kirche zu dieser Zeit.
Physically-based distributed rainfall-runoff models as the standard analysis tools for hydro-logical processes have been used to simulate the water system in detail, which includes spa-tial patterns and temporal dynamics of hydrological variables and processes (Davison et al., 2015; Ek and Holtslag, 2004). In general, catchment models are parameterized with spatial information on soil, vegetation and topography. However, traditional approaches for eval-uation of the hydrological model performance are usually motivated with respect to dis-charge data alone. This may thus cloud model realism and hamper understanding of the catchment behavior. It is necessary to evaluate the model performance with respect to in-ternal hydrological processes within the catchment area as well as other components of wa-ter balance rather than runoff discharge at the catchment outlet only. In particular, a consid-erable amount of dynamics in a catchment occurs in the processes related to interactions of the water, soil and vegetation. Evapotranspiration process, for instance, is one of those key interactive elements, and the parameterization of soil and vegetation in water balance mod-eling strongly influences the simulation of evapotranspiration. Specifically, to parameterize the water flow in unsaturated soil zone, the functional relationships that describe the soil water retention and hydraulic conductivity characteristics are important. To define these functional relationships, Pedo-Transfer Functions (PTFs) are common to use in hydrologi-cal modeling. Opting the appropriate PTFs for the region under investigation is a crucial task in estimating the soil hydraulic parameters, but this choice in a hydrological model is often made arbitrary and without evaluating the spatial and temporal patterns of evapotran-spiration, soil moisture, and distribution and intensity of runoff processes. This may ulti-mately lead to implausible modeling results and possibly to incorrect decisions in regional water management. Therefore, the use of reliable evaluation approaches is continually re-quired to analyze the dynamics of the current interactive hydrological processes and predict the future changes in the water cycle, which eventually contributes to sustainable environ-mental planning and decisions in water management.
Remarkable endeavors have been made in development of modelling tools that provide insights into the current and future of hydrological patterns in different scales and their im-pacts on the water resources and climate changes (Doell et al., 2014; Wood et al., 2011). Although, there is a need to consider a proper balance between parameter identifiability and the model's ability to realistically represent the response of the natural system. Neverthe-less, tackling this issue entails investigation of additional information, which usually has to be elaborately assembled, for instance, by mapping the dominant runoff generation pro-cesses in the intended area, or retrieving the spatial patterns of soil moisture and evapotran-spiration by using remote sensing methods, and evaluation at a scale commensurate with hydrological model (Koch et al., 2022; Zink et al., 2018). The present work therefore aims to give insights into the modeling approaches to simulate water balance and to improve the soil and vegetation parameterization scheme in the hydrological model subject to producing more reliable spatial and temporal patterns of evapotranspiration and runoff processes in the catchment.
An important contribution to the overall body of work is a book chapter included among publications. The book chapter provides a comprehensive overview of the topic and valua-ble insights into the understanding the water balance and its estimation methods.
Moreover, the first paper aimed to evaluate the hydrological model behavior with re-spect to contribution of various sources of information. To do so, a multi-criteria evaluation metric including soft and hard data was used to define constraints on outputs of the 1-D hydrological model WaSiM-ETH. Applying this evaluation metric, we could identify the optimal soil and vegetation parameter sets that resulted in a “behavioral” forest stand water balance model. It was found out that even if simulations of transpiration and soil water con-tent are consistent with measured data, but still the dominant runoff generation processes or total water balance might be wrongly calculated. Therefore, only using an evaluation scheme which looks over different sources of data and embraces an understanding of the local controls of water loss through soil and plant, allowed us to exclude the unrealistic modeling outputs. The results suggested that we may need to question the generally accept-ed soil parameterization procedures that apply default parameter sets.
The second paper attempts to tackle the pointed model evaluation hindrance by getting down to the small-scale catchment (in Bavaria). Here, a methodology was introduced to analyze the sensitivity of the catchment water balance model to the choice of the Pedo-Transfer Functions (PTF). By varying the underlying PTFs in a calibrated and validated model, we could determine the resulting effects on the spatial distribution of soil hydraulic properties, total water balance in catchment outlet, and the spatial and temporal variation of the runoff components. Results revealed that the water distribution in the hydrologic system significantly differs amongst various PTFs. Moreover, the simulations of water balance components showed high sensitivity to the spatial distribution of soil hydraulic properties. Therefore, it was suggested that opting the PTFs in hydrological modeling should be care-fully tested by looking over the spatio-temporal distribution of simulated evapotranspira-tion and runoff generation processes, whether they are reasonably represented.
To fulfill the previous studies’ suggestions, the third paper then aims to focus on evalu-ating the hydrological model through improving the spatial representation of dominant run-off processes. It was implemented in a mesoscale catchment in southwestern Germany us-ing the hydrological model WaSiM-ETH. Dealing with the issues of inadequate spatial ob-servations for rigorous spatial model evaluation, we made use of a reference soil hydrologic map available for the study area to discern the expected dominant runoff processes across a wide range of hydrological conditions. The model was parameterized by applying 11 PTFs and run by multiple synthetic rainfall events. To compare the simulated spatial patterns to the patterns derived by digital soil map, a multiple-component spatial performance metric (SPAEF) was applied. The simulated DRPs showed a large variability with regard to land use, topography, applied rainfall rates, and the different PTFs, which highly influence the rapid runoff generation under wet conditions.
The three published manuscripts proceeded towards the model evaluation viewpoints that ultimately attain the behavioral model outputs. It was performed through obtaining information about internal hydrological processes that lead to certain model behaviors, and also about the function and sensitivity of some of the soil and vegetation parameters that may primarily influence those internal processes in a catchment. Accordingly, using this understanding on model reactions, and by setting multiple evaluation criteria, it was possi-ble to identify which parameterization could lead to behavioral model realization. This work, in fact, will contribute to solving some of the issues (e.g., spatial variability and modeling methods) identified as the 23 unsolved problems in hydrology in the 21st century (Blöschl et al., 2019). The results obtained in the present work encourage the further inves-tigations toward a comprehensive model calibration procedure considering multiple data sources simultaneously. This will enable developing the new perspectives to the current parameter estimation methods, which in essence, focus on reproducing the plausible dy-namics (spatio-temporal) of the other hydrological processes within the watershed.
Sozialunternehmen haben mindestens zwei Ziele: die Erfüllung ihrer sozialen bzw. ökologischen Mission und finanzielle Ziele. Zwischen diesen Zielen können Spannungen entstehen. Wenn sie sich in diesem Spannungsfeld wiederholt zugunsten der finanziellen Ziele entscheiden, kommt es zum Mission Drift. Die Priorisierung der finanziellen Ziele überlagert dabei die soziale Mission. Auch wenn das Phänomen in der Praxis mehrfach beobachtet und in Einzelfallanalysen beschrieben wurde, gibt es bislang wenig Forschung zu Mission Drift. Der Fokus der vorliegenden Arbeit liegt darauf, diese Forschungslücke zu schließen und eigene Erkenntnisse für die Auslöser und Treiber des Mission Drifts von Sozialunternehmen zu ermitteln. Ein Augenmerk liegt auf den verhaltensökonomischen Theorien und der Mixed-Gamble-Logik. Dieser Logik zufolge liegt bei Entscheidungen immer eine Gleichzeitigkeit von Gewinnen und Verlusten vor, sodass Entscheidungsträger die Furcht vor Verlusten gegenüber der Aussicht auf Gewinne abwägen müssen. Das Modell wird genutzt, um eine neue theoretische Betrachtungsweise auf die Abwägung zwischen sozialen und finanziellen Zielen bzw. Mission Drift zu erhalten. Mit einem Conjoint Experiment werden Daten über das Entscheidungsverhalten von Sozialunternehmern generiert. Im Zentrum steht die Abwägung zwischen sozialen und finanziellen Zielen in verschiedenen Szenarien (Krisen- und Wachstumssituationen). Mithilfe einer eigens erstellten Stichprobe von 1.222 Sozialunternehmen aus Deutschland, Österreich und der Schweiz wurden 187 Teilnehmende für die Studie gewonnen. Die Ergebnisse dieser Arbeit zeigen, dass eine Krisensituation Auslöser für Mission Drift von Sozialunternehmen sein kann, weil in diesem Szenario den finanziellen Zielen die größte Bedeutung zugemessen wird. Für eine Wachstumssituation konnten hingegen keine solche Belege gefunden werden. Hinzu kommen weitere Einflussfaktoren, welche die finanzielle Orientierung verstärken können, nämlich die Gründeridentitäten der Sozialunternehmer, eine hohe Innovativität der Unternehmen und bestimmte Stakeholder. Die Arbeit schließt mit einer ausführlichen Diskussion der Ergebnisse. Es werden Empfehlungen gegeben, wie Sozialunternehmen ihren Zielen bestmöglich treu bleiben können. Außerdem werden die Limitationen der Studie und Wege für zukünftige Forschung im Bereich Mission Drift aufgezeigt.
Semantic-Aware Coordinated Multiple Views for the Interactive Analysis of Neural Activity Data
(2024)
Visualizing brain simulation data is in many aspects a challenging task. For one, data used in brain simulations and the resulting datasets is heterogeneous and insight is derived by relating all different kinds of it. Second, the analysis process is rapidly changing while creating hypotheses about the results. Third, the scale of data entities in these heterogeneous datasets is manifold, reaching from single neurons to brain areas interconnecting millions. Fourth, the heterogeneous data consists of a variety of modalities, e.g.: from time series data to connectivity data, from single parameters to a set of parameters spanning parameter spaces with multiple possible and biological meaningful solutions; from geometrical data to hierarchies and textual descriptions, all on mostly different scales. Fifth, visualizing includes finding suitable representations and providing real-time interaction while supporting varying analysis workflows. To this end, this thesis presents a scalable and flexible software architecture for visualizing, integrating and interacting with brain simulations data. The scalability and flexibility is achieved by interconnected services forming in a series of Coordinated Multiple View (CMV) systems. Multiple use cases are presented, introducing views leveraging this architecture, extending its ecosystem and resulting in a Problem Solving Environment (PSE) from which custom-tailored CMV systems can be build. The construction of such CMV system is assisted by semantic reasoning hence the term semantic-aware CMVs.
Information in der vorvertraglichen Phase – das heißt, Informationspflichten sowie Rechtsfolgen von Informationserteilung und -nichterteilung – in Bezug auf Kaufvertrag und Wahl des optionalen Instruments hat im Vorschlag der Europäischen Kommission für ein Gemeinsames Europäisches Kaufrecht (GEK; KOM(2011) 635) vielfältige Regelungen erfahren. Die vorliegende Arbeit betrachtet diese Regelungen auch in ihrem Verhältnis zu den Textstufen des Europäischen Privatrechts – Modellregeln und verbraucherschützende EU-Richtlinien – und misst sie an ökonomischen Rahmenbedingungen, die die Effizienz von Transaktionen gebieten und Grenzen des Nutzens von (Pflicht-)Informationen aufzeigen.
Vom Grundsatz der Vertragsfreiheit ausgehend ist jeder Partei das Risiko zugewiesen, unzureichend informiert zu sein, während die Gegenseite nur punktuell zur Information verpflichtet ist. Zwischen Unternehmern bleibt es auch nach dem GEK hierbei, doch zwischen Unternehmer und Verbraucher wird dieses Verhältnis umgekehrt. Dort gelten, mit Differenzierung nach Vertragsschlusssituationen, umfassende Kataloge von Informationspflichten hinsichtlich des Kaufvertrags. Als Konzept ist dies grundsätzlich sinnvoll; die Pflichten dienen dem Verbraucherschutz, insbesondere der Informiertheit und Transparenz vor der Entscheidung über den Vertragsschluss. Teilweise gehen die Pflichten aber zu weit. Die Beeinträchtigung der Vertragsfreiheit des Unternehmers durch die Pflichten und die Folgen ihrer Verletzung lässt sich nicht vollständig mit dem Ziel des Verbraucherschutzes rechtfertigen. Durch das Übermaß an Information fördern die angeordneten Pflichten den Verbraucherschutz nur eingeschränkt; sie genügen nicht verhaltensökonomischen Maßstäben. Es empfiehlt sich daher, zwischen Unternehmern und Verbrauchern bestimmte verpflichtende Informationsinhalte ganz zu streichen, auf im konkreten Fall nicht erforderliche Information zu verzichten, erst nach Vertragsschluss relevante Informationen auf diese Zeit zu verschieben und die verbleibenden vorvertraglichen Pflichtinformationen in einer für den Verbraucher besser zu verarbeitenden Weise zu präsentieren. Von den einem Verbraucher zu erteilenden Informationen sollte stets verlangt werden, dass sie klar und verständlich sind; die Beweislast für ihre ordnungsgemäße Erteilung sollte generell dem Unternehmer obliegen.
Neben die ausdrücklich angeordneten Informationspflichten treten ungeachtet der Verbraucher- oder Unternehmereigenschaft sowie der Käufer- oder Verkäuferrolle stark einzelfallabhängige Informationspflichten nach Treu und Glauben, die im Recht der Willensmängel niedergelegt sind. Hier ist der Grundsatz verwirklicht, dass mangelnde Information zunächst das eigene Risiko jeder Partei ist; berechtigtes Vertrauen und freie Willensbildung werden geschützt. Diese Pflichten berücksichtigen auch das Ziel der Effizienz und achten die Vertragsfreiheit. Das Vertrauen auf jegliche erteilten Informationen wird zudem dadurch geschützt, dass sie den Vertragsinhalt – allerdings in Verbraucherverträgen nicht umfassend genug – mitbestimmen können und dass ihre Unrichtigkeit sanktioniert wird.
Die Verletzung jeglicher Arten von Informationspflichten kann insbesondere einen Schadensersatzanspruch sowie – über das Recht der Willensmängel – die Möglichkeit zur Lösung vom Vertrag nach sich ziehen. Das Zusammenspiel der unterschiedlichen Mechanismen führt allerdings zu Friktionen sowie zu Lücken in den Rechtsfolgen von Informationspflichtverletzungen. Daher empfiehlt sich die Schaffung eines Schadensersatzanspruchs für jede treuwidrig unterlassene Informationserteilung; hierdurch wird das Gebot von Treu und Glauben auch außerhalb des Rechts der Willensmängel zu einer eigentlichen einzelfallabhängigen Informationspflicht aufgewertet.
Social entrepreneurship is a successful activity to solve social problems and economic challenges. Social entrepreneurship uses for-profit industry techniques and tools to build financially sound businesses that provide nonprofit services. Social entrepreneurial activities also lead to the achievement of sustainable development goals. However, due to the complex, hybrid nature of the business, social entrepreneurial activities are typically supported by macrolevel determinants. To expand our knowledge of how beneficial macro-level determinants can be, this work examines empirical evidence about the impact of macro-level determinants on social entrepreneurship. Another aim of this dissertation is to examine the impact at the micro level, as the growth ambitions of social and commercial entrepreneurs differ. At the beginning, the introductory section is explained in Chapter 1, which contains the motivation for the research, the research question, and the structure of the work.
There is an ongoing debate about the origin and definition of social entrepreneurship. Therefore, the numerous phenomena of social entrepreneurship are examined theoretically in the previous literature. To determine the common consensus on the topic, Chapter 2 presents
the theoretical foundations and definition of social entrepreneurship. The literature shows that a variety of determinants at the micro and macro levels are essential for the emergence of social entrepreneurship as a distinctive business model (Hartog & Hoogendoorn, 2011; Stephan et al., 2015; Hoogendoorn, 2016). It is impossible to create a society based on a social mission without the support of micro and macro-level-level determinants. This work examines the determinants and consequences of social entrepreneurship from different methodological perspectives. The theoretical foundations of the micro- and macro-level determinants influencing social entrepreneurial activities were discussed in Chapter 3. The purpose of reproducibility in research is to confirm previously published results (Hubbard et al., 1998; Aguinis & Solarino, 2019). However, due to the lack of data, lack of transparency of methodology, reluctance to publish, and lack of interest from researchers, there is a lack of promoting replication of the existing research study (Baker, 2016; Hedges & Schauer, 2019a). Promoting replication studies has been regularly emphasized in the business and management literature (Kerr et al., 2016; Camerer et al., 2016). However, studies that provide replicability of the reported results are considered rare in previous research (Burman et al., 2010; Ryan & Tipu, 2022). Based on the research of Köhler and Cortina (2019), an empirical study on this topic is carried out in Chapter 4 of this work.
Given this focus, researchers have published a large body of research on the impact of microand macro-level determinants on social inclusion, although it is still unclear whether these studies accurately reflect reality. It is important to provide conceptual underpinnings to the field through a reassessment of published results (Bettis et al., 2016). The results of their research make it abundantly clear that the macro determinants support social entrepreneurship.
In keeping with the more narrative approach, which is a crucial concern and requires attention, Chapter 5 considered the reproducibility of previous results, particularly on the topic of social entrepreneurship. We replicated the results of Stephan et al. (2015) to establish the trend of reproducibility and validate the specific conclusions they drew. The literal and constructive replication in the dissertation inspired us to explore technical replication research on social entrepreneurship. Chapter 6 evaluates the fundamental characteristics that have proven to be key factors in the growth of social ventures. The current debate reviews and references literature that has specifically focused on the development of social entrepreneurship. An empirical analysis of factors directly related to the ambitious growth of social entrepreneurship is also carried out.
Numerous social entrepreneurial groups have been studied concerning this association. Chapter 6 compares the growth ambitions of social and traditional (commercial) entrepreneurship as consequences at the micro level. This study examined many characteristics of social and commercial entrepreneurs' growth ambitions. Scholars have claimed to some extent that the growth of social entrepreneurship differs from commercial entrepreneurial activities due to objectivity differences (Lumpkin et al., 2013; Garrido-Skurkowicz et al., 2022). Qualitative research has been used in studies to support the evidence on related topics, including Gupta et al (2020) emphasized that research needs to focus on specific concepts of social entrepreneurship for the field to advance. Therefore, this study provides a quantitative, analysis-based assessment of facts and data. For this purpose, a data set from the Global Entrepreneurship Monitor (GEM) 2015 was used, which examined 12,695 entrepreneurs from 38 countries. Furthermore, this work conducted a regression analysis to evaluate the influence of various social and commercial characteristics of entrepreneurship on economic growth in developing countries. Chapter 7 briefly explains future directions and practical/theoretical implications.
Data fusions are becoming increasingly relevant in official statistics. The aim of a data fusion is to combine two or more data sources using statistical methods in order to be able to analyse different characteristics that were not jointly observed in one data source. Record linkage of official data sources using unique identifiers is often not possible due to methodological and legal restrictions. Appropriate data fusion methods are therefore of central importance in order to use the diverse data sources of official statistics more effectively and to be able to jointly analyse different characteristics. However, the literature lacks comprehensive evaluations of which fusion approaches provide promising results for which data constellations. Therefore, the central aim of this thesis is to evaluate a concrete plethora of possible fusion algorithms, which includes classical imputation approaches as well as statistical and machine learning methods, in selected data constellations.
To specify and identify these data contexts, data and imputation-related scenario types of a data fusion are introduced: Explicit scenarios, implicit scenarios and imputation scenarios. From these three scenario types, fusion scenarios that are particularly relevant for official statistics are selected as the basis for the simulations and evaluations. The explicit scenarios are the fulfilment or violation of the Conditional Independence Assumption (CIA) and varying sample sizes of the data to be matched. Both aspects are likely to have a direct, that is, explicit, effect on the performance of different fusion methods. The summed sample size of the data sources to be fused and the scale level of the variable to be imputed are considered as implicit scenarios. Both aspects suggest or exclude the applicability of certain fusion methods due to the nature of the data. The univariate or simultaneous, multivariate imputation solution and the imputation of artificially generated or previously observed values in the case of metric characteristics serve as imputation scenarios.
With regard to the concrete plethora of possible fusion algorithms, three classical imputation approaches are considered: Distance Hot Deck (DHD), the Regression Model (RM) and Predictive Mean Matching (PMM). With Decision Trees (DT) and Random Forest (RF), two prominent tree-based methods from the field of statistical learning are discussed in the context of data fusion. However, such prediction methods aim to predict individual values as accurately as possible, which can clash with the primary objective of data fusion, namely the reproduction of joint distributions. In addition, DT and RF only comprise univariate imputation solutions and, in the case of metric variables, artificially generated values are imputed instead of real observed values. Therefore, Predictive Value Matching (PVM) is introduced as a new, statistical learning-based nearest neighbour method, which could overcome the distributional disadvantages of DT and RF, offers a univariate and multivariate imputation solution and, in addition, imputes real and previously observed values for metric characteristics. All prediction methods can form the basis of the new PVM approach. In this thesis, PVM based on Decision Trees (PVM-DT) and Random Forest (PVM-RF) is considered.
The underlying fusion methods are investigated in comprehensive simulations and evaluations. The evaluation of the various data fusion techniques focusses on the selected fusion scenarios. The basis for this is formed by two concrete and current use cases of data fusion in official statistics, the fusion of EU-SILC and the Household Budget Survey on the one hand and of the Tax Statistics and the Microcensus on the other. Both use cases show significant differences with regard to different fusion scenarios and thus serve the purpose of covering a variety of data constellations. Simulation designs are developed from both use cases, whereby the explicit scenarios in particular are incorporated into the simulations.
The results show that PVM-RF in particular is a promising and universal fusion approach under compliance with the CIA. This is because PVM-RF provides satisfactory results for both categorical and metric variables to be imputed and also offers a univariate and multivariate imputation solution, regardless of the scale level. PMM also represents an adequate fusion method, but only in relation to metric characteristics. The results also imply that the application of statistical learning methods is both an opportunity and a risk. In the case of CIA violation, potential correlation-related exaggeration effects of DT and RF, and in some cases also of RM, can be useful. In contrast, the other methods induce poor results if the CIA is violated. However, if the CIA is fulfilled, there is a risk that the prediction methods RM, DT and RF will overestimate correlations. The size ratios of the studies to be fused in turn have a rather minor influence on the performance of fusion methods. This is an important indication that the larger dataset does not necessarily have to serve as a donor study, as was previously the case.
The results of the simulations and evaluations provide concrete implications as to which data fusion methods should be used and considered under the selected data and imputation constellations. Science in general and official statistics in particular benefit from these implications. This is because they provide important indications for future data fusion projects in order to assess which specific data fusion method could provide adequate results along the data constellations analysed in this thesis. Furthermore, with PVM this thesis offers a promising methodological innovation for future data fusions and for imputation problems in general.
Der zentrale Gegenstand der Untersuchung ist die Rechtsfigur des Indigenats im Kontext der württembergischen und preußischen Staatenlandschaft. Das Indigenat lässt sich als ein Recht bestimmen, das seine potenziellen Rechtsträger maßgeblich über das Abstammungsprinzip definiert und ein Verhältnis zwischen Rechtsträger und einem übergeordneten Rechtssubjekt zum Ausdruck bringt, sei es lehns- oder standes-, staats- oder auch bundes- beziehungsweise reichsrechtlicher Natur. Der zeitliche Schwerpunkt der Betrachtung liegt auf dem 19. Jahrhundert. Es werden jedoch auch Rückblicke in die Frühe Neuzeit geworfen, weil Wandel und Kontinuität in der Entwicklung des Indigenats in einer solch langen Perspektive besonders klar hervortreten können. Die zentrale These dieser Arbeit ist, dass ein enger Zusammenhang zwischen der im 19. Jahrhundert entstehenden und bis heute geläufigen Form der Zuordnung von Menschen zum Staat und den aus diesem Verhältnis entspringenden Rechten einerseits und dem frühneuzeitlichen Indigenat andererseits besteht. Dabei kann gezeigt werden, dass Gesellschaften ihre politischen Machtpositionen gegenüber „fremdstämmigen“, etwa zuwandernden Personen abschirmten, indem sie sich auf indigenatrechtliche, ethnische Bestimmungen beriefen.
This thesis contains four parts that are all connected by their contributions to the Efficient Market Hypothesis and decision-making literature. Chapter two investigates how national stock market indices reacted to the news of national lockdown restrictions in the period from January to May 2020. The results show that lockdown restrictions led to different reactions in a sample of OECD and BRICS countries: there was a general negative effect resulting from the increase in lockdown restrictions, but the study finds strong evidence for underreaction during the lockdown announcement, followed by some overreaction that is corrected subsequently. This under-/overreaction pattern, however, is observed mostly during the first half of our time series, pointing to learning effects. Relaxation of the lockdown restrictions, on the other hand, had a positive effect on markets only during the second half of our sample, while for the first half of the sample, the effect was negative. The third chapter investigates the gender differences in stock selection preferences on the Taiwan Stock Exchange. By utilizing trading data from the Taiwan Stock Exchange over a span of six years, it becomes possible to analyze trading behavior while minimizing the self-selection bias that is typically present in brokerage data. To study gender differences, this study uses firm-level data. The percentage of male traders in a company is the dependent variable, while the company’s industry and fundamental/technical aspects serve as independent variables. The results show that the percentage of women trading a company rises with a company’s age, market capitalization, a company’s systematic risk, and return. Men trade more frequently and show a preference for dividend-paying stocks and for industries with which they are more familiar. The fourth chapter investigated the relationship between regret and malicious and benign envy. The relationship is analyzed in two different studies. In experiment 1, subjects had to fill out psychological scales that measured regret, the two types of envy, core self-evaluation and the big 5 personality traits. In experiment 2, felt regret is measured in a hypothetical scenario, and the subject’s felt regret was regressed on the other variables mentioned above. The two experiments revealed that there is a positive direct relationship between regret and benign envy. The relationship between regret and malicious envy, on the other hand, is mostly an artifact of core self-evaluation and personality influencing both malicious envy and regret. The relationship can be explained by the common action tendency of self-improvement for regret and benign envy. Chapter five discusses the differences in green finance regulation and implementation between the EU and China. China introduced the Green Silk Road, while the EU adopted the Green Deal and started working with its own green taxonomy. The first difference comes from the definition of green finance, particularly with regard to coal-fired power plants. Especially the responsibility of nation-states’ emissions abroad. China is promoting fossil fuel projects abroad through its Belt and Road Initiative, but the EU’s Green Deal does not permit such actions. Furthermore, there are policies in both the EU and China that create contradictory incentives for economic actors. On the one hand, the EU and China are improving the framework conditions for green financing while, on the other hand, still allowing the promotion of conventional fuels. The role of central banks is also different between the EU and China. China’s central bank is actively working towards aligning the financial sector with green finance. A possible new role of the EU central bank or the priority financing of green sectors through political decision-making is still being debated.
When humans encounter attitude objects (e.g., other people, objects, or constructs), they evaluate them. Often, these evaluations are based on attitudes. Whereas most research focuses on univalent (i.e., only positive or only negative) attitude formation, little research exists on ambivalent (i.e., simultaneously positive and negative) attitude formation. Following a general introduction into ambivalence, I present three original manuscripts investigating ambivalent attitude formation. The first manuscript addresses ambivalent attitude formation from previously univalent attitudes. The results indicate that responding to a univalent attitude object incongruently leads to ambivalence measured via mouse tracking but not ambivalence measured via self-report. The second manuscript addresses whether the same number of positive and negative statements presented block-wise in an impression formation task leads to ambivalence. The third manuscript also used an impression formation task and addresses the question of whether randomly presenting the same number of positive and negative statements leads to ambivalence. Additionally, the effect of block size of the same valent statements is investigated. The results of the last two manuscripts indicate that presenting all statements of one valence and then all statements of the opposite valence leads to ambivalence measured via self-report and mouse tracking. Finally, I discuss implications for attitude theory and research as well as future research directions.
Left ventricular assist devices (LVADs) have become a valuable treatment for patients with advanced heart failure. Women appear to be disadvantaged in the usage of LVADs and concerning clinical outcomes such as death and adverse events after LVAD implant. Contrary to typical clinical characteristics (e.g., disease severity), device-related factors such as the intended device strategy, bridge to a heart transplantation or destination therapy, are often not considered in research on gender differences. In addition, the relevance of pre-implant psychosocial risk factors, such as substance abuse and limited social support, for LVAD outcomes is currently unclear. Thus, the aim of this dissertation is to explore the role of pre-implant psychosocial risk factors for gender differences in clinical outcomes, accounting for clinical and device-related risk factors.
In the first article, gender differences in pre-implant characteristics of patients registered in The European Registry for Patients with Mechanical Circulatory Support (EUROMACS) were investigated. It was found that women and men differed in multiple pre-implant characteristics depending on device strategy. In the second article, gender differences in major clinical outcomes (i.e., death, heart transplant, device explant due to cardiac recovery, device replacement due to complications) were evaluated for patients in the device strategy destination therapy in the Interagency Registry for Mechanically Assisted Circulation (INTERMACS). Additionally, the association of gender and psychosocial risk factors with the major outcomes were analyzed. Women had similar probabilities to die on LVAD support, and even higher probabilities to experience explant of the device due to cardiac recovery compared with men in the destination therapy subgroup. Pre-implant psychosocial risk factors were not associated with major outcomes. The third article focused on gender differences in 10 adverse events (e.g., device malfunction, bleeding) after LVAD implant in INTERMACS. The association of a psychosocial risk indicator with gender and adverse events after LVAD implantation was evaluated. Women were less likely to have psychosocial risk pre-implant but more likely to experience seven out of 10 adverse events compared with men. Pre-implant psychosocial risk was associated with adverse events, even suggesting a dose response-relationship. These associations appeared to be more pronounced in women.
In conclusion, women appear to have similar survival to men when accounting for device strategy. They have higher probabilities of recovery, but higher probabilities of device replacement and adverse events compared with men. Regarding these adverse events, women may be more susceptible to psychosocial risk factors than men. The results of this dissertation illustrate the importance of gender-sensitive research and suggest considering device strategy when studying gender differences in LVAD recipients. Further research is warranted to elucidate the role of specific psychosocial risk factors that lead to higher probabilities of adverse events, to intervene early and improve patient care in both, women and men
Anmerkung: Es handelt sich um die 2. überarbeitete Auflage der Dissertation.
1. Auflage siehe:
"https://ubt.opus.hbz-nrw.de/frontdoor/index/index/docId/2083".
Ausgangspunkt der politisch-ikonographischen Untersuchung, in deren
Zentrum zwei Staatsporträts König Maximilians II. von Bayern stehen, ist die Beobachtung, dass diese beiden Bildnisse grundsätzlich unterschiedliche Inszenierungsformen wählen. Das erste von Max Hailer gefertigte Werk zeigt Maximilian II. im vollen bayerischen Krönungsornat und greift eine tradierte Darstellungsweise im Staatsporträt auf. Es entstand zwei Jahre nach Maximilians II. Thronbesteigung und damit nach den revolutionären Unruhen der Jahre 1848/49 im Jahr 1850. Das zweite wurde von Joseph Bernhardt 1857 bis 1858 gemalt und im Jahr 1858 zum zehnjährigen Thronjubiläum des Monarchen erstmals präsentiert. Die Inszenierung ändert sich im zweiten Bildnis: Das bayerische Krönungsornat ist der Generalsuniform gewichen, ebenso weitere Details, die sich noch in der ersten Darstellung finden: Draperie und Wappen fehlen, der übliche bayerisch-königliche Thronsessel ist durch einen anderen ersetzt. In den Hintergrund gedrängt ist die Verfassung, immerhin seit 1818 staatliche Rechtsgrundlage des bayerischen Königreichs. Die beiden Staatsporträts Maximilians II. leiten offensichtlich von den Herrscherbildnissen im vollen bayerischen Krönungsornat seines Großvaters Maximilian I. und Vaters Ludwig I. über zu einer solchen in Uniform mit Krönungsmantel wie sie sich bei Napoleon III. und Friedrich Wilhelm IV. finden und wie sie sein Sohn Ludwig II. weiterführte. Es stellt sich somit die Frage, welche Faktoren zu diesem prägnanten Wandel in der Inszenierung Maximilians II. als König von Bayern führten. Die Arbeit geht der These nach, dass beide Darstellungen grundlegend auf eine reaktionäre, gegen die Revolution 1848/49 gerichtete Politik ausgelegt sind, wobei dieser reaktionäre Charakter in Maximilians II. Bildnis von 1858 noch eine Steigerung im Vergleich zu derjenigen von 1850 erfährt. Zudem wandelt sich die innenpolitisch-historische Ausrichtung des ersten Porträts bei der zweiten Darstellung des bayerischen Monarchen in eine außenpolitisch-progressive. Die Legitimation Maximilians II. begründet sich nicht mehr, wie bei ersterem, in der Geschichte und der Herrschaft der Wittelsbacher, sondern in seinen eigenen Errungenschaften und seiner eigenen Herrschaft. Dieser Wechsel der politischen Bildaussage fußt sowohl auf den politischen Veränderungen und Entwicklungen innerhalb und außerhalb Bayerns als auch auf der Entwicklung des Staatsporträts in der Mitte des 19. Jahrhunderts. Nach nur zehn Jahren wird so eine veränderte Botschaft über Maximilians II. Position und Machtanspruch ausgesendet.
Knowledge acquisition comprises various processes. Each of those has its dedicated research domain. Two examples are the relations between knowledge types and the influences of person-related variables. Furthermore, the transfer of knowledge is another crucial domain in educational research. I investigated these three processes through secondary analyses in this dissertation. Secondary analyses comply with the broadness of each field and yield the possibility of more general interpretations. The dissertation includes three meta-analyses: The first meta-analysis reports findings on the predictive relations between conceptual and procedural knowledge in mathematics in a cross-lagged panel model. The second meta-analysis focuses on the mediating effects of motivational constructs on the relationship between prior knowledge and knowledge after learning. The third meta-analysis deals with the effect of instructional methods in transfer interventions on knowledge transfer in school students. These three studies provide insights into the determinants and processes of knowledge acquisition and transfer. Knowledge types are interrelated; motivation mediates the relation between prior and later knowledge, and interventions influence knowledge transfer. The results are discussed by examining six key insights that build upon the three studies. Additionally, practical implications, as well as methodological and content-related ideas for further research, are provided.
There is no longer any doubt about the general effectiveness of psychotherapy. However, up to 40% of patients do not respond to treatment. Despite efforts to develop new treatments, overall effectiveness has not improved. Consequently, practice-oriented research has emerged to make research results more relevant to practitioners. Within this context, patient-focused research (PFR) focuses on the question of whether a particular treatment works for a specific patient. Finally, PFR gave rise to the precision mental health research movement that is trying to tailor treatments to individual patients by making data-driven and algorithm-based predictions. These predictions are intended to support therapists in their clinical decisions, such as the selection of treatment strategies and adaptation of treatment. The present work summarizes three studies that aim to generate different prediction models for treatment personalization that can be applied to practice. The goal of Study I was to develop a model for dropout prediction using data assessed prior to the first session (N = 2543). The usefulness of various machine learning (ML) algorithms and ensembles was assessed. The best model was an ensemble utilizing random forest and nearest neighbor modeling. It significantly outperformed generalized linear modeling, correctly identifying 63.4% of all cases and uncovering seven key predictors. The findings illustrated the potential of ML to enhance dropout predictions, but also highlighted that not all ML algorithms are equally suitable for this purpose. Study II utilized Study I’s findings to enhance the prediction of dropout rates. Data from the initial two sessions and observer ratings of therapist interventions and skills were employed to develop a model using an elastic net (EN) algorithm. The findings demonstrated that the model was significantly more effective at predicting dropout when using observer ratings with a Cohen’s d of up to .65 and more effective than the model in Study I, despite the smaller sample (N = 259). These results indicated that generating models could be improved by employing various data sources, which provide better foundations for model development. Finally, Study III generated a model to predict therapy outcome after a sudden gain (SG) in order to identify crucial predictors of the upward spiral. EN was used to generate the model using data from 794 cases that experienced a SG. A control group of the same size was also used to quantify and relativize the identified predictors by their general influence on therapy outcomes. The results indicated that there are seven key predictors that have varying effect sizes on therapy outcome, with Cohen's d ranging from 1.08 to 12.48. The findings suggested that a directive approach is more likely to lead to better outcomes after an SG, and that alliance ruptures can be effectively compensated for. However, these effects
were reversed in the control group. The results of the three studies are discussed regarding their usefulness to support clinical decision-making and their implications for the implementation of precision mental health.
The publication of statistical databases is subject to legal regulations, e.g. national statistical offices are only allowed to publish data if the data cannot be attributed to individuals. Achieving this privacy standard requires anonymizing the data prior to publication. However, data anonymization inevitably leads to a loss of information, which should be kept minimal. In this thesis, we analyze the anonymization method SAFE used in the German census in 2011 and we propose a novel integer programming-based anonymization method for nominal data.
In the first part of this thesis, we prove that a fundamental variant of the underlying SAFE optimization problem is NP-hard. This justifies the use of heuristic approaches for large data sets. In the second part, we propose a new anonymization method belonging to microaggregation methods, specifically designed for nominal data. This microaggregation method replaces rows in a microdata set with representative values to achieve k-anonymity, ensuring each data row is identical to at least k − 1 other rows. In addition to the overall dissimilarities of the data rows, the method accounts for errors in resulting frequency tables, which are of high interest for nominal data in practice. The method employs a typical two-step structure: initially partitioning the data set into clusters and subsequently replacing all cluster elements with representative values to achieve k-anonymity. For the partitioning step, we propose a column generation scheme followed by a heuristic to obtain an integer solution, which is based on the dual information. For the aggregation step, we present a mixed-integer problem formulation to find cluster representatives. To this end, we take errors in a subset of frequency tables into account. Furthermore, we show a reformulation of the problem to a minimum edge-weighted maximal clique problem in a multipartite graph, which allows for a different perspective on the problem. Moreover, we formulate a mixed-integer program, which combines the partitioning and the aggregation step and aims to minimize the sum of chi-squared errors in frequency tables.
Finally, an experimental study comparing the methods covered or developed in this work shows particularly strong results for the proposed method with respect to relative criteria, while SAFE shows its strength with respect to the maximum absolute error in frequency tables. We conclude that the inclusion of integer programming in the context of data anonymization is a promising direction to reduce the inevitable information loss inherent in anonymization, particularly for nominal data.
Building Fortress Europe Economic realism, China, and Europe’s investment screening mechanisms
(2023)
This thesis deals with the construction of investment screening mechanisms across the major economic powers in Europe and at the supranational level during the post-2015 period. The core puzzle at the heart of this research is how, in a traditional bastion of economic liberalism such as Europe, could a protectionist tool such as investment screening be erected in such a rapid manner. Within a few years, Europe went from a position of being highly welcoming towards foreign investment to increasingly implementing controls on it, with the focus on China. How are we to understand this shift in Europe? I posit that Europe’s increasingly protectionist shift on inward investment can be fruitfully understood using an economic realist approach, where the introduction of investment screening can be seen as part of a process of ‘balancing’ China’s economic rise and reasserting European competitiveness. China has moved from being the ‘workshop of the world’ to becoming an innovation-driven economy at the global technological frontier. As China has become more competitive, Europe, still a global economic leader, broadly situated at the technological frontier, has begun to sense a threat to its position, especially in the context of the fourth industrial revolution. A ‘balancing’ process has been set in motion, in which Europe seeks to halt and even reverse the narrowing competitiveness gap between it and China. The introduction of investment screening measures is part of this process.
In Luxemburg helfen externe Schulmediator*innen bei schulischen Konflikten. Die Anlaufstelle unterstützt bei drohenden Schulabbrüchen und Konflikten, die bei der Inklusion und Integration von Schüler*innen mit besonderem Förderbedarf oder mit Migrationshintergrund entstehen. Michèle Schilt sprach mit der Leiterin der Servicestelle, Lis De Pina, über die Arbeit der Schulmediation.
Emotionen gelten als Spiegelbild unserer persönlichen Bedürfnislage. Insbesondere in Konflikt- oder Mediationsgesprächen ist es demnach wichtig, nicht nur über den Moment zu sprechen, an dem ein Streit entstanden ist, sondern auch Bedürfnisse und Gefühle aufzudecken, die unser Handeln, Denken und Fühlen beeinflusst haben. Die folgenden Materialien zeigen, wie man als Lehrkraft Emotionen und Streit mit Grundschulkindern behandeln kann.
Sie haben eine spannende politische Diskussion in der Klasse. Das Gros Ihrer Schüler*innen ist wach, interessiert und engagiert. Alles läuft prima. Doch dann passiert's: Einer oder eine von ihnen stellt – absichtlich oder unreflektiert – eine extremistische oder verschwörungstheoretische Aussage in den Raum. Und nun?
Die Praxishefte Demokratische Schulkultur erscheinen halbjährlich und bieten Schulleitungen und Schulpersonal theoretische Grundlagen und praxisorientierte Anleitungen zur demokratiepädagogischen Schulentwicklung. Jedes Themenheft ist jeweils einer demokratiepädagogischen Bauform oder strategischen Frage der Schulentwicklung gewidmet. Die Praxishefte werden allen Luxemburger Schulen als Printausgabe zur Verfügung gestellt und online mit zusätzlichen Materialien und in französischer Fassung vorgehalten.
While humans find it easy to process visual information from the real world, machines struggle with this task due to the unstructured and complex nature of the information. Computer vision (CV) is the approach of artificial intelligence that attempts to automatically analyze, interpret, and extract such information. Recent CV approaches mainly use deep learning (DL) due to its very high accuracy. DL extracts useful features from unstructured images in a training dataset to use them for specific real-world tasks. However, DL requires a large number of parameters, computational power, and meaningful training data, which can be noisy, sparse, and incomplete for specific domains. Furthermore, DL tends to learn correlations from the training data that do not occur in reality, making DNNs poorly generalizable and error-prone.
Therefore, the field of visual transfer learning is seeking methods that are less dependent on training data and are thus more applicable in the constantly changing world. One idea is to enrich DL with prior knowledge. Knowledge graphs (KG) serve as a powerful tool for this purpose because they can formalize and organize prior knowledge based on an underlying ontological schema. They contain symbolic operations such as logic, rules, and reasoning, and can be created, adapted, and interpreted by domain experts. Due to the abstraction potential of symbols, KGs provide good prerequisites for generalizing their knowledge. To take advantage of the generalization properties of KG and the ability of DL to learn from large-scale unstructured data, attempts have long been made to combine explicit graph and implicit vector representations. However, with the recent development of knowledge graph embedding methods, where a graph is transferred into a vector space, new perspectives for a combination in vector space are opening up.
In this work, we attempt to combine prior knowledge from a KG with DL to improve visual transfer learning using the following steps: First, we explore the potential benefits of using prior knowledge encoded in a KG for DL-based visual transfer learning. Second, we investigate approaches that already combine KG and DL and create a categorization based on their general idea of knowledge integration. Third, we propose a novel method for the specific category of using the knowledge graph as a trainer, where a DNN is trained to adapt to a representation given by prior knowledge of a KG. Fourth, we extend the proposed method by extracting relevant context in the form of a subgraph of the KG to investigate the relationship between prior knowledge and performance on a specific CV task. In summary, this work provides deep insights into the combination of KG and DL, with the goal of making DL approaches more generalizable, more efficient, and more interpretable through prior knowledge.
Repeatedly encountering a stimulus biases the observer’s affective response and evaluation of the stimuli. Here we provide evidence for a causal link between mere exposure to fictitious news reports and subsequent voting behavior. In four pre-registered online experiments, participants browsed through newspaper webpages and were tacitly exposed to names of fictitious politicians. Exposure predicted voting behavior in a subsequent mock election, with a consistent preference for frequent over infrequent names, except when news items were decidedly negative. Follow-up analyses indicated that mere media presence fuels implicit personality theories regarding a candidate’s vigor in political contexts. News outlets should therefore be mindful to cover political candidates as evenly as possible.
The benefits of prosocial power motivation in leadership: Action orientation fosters a win-win
(2023)
Power motivation is considered a key component of successful leadership. Based on its dualistic nature, the need for power (nPower) can be expressed in a dominant or a prosocial manner. Whereas dominant motivation is associated with antisocial behaviors, prosocial motivation is characterized by more benevolent actions (e.g., helping, guiding). Prosocial enactment of the power motive has been linked to a wide range of beneficial outcomes, yet less has been investigated what determines a prosocial enactment of the power motive. According to Personality Systems Interactions (PSI) theory, action orientation (i.e., the ability to self-regulate affect) promotes prosocial enactment of the implicit power motive and initial findings within student samples verify this assumption. In the present study, we verified the role of action orientation as an antecedent for prosocial power enactment in a leadership sample (N = 383). Additionally, we found that leaders personally benefited from a prosocial enactment strategy. Results show that action orientation through prosocial power motivation leads to reduced power-related anxiety and, in turn, to greater leader well-being. The integration of motivation and self-regulation research reveals why leaders enact their power motive in a certain way and helps to understand how to establish a win-win situation for both followers and leaders.
We use a novel sea-ice lead climatology for the winters of 2002/03 to 2020/21 based on satellite observations with 1 km2 spatial resolution to identify predominant patterns in Arctic wintertime sea-ice leads. The causes for the observed spatial and temporal variabilities are investigated using ocean surface current velocities and eddy kinetic energies from an ocean model (Finite Element Sea Ice–Ice-Shelf–Ocean Model, FESOM) and winds from a regional climate model (CCLM) and ERA5 reanalysis, respectively. The presented investigation provides evidence for an influence of ocean bathymetry and associated currents on the mechanic weakening of sea ice and the accompanying occurrence of sea-ice leads with their characteristic spatial patterns. While the driving mechanisms for this observation are not yet understood in detail, the presented results can contribute to opening new hypotheses on ocean–sea-ice interactions. The individual contribution of ocean and atmosphere to regional lead dynamics is complex, and a deeper insight requires detailed mechanistic investigations in combination with considerations of coastal geometries. While the ocean influence on lead dynamics seems to act on a rather long-term scale (seasonal to interannual), the influence of wind appears to trigger sea-ice lead dynamics on shorter timescales of weeks to months and is largely controlled by individual events causing increased divergence. No significant pan-Arctic trends in wintertime leads can be observed.
The microbial enzyme alkaline phosphatase contributes to the removal of organic phosphorus compounds from wastewaters. To cope with regulatory threshold values for permitted maximum phosphor concentrations in treated wastewaters, a high activity of this enzyme in the biological treatment stage, e.g., the activated sludge process, is required. To investigate the reaction dynamics of this enzyme, to analyze substrate selectivities, and to identify potential inhibitors, the determination of enzyme kinetics is necessary. A method based on the application of the synthetic fluorogenic substrate 4-methylumbelliferyl phosphate is proven for soils, but not for activated sludges. Here, we adapt this procedure to the latter. The adapted method offers the additional benefit to determine inhibition kinetics. In contrast to conventional photometric assays, no particle removal, e.g., of sludge pellets, is required enabling the analysis of the whole sludge suspension as well as of specific sludge fractions. The high sensitivity of fluorescence detection allows the selection of a wide substrate concentration range for sound modeling of kinetic functions.
- Fluorescence array technique for fast and sensitive analysis of high sample numbers
- No need for particle separation – analysis of the whole (diluted) sludge suspension
- Simultaneous determination of standard and inhibition kinetics
The forensic application of phonetics relies on individuality in speech. In the forensic domain, individual patterns of verbal and paraverbal behavior are of interest which are readily available, measurable, consistent, and robust to disguise and to telephone transmission. This contribution is written from the perspective of the forensic phonetic practitioner and seeks to establish a more comprehensive concept of disfluency than previous studies have. A taxonomy of possible variables forming part of what can be termed disfluency behavior is outlined. It includes the “classical” fillers, but extends well beyond these, covering, among others, additional types of fillers as well as prolongations, but also the way in which fillers are combined with pauses. In the empirical section, the materials collected for an earlier study are re-examined and subjected to two different statistical procedures in an attempt to approach the issue of individuality. Recordings consist of several minutes of spontaneous speech by eight speakers on three different occasions. Beyond the established set of hesitation markers, additional aspects of disfluency behavior which fulfill the criteria outlined above are included in the analysis. The proportion of various types of disfluency markers is determined. Both statistical approaches suggest that these speakers can be distinguished at a level far above chance using the disfluency data. At the same time, the results show that it is difficult to pin down a single measure which characterizes the disfluency behavior of an individual speaker. The forensic implications of these findings are discussed.
Redox-driven biogeochemical cycling of iron plays an integral role in the complex process network of ecosystems, such as carbon cycling, the fate of nutrients and greenhouse gas emissions. We investigate Fe-(hydr)oxide (trans)formation pathways from rhyolitic tephra in acidic topsoils of South Patagonian Andosols to evaluate the ecological relevance of terrestrial iron cycling for this sensitive fjord ecosystem. Using bulk geochemical analyses combined with micrometer-scale-measurements on individual soil aggregates and tephra pumice, we document biotic and abiotic pathways of Fe released from the glassy tephra matrix and titanomagnetite phenocrysts. During successive redox cycles that are controlled by frequent hydrological perturbations under hyper-humid climate, (trans)formations of ferrihydrite-organic matter coprecipitates, maghemite and hematite are closely linked to tephra weathering and organic matter turnover. These Fe-(hydr)oxides nucleate after glass dissolution and complexation with organic ligands, through maghemitization or dissolution-(re)crystallization processes from metastable precursors. Ultimately, hematite represents the most thermodynamically stable Fe-(hydr)oxide formed under these conditions and physically accumulates at redox interfaces, whereas the ferrihydrite coprecipitates represent a so far underappreciated terrestrial source of bio-available iron for fjord bioproductivity. The insights into Fe-(hydr)oxide (trans)formation in Andosols have implications for a better understanding of biogeochemical cycling of iron in this unique Patagonian fjord ecosystem.
Regional climate models are a valuable tool for the study of the climate processes and climate change in polar regions, but the performance of the models has to be evaluated using experimental data. The regional climate model CCLM was used for simulations for the MOSAiC period with a horizontal resolution of 14 km (whole Arctic). CCLM was used in a forecast mode (nested in ERA5) and used a thermodynamic sea ice model. Sea ice concentration was taken from AMSR2 data (C15 run) and from a high-resolution data set (1 km) derived from MODIS data (C15MOD0 run). The model was evaluated using radiosonde data and data of different profiling systems with a focus on the winter period (November–April). The comparison with radiosonde data showed very good agreement for temperature, humidity, and wind. A cold bias was present in the ABL for November and December, which was smaller for the C15MOD0 run. In contrast, there was a warm bias for lower levels in March and April, which was smaller for the C15 run. The effects of different sea ice parameterizations were limited to heights below 300 m. High-resolution lidar and radar wind profiles as well as temperature and integrated water vapor (IWV) data from microwave radiometers were used for the comparison with CCLM for case studies, which included low-level jets. LIDAR wind profiles have many gaps, but represent a valuable data set for model evaluation. Comparisons with IWV and temperature data of microwave radiometers show very good agreement.
COVID-19 was a harsh reminder that diseases are an aspect of human existence and mortality. It was also a live experiment in the formation and alteration of disease-related attitudes. Not only are these attitudes relevant to an individual’s self-protective behavior, but they also seem to be associated with social and political attitudes more broadly. One of these attitudes is Social Darwinism, which holds that a pandemic benefits society by enabling nature “to weed out the weak”. In two countries (N = 300, N = 533), we introduce and provide evidence for the reliability, validity, and usefulness of the Disease-Related Social Darwinism (DRSD) Short Scale measuring this concept. Results indicate that DRSD is meaningful related to other central political attitudes like Social Dominance Orientation, Authoritarianism and neoliberalism. Importantly, the scale significantly predicted people’s protective behavior during the Pandemic over and above general social Darwinism. Moreover, it significantly predicted conservative attitudes, even after controlling for Social Dominance Orientation.
People are increasingly concerned about how meat affects the environment, human health, and animal welfare, yet eating and enjoying meat remains a norm. Unsurprisingly, many people are ambivalent about meat—evaluating it as both positive and negative. Here, we propose that meat-related conflict is multidimensional and depends on people’s dietary group: Omnivores’ felt ambivalence relates to multiple negative associations that oppose a predominantly positive attitude towards meat, and veg*ans’ ambivalence relates to various positive associations that oppose a predominantly negative attitude. A qualitative study (N = 235; German) revealed that omnivores and veg*ans experience meat-related ambivalence due to associations with animals, sociability, sustainability, health, and sensory experiences. To quantify felt ambivalence in these domains, we developed the Meat Ambivalence Questionnaire (MAQ). We validated the MAQ in four pre-registered studies using self-report and behavioral data (N = 3,485; German, UK, representative US). Both omnivores and veg*ans reported meat-related ambivalence, but with differences across domains and their consequences for meat consumption. Specifically, ambivalence was associated with less meat consumption in omnivores (especially sensory-/animal-based ambivalence) and more meat consumption in veg*ans (especially sensory-/socially-based ambivalence). Network analyses shed further light on the nomological net of the MAQ while controlling for a comprehensive set of determinants of meat consumption. By introducing the MAQ, we hope to provide researchers with a tool to better understand how ambivalence accompanies behavior change and maintenance.
The COVID-19 pandemic has affected schooling worldwide. In many places, schools closed for weeks or months, only part of the student body could be educated at any one time, or students were taught online. Previous research discloses the relevance of schooling for the development of cognitive abilities. We therefore compared the intelligence test performance of 424 German secondary school students in Grades 7 to 9 (42% female) tested after the first six months of the COVID-19 pandemic (i.e., 2020 sample) to the results of two highly comparable student samples tested in 2002 (n = 1506) and 2012 (n = 197). The results revealed substantially and significantly lower intelligence test scores in the 2020 sample than in both the 2002 and 2012 samples. We retested the 2020 sample after another full school year of COVID-19-affected schooling in 2021. We found mean-level changes of typical magnitude, with no signs of catching up to previous cohorts or further declines in cognitive performance. Perceived stress during the pandemic did not affect changes in intelligence test results between the two measurements.
Addition of Phosphogypsum to Fire-Resistant Plaster Panels:
A Physic–Mechanical Investigation
(2023)
Gypsum (GPS) has great potential for structural fire protection and is increasingly used in construction due to its high-water retention and purity. However, many researchers aim to improve its physical and mechanical properties by adding other organic or inorganic materials such as fibers, recycled GPS, and waste residues. This study used a novel method to add non-natural GPS from factory waste (phosphogypsum (PG)) as a secondary material for GPS. This paper proposes to mix these two materials to properly study the effect of PG on the physico-mechanical properties and fire performance of two Tunisian GPSs (GPS1 and GPS2). PG initially replaced GPS at 10, 20, 30, 40, and 50% weight percentage (mixing plan A). The PGs were then washed with distilled water several times. Two more mixing plans were run when the pH of the PG was equal to 2.4 (mixing plan B), and the pH was equal to 5 (mixing plan C). Finally, a comparative study was conducted on the compressive strength, flexural strength, density, water retention, and mass loss levels after 90 days of drying, before/after incineration of samples at 15, 30, 45, and 60 min. The results show that the mixture of GPS1 and 30% PG (mixing plan B) obtained the highest compressive strength (41.31%) and flexural strength (35.03%) compared to the reference sample. The addition of 10% PG to GPS1 (mixing plan A) improved fire resistance (33.33%) and the mass loss (17.10%) of the samples exposed to flame for 60 min compared to GPS2. Therefore, PG can be considered an excellent insulating material, which can increase physico-mechanical properties and fire resistance time of plaster under certain conditions.
Properties Evaluation of Composite Materials Based on Gypsum Plaster and Posidonia Oceanica Fibers
(2023)
Estimating the amount of material without significant losses at the end of hybrid casting is a problem addressed in this study. To minimize manufacturing costs and improve the accuracy of results, a correction factor (CF) was used in the formula to estimate the volume percent of the material in order to reduce material losses during the sample manufacturing stage, allowing for greater confidence between the approved blending plan and the results obtained. In this context, three material mixing schemes of different sizes and shapes (gypsum plaster, sand (0/2), gravel (2/4), and Posidonia oceanica fibers (PO)) were created to verify the efficiency of CF and more precisely study the physico-mechanical effects on the samples. The results show that the use of a CF can reduce mixing loss to almost 0%. The optimal compressive strength of the sample (S1B) with the lowest mixing loss was 7.50 MPa. Under optimal conditions, the addition of PO improves mix volume percent correction (negligible), flexural strength (5.45%), density (18%), and porosity (3.70%) compared with S1B. On the other hand, the addition of PO thermo-chemical treatment by NaOH increases the compressive strength (3.97%) compared with PO due to the removal of impurities on the fiber surface, as shown by scanning electron microscopy. We then determined the optimal mixture ratio (PO divided by a mixture of plaster, sand, and gravel), which equals 0.0321 because Tunisian gypsum contains small amounts of bassanite and calcite, as shown by the X-ray diffraction results.
This study scrutinizes press photographs published during the first 6 weeks of the Russian War in Ukraine, beginning February 24th, 2022. Its objective is to shed light on the emotions evoked in Internet-savvy audiences. This empirical research aims to contribute to the understanding of emotional media effects that shape attitudes and actions of ordinary citizens. Main research questions are: What kind of empathic reactions are observed during the Q-sort study? Which visual patterns are relevant for which emotional evaluations and attributions? The assumption is that the evaluations and attributions of empathy are not random, but follow specific patterns. The empathic reactions are based on visual patterns which, in turn, influence the type of empathic reaction. The identification of specific categories for visual and emotional reaction patterns are arrived at in different methodological processes. Visual pattern categories were developed inductively, using the art history method of iconography-iconology to identify six distinct types of visual motifs in a final sample of 33 war photographs. The overarching categories for empathic reactions—empty empathy, vicarious traumatization and witnessing—were applied deductively, building on E. Ann Kaplan's pivotal distinctions. The main result of this research are three novel categories that combine visual patterns with empathic reaction patterns. The labels for these categories are a direct result of the Q-factorial analysis, interpreted through the lense of iconography-iconology. An exploratory nine-scale forced-choice Q-sort study (Nstimuli = 33) was implemented, followed by self-report interviews with a total of 25 participants [F = 16 (64%), M = 9 (36%), Mage = 26.4 years]. Results from this exploratory research include motivational statements on the meanings of war photography from semi-structured post-sort-interviews. The major result of this study are three types of visual patterns (“factors”) that govern distinct empathic reactions in participants: Factor 1 is “veiled empathy” with highest empathy being attributed to photos showing victims whose corpses or faces were veiled. Additional features of “veiled empathy” are a strong anti-politician bias and a heightened awareness of potential visual manipulation. Factor 2 is “mirrored empathy” with highest empathy attributions to photos displaying human suffering openly. Factor 3 focused on the context. It showed a proclivity for documentary style photography. This pattern ranked photos without clear contextualization lower in empathy than those photos displaying the fully contextualized setting. To the best of our knowledge, no study has tested empathic reactions to war photography empirically. In this respect, the study is novel, but also exploratory. Findings like the three patterns of visual empathy might be helpful for photo selection processes in journalism, for political decision-making, for the promotion of relief efforts, and for coping strategies in civil society to deal with the potentially numbing or traumatizing visual legacy of the War in Ukraine.
Family firms play a crucial role in the DACH region (Germany, Austria, Switzerland). They are characterized by a long tradition, a strong connection to the region, and a well-established network. However, family firms also face challenges, especially in finding a suitable successor. Wealthy entrepreneurial families are increasingly opting to establish Single Family Offices (SFOs) as a solution to this challenge. An SFO takes on the management and protection of family wealth. Its goal is to secure and grow the wealth over generations. In Germany alone, there are an estimated 350 to 450 SFOs, with 70% of them being established after the year 2000. However, research on SFOs is still in its early stages, particularly regarding the role of SFOs as firm owners. This dissertation delves into an exploration of SFOs through four quantitative empirical studies. The first study provides a descriptive overview of 216 SFOs from the DACH-region. Findings reveal that SFOs exhibit a preference for investing in established companies and real estate. Notably, only about a third of SFOs engage in investments in start-ups. Moreover, SFOs as a group are heterogeneous. Categorizing them into three groups based on their relationship with the entrepreneurial family and the original family firm reveals significant differences in their asset allocation strategies. Subsequent studies in this dissertation leverage a hand-collected sample of 173 SFO-owned firms from the DACH region, meticulously matched with 684 family-owned firms from the same region. The second study focusing on financial performance indicates that SFO-owned firms tend to exhibit comparatively poorer financial performance than family-owned firms. However, when members of the SFO-owning family hold positions on the supervisory or executive board of the firm, there's a notable improvement. The third study, concerning cash holdings, reveals that SFO-owned firms maintain a higher cash holding ratio compared to family-owned firms. Notably, this effect is magnified when the SFO has divested its initial family firms. Lastly, the fourth study regarding capital structure highlights that SFO-owned firms tend to display a higher long-term debt ratio than family-owned firms. This suggests that SFO-owned firms operate within a trade-off theory framework, like private equity-owned firms. Furthermore, this effect is stronger for SFOs that sold their original family firm. The outcomes of this research are poised to provide entrepreneurial families with a practical guide for effectively managing and leveraging SFOs as a strategic long-term instrument for succession and investment planning.
Some of the largest firms in the DACH region (Germany, Austria, Switzerland) are (partially) owned by a foundation and/or a family office, such as Aldi, Bosch, or Rolex. Despite their growing importance, prior research neglected to analyze the impact of these intermediaries on the firms they own. This dissertation closes this research gap by contributing to a deeper understanding of two increasingly used family firm succession vehicles, through four empirical quantitative studies. The first study focuses on the heterogeneity in foundation-owned firms (FOFs) by applying a descriptive analysis to a sample of 169 German FOFs. The results indicate that the family as a central stakeholder in a family foundation fosters governance that promotes performance and growth. The second study examines the firm growth of 204 FOFs compared to matched non-FOFs from the DACH region. The findings suggest that FOFs grow significantly less in terms of sales but not with regard to employees. In addition, it seems that this negative effect is stronger for the upper than for the middle or lower quantiles of the growth distribution. Study three adopts an agency perspective and investigates the acquisition behavior within the group of 164 FOFs. The results reveal that firms with charitable foundations as owners are more likely to undertake acquisitions and acquire targets that are geographically and culturally more distant than firms with a family foundation as owner. At the same time, they favor target companies from the same or related industries. Finally, the fourth study scrutinizes the capital structure of firms owned by single family-offices (SFOs). Drawing on a hand-collected sample of 173 SFO-owned firms in the DACH region, the results show that SFO-owned firms display a higher long-term debt ratio than family-owned firms, indicating that SFO-owned firms follow trade-off theory, similar to private equity-owned firms. Additional analyses show that this effect is stronger for SFOs that sold their original family firm. In conclusion, the outcomes of this dissertation furnish valuable research contributions and offer practical insights for families navigating such intermediaries or succession vehicles in the long term.
Diese Dissertationsschrift befasst sich mit der Erforschung des motorischen Gedächtnisses. Wir gehen der Frage nach, ob sich dort Analogien zu im deklarativen Gedächtnis bekannten kontextuellen und inhibitorischen Effekten finden lassen.
Der erste von drei peer reviewed Artikeln setzt sich mit der generellen Bedeutung von externen Kontextmerkmalen für einen motorischen Gedächtnisabruf auseinander. Wir veränderten zwei verschiedene Sätze motorischer Sequenzen entlang einer hohen Zahl entsprechender Merkmale. Signifikant unterschiedliche Erinnerungsleistungen wiesen auf eine Kontextabhängigkeit motorischer Inhalte hin. Die Erinnerungsleistung variierte entlang der seriellen Output-Position. Bei einem Kontextwechsel blieb die Erinnerungsleistung über den Abrufverlauf nahezu stabil, bei Kontextbeibehaltung fiel diese schnell signifikant ab.
Beide weiteren peer reviewed Artikel wenden sich dann der Inhibition motorischer Sequenzen zu. Im zweiten Artikel begutachten wir drei Sätze motorischer Sequenzen, die wir mit verschiedenen Händen ausführen ließen, auf ein selektives gerichtetes Vergessen. Die Vergessen-Gruppe zeigte dies nur, wenn für Satz Zwei und Drei dieselbe Hand benutzt wurde und somit ein hohes Interferenzpotenzial zwischen diesen Listen bestand. War dieses im Vergleich niedrig, indem beide Sätze mit verschiedenen Händen auszuführen waren, trat kein selektives gerichtetes Vergessen auf. Das deutet auf kognitive Inhibition als wirkursächlichen Prozess.
Im dritten Artikel schließlich untersuchen wir Effekte willentlicher kognitiver Unterdrückung sowohl des Gedächtnisabrufs als auch des Ausführens in einer motorischen Adaptation des TNT (think/no-think) – Paradigmas (Anderson & Green, 2001). Waren die Sequenzen in Experiment 1 anfänglich stärker trainiert worden, so zeigten willentlich unterdrückte (no-think) motorische Repräsentationen eine deutliche Verlangsamung in deren Zugänglichkeit und tendenziell auch in der Ausführung, - im Vergleich zu Basisraten-Sequenzen. Waren die Sequenzen in Experiment 2 dagegen nur moderat trainiert, wurden diese auch schlechter erinnert und deutlich verlangsamt ausgeführt. Willentliche kognitive Unterdrückung kann motorische Gedächtnisrepräsentation und deren Ausführung beeinflussen.
Unsere drei Artikel bestätigen motorische Analogien bekannter Kontext- und Inhibitionseffekte im deklarativen Gedächtnis. Wir führen ein selektives gerichtetes Vergessen motorischer Inhalte eindeutig auf Inhibition zurück und bestätigen darüber hinaus Effekte der willentlichen Unterdrückung motorischer Gedächtnisrepräsentation.
Anmerkung: Es handelt sich um die 1. Auflage der Dissertation.
2. überarbeitete Auflage siehe:
"https://ubt.opus.hbz-nrw.de/frontdoor/index/index/docId/2166".
Ausgangspunkt der politisch-ikonographischen Untersuchung, in deren Zentrum zwei Staatsporträts König Maximilians II. von Bayern stehen, ist die Beobachtung, dass diese beiden Bildnisse grundsätzlich unterschiedliche Inszenierungsformen wählen. Das erste von Max Hailer gefertigte Werk zeigt Maximilian II. im vollen bayerischen Krönungsornat und greift eine tradierte Darstellungsweise im Staatsporträt auf. Es entstand zwei Jahre nach Maximilians II. Thronbesteigung und damit nach den revolutionären Unruhen der Jahre 1848/49 im Jahr 1850. Das zweite wurde von Joseph Bernhardt 1857 bis 1858 gemalt und im Jahr 1858 zum zehnjährigen Thronjubiläum des Monarchen erstmals präsentiert. Die Inszenierung ändert sich im zweiten Bildnis: Das bayerische Krönungsornat ist der Generalsuniform gewichen, ebenso weitere Details, die sich noch in der ersten Darstellung finden: Draperie und Wappen fehlen, der übliche bayerisch-königliche Thronsessel ist durch einen anderen ersetzt. In den Hintergrund gedrängt ist die Verfassung, immerhin seit 1818 staatliche Rechtsgrundlage des bayerischen Königreichs. Die beiden Staatsporträts Maximilians II. leiten offensichtlich von den Herrscherbildnissen im vollen bayerischen Krönungsornat seines Großvaters Maximilian I. und Vaters Ludwig I. über zu einer solchen in Uniform mit Krönungsmantel wie sie sich bei Napoleon III. und Friedrich Wilhelm IV. finden und wie sie sein Sohn Ludwig II. weiterführte. Es stellt sich somit die Frage, welche Faktoren zu diesem prägnanten Wandel in der Inszenierung Maximilians II. als König von Bayern führten. Die Arbeit geht der These nach, dass beide Darstellungen grundlegend auf eine reaktionäre, gegen die Revolution 1848/49 gerichtete Politik ausgelegt sind, wobei dieser reaktionäre Charakter in Maximilians II. Bildnis von 1858 noch eine Steigerung im Vergleich zu derjenigen von 1850 erfährt. Zudem wandelt sich die innenpolitisch-historische Ausrichtung des ersten Porträts bei der zweiten Darstellung des bayerischen Monarchen in eine außenpolitisch-progressive. Die Legitimation Maximilians II. begründet sich nicht mehr, wie bei ersterem, in der Geschichte und der Herrschaft der Wittelsbacher, sondern in seinen eigenen Errungenschaften und seiner eigenen Herrschaft. Dieser Wechsel der politischen Bildaussage fußt sowohl auf den politischen Veränderungen und Entwicklungen innerhalb und außerhalb Bayerns als auch auf der Entwicklung des Staatsporträts in der Mitte des 19. Jahrhunderts. Nach nur zehn Jahren wird so eine veränderte Botschaft über Maximilians II. Position und Machtanspruch ausgesendet.
Debatten führen nicht immer zu einem Konsens. Selbst die Vorlage von Beweisen bewirkt nicht immer eine Überzeugung der Gegenseite. Dies zeigt sich nicht nur in der Geschichte der Wissenschaften (vgl. Ludwik Fleck, Bruno Latour), sondern auch in der in unterschiedlichen Disziplinen geführten zeitgenössischen Debatte unter dem Label ‚science wars‘ zwischen einem Realismus und Konstruktivismus beziehungsweise Relativismus. Unterschiede in ihren Legitimierungen zeigen systematisch verschiedene Wirklichkeits- und Wahrheitsverständnisse, die sich aus den vom Seinsstandort der Perspektive abhängigen Grundannahmen konstituieren. Über einen wissenssoziologischen Zugriff wird es möglich die (sozio-)strukturlogische Konstitution von Perspektivität zu analysieren, die eine epistemologisch vorstrukturierte Revolvierung untereinander inkommensurabler Beiträge in der Debatte aufdeckt, was als Erklärung für ungelöste Debatten in Wissenschaft, Politik und Alltag überhaupt fungieren kann.
Die vorliegende Arbeit orientiert sich in ihrem Vorgehen an dem von Paul Boghossian veröffentlichten Werk ‚Angst vor der Wahrheit‘ als zeitgenössischen Vertreter eines Neuen Realismus. Hierbei werden zum einen den direkten Bezügen von Boghossian die Aussagen der kritisierten Perspektiven (v.a. Latour und Goodman) gegenübergestellt, als auch zum anderen weitere Spielarten eines Konstruktivismus (kognitionstheoretischer Konstruktivismus nach Maturana und Varela, soziologischer Konstruktivismus nach Berger und Luckmann, Wissenschaftssoziologie am Beispiel von Bloor und Latour, die Systemtheorie von Luhmann sowie postkonstruktivistische Positionen) in den Dimensionen ‚Wissensverständnis‘, ‚Subjektrelevanz‘ und ‚Einstellung zu einer naturalistischen Grundlage‘ vorgestellt. Es wird eine systematische und beidseitige Fehlinterpretation in der Debatte zwischen Realismus und Konstruktivismus sichtbar. Diese wird auf die Seinsgebundenheit von Perspektiven nach dem Verständnis einer mannheimschen Wissenssoziologie zurückgeführt. Anhand einer Rekonstruktion der Erkenntnistheorie des frühen Mannheims (1922: ‚Strukturanalyse der Erkenntnistheorie‘) wird die (sozio-)strukturlogische Konstitution erkenntnistheoretischer Elemente von Grundwissenschaften herausgearbeitet, wodurch denkstilgemäße Objektivierungen (und damit Wahrheitsverständnisse) unterschieden werden können. Diese Unterschiede erklären nicht nur die Inkommensurabilität von heterogenen Perspektiven in Debatten, sondern zeigen auf, dass das Aufeinandertreffen der Debattierenden vorstrukturiert sind. Der Ablauf einer Debatte ist soziostrukturell determiniert. Abschließend wird in der vorliegenden Arbeit diskutiert, inwiefern der verfahrenen Situation einer Debatte entgegengewirkt werden kann und auf welche Weise eine wissenssoziologische Analyse zu einem gegenseitigen Verständnis zwischen debattierenden Parteien beitragen kann.
Diese Dissertation beschäftigt sich mit der Fragestellung, ob und wie Intersektionalität als analytische Perspektive für literarische Texte eine nützliche Ergänzung für ethnisch geordnete Literaturfelder darstellt. Diese Fragestellung wird anhand der Analyse dreier zeitgenössischer chinesisch-kanadischer Romane untersucht.
In der Einleitung wird die Relevanz der Themenbereiche Intersektionalität und asiatisch-kanadische Literatur erörtert. Das darauffolgende Kapitel bietet einen historischen Überblick über die chinesisch-kanadische Einwanderung und geht detailliert auf die literarischen Produktionen ein. Es wird aufgezeigt, dass, obwohl kulturelle Güter auch zur Artikulation von Ungleichheitsverhältnissen aufgrund von zugeschriebener ethnischer Zugehörigkeit entstehen, ein Diversifizierungsbestreben innerhalb der literarischen Gemeinschaft von chinesisch-kanadischen Autor:innen identifiziert werden kann. Das dritte Kapitel widmet sich dem Begriff „Intersektionalität“ und stellt, nach einer historischen Einordnung des Konzeptes mit seinen Ursprüngen im Black Feminism, Intersektionalität als bindendes Element zwischen Postkolonialismus, Diversität und Empowerment dar – Konzepte, die für die Analyse (kanadischer) Literatur in dieser Dissertation von besonderer Relevanz sind. Anschließend wird die Rolle von Intersektionalität in der Literaturwissenschaft aufgegriffen. Die darauffolgenden exemplarischen Analysen von Kim Fus For Today I Am a Boy, Wayson Choys The Jade Peony und Yan Lis Lily in the Snow veranschaulichen die vorangegangen methodischen Überlegungen. Allen drei Romanen vorangestellt ist die Kontextualisierung des jeweiligen Werkes als chinesisch-kanadisch, aber auch bisher vorgenommene Überlegungen, die diese Einordnung infrage stellen. Nach einer Zusammenfassung des Inhalts folgt eine intersektionale Analyse auf der inhaltlichen Ebene, die in den familiären und weiteren sozialen Bereich unterteilt ist, da sich die Hierarchiemechanismen innerhalb dieser Bereiche unterscheiden oder gegenseitig verstärken, wie aus den Analysen hervorgeht. Anschließend wird die formale Analyse mit einem intersektionalen Schwerpunkt in einem separaten Unterkapitel näher beleuchtet. Ein drittes Unterkapitel widmet sich einem dem jeweiligen Roman spezifischen Aspekt, der im Zusammenhang mit einer intersektionalen Analyse von besonderer Relevanz ist. Die Arbeit schließt mit einem übergreifenden Fazit, welches die wichtigsten Ergebnisse aus der Analyse zusammenfasst und mit weiteren Überlegungen zu den Implikationen dieser Dissertation, vor allem im Hinblick auf sogenannte kanadische „master narratives“, die eine weitreichende, kontextuelle Relevanz für das Arbeiten mit literarischen Texten aufweisen und durch einen intersektionalen literarischen Ansatz in Zukunft gegebenenfalls gewinnbringend ergänzt werden können.
In recent years, the establishment of new makerspaces in Germany has increased significantly. The underlying phenomenon of the Maker Movement is a cultural and technological movement focused on making physical and digital products using open source principles, collaborative production, and individual empowerment. Because of its potential to democratize the innovation and production process, empower individuals and communities, and enable innovators to solve problems at the local level, the Maker Movement has received considerable attention in recent years. Despite numerous indicators, little is known about the phenomenon and its individual members, especially in Germany. Initial research suggests that the Maker Movement holds great potential for innovation and entrepreneurship. However, there is still a gap in understanding how Makers discover, evaluate and exploit entrepreneurial opportunities. Moreover, there is still controversy - both among policy makers and within the maker community itself - about the impact the maker movement has and can have on innovation and entrepreneurship in the future. This dissertation uses a mixed-methods approach to explore these questions. In addition to a quantitative analysis of maker characteristics, the results show that social impact, market size, and property rights have significant effects on the evaluation of entrepreneurial opportunities. The findings within this dissertation expand research in the field of the Maker Movement and offer multiple implications for practice. This dissertation provides the first quantitative data on makers in makerspaces in Germany, their characteristics and motivations. In particular, the relationship between the Maker Movement and entrepreneurship is explored in depth for the first time. This is complemented by the presentation of different identity profiles of the individuals involved. In this way, policy-makers can develop a better understanding of the movement, its personalities and values, and consider them in initiatives and formats.
This thesis deals with REITs, their capital structure and the effects on leverage that regulatory requirements might have. The data used results from a combination of Thomson Reuters data with hand-collected data regarding the REIT status, regulatory information and law variables. Overall, leverage is analysed across 20 countries in the years 2007 to 2018. Country specific data, manually extracted from yearly EPRA reportings, is merged with company data in order to analyse the influence of different REIT restrictions on a firm's leverage.
Observing statistically significant differences in means across NON-REITs and REITs, causes motivation for further investigations. My results show that variables beyond traditional capital structure determinants impact the leverage of REITs. I find that explicit restrictions on leverage and the distribution of profits have a significant effect on leverage decisions. This supports the notion that the restrictions from EPRA reportings are mandatory. I test for various combinations of regulatory variables that show both in isolation as well as in combination significant effects on leverage.
My main result is the following: Firms that operate under regulation that specifies a maximum leverage ratio, in addition to mandatory high dividend distributions, have on average lower leverage ratios. Further the existence of sanctions has a negative effect on REITs' leverage ratios, indicating that regulation is binding. The analysis clearly shows that traditional capital structure determinants are of second order relevance. This relationship highlights the impact on leverage and financing decisions caused by regulation. These effects are supported by further analysis. Results based on an event study show that REITs have statistically lower leverage ratios compared to NON-REITs. Based on a structural break model, the following effect becomes apparent: REITs increase their leverage ratios in years prior REIT status. As a consequence, the ex ante time frame is characterised by a bunker and adaption process, followed by the transformation in the event. Using an event study and a structural break model, the analysis highlights the dominance of country-specific regulation.
Striving for sustainable development by combating climate change and creating a more social world is one of the most pressing issues of our time. Growing legal requirements and customer expectations require also Mittelstand firms to address sustainability issues such as climate change. This dissertation contributes to a better understanding of sustainability in the Mittelstand context by examining different Mittelstand actors and the three dimensions of sustainability - social, economic, and environmental sustainability - in four quantitative studies. The first two studies focus on the social relevance and economic performance of hidden champions, a niche market leading subgroup of Mittelstand firms. At the regional level, the impact of 1,645 hidden champions located in Germany on various dimensions of regional development is examined. A higher concentration of hidden champions has a positive effect on regional employment, median income, and patents. At the firm level, analyses of a panel dataset of 4,677 German manufacturing firms, including 617 hidden champions, show that the latter have a higher return on assets than other Mittelstand firms. The following two chapters deal with environmental strategies and thus contribute to the exploration of the environmental dimension of sustainability. First, the consideration of climate aspects in investment decisions is compared using survey data from 468 European venture capital and private equity investors. While private equity firms respond to external stakeholders and portfolio performance and pursue an active ownership strategy, venture capital firms are motivated by product differentiation and make impact investments. Finally, based on survey data from 443 medium-sized manufacturing firms in Germany, 54% of which are family-owned, the impact of stakeholder pressures on their decarbonization strategies is analyzed. A distinction is made between symbolic (compensation of CO₂-emissions) and substantive decarbonization strategies (reduction of CO₂-emissions). Stakeholder pressures lead to a proactive pursuit of decarbonization strategies, with internal and external stakeholders varying in their influence on symbolic and substantial decarbonization strategies, and the relationship influenced by family ownership.
The German Mittelstand is closely linked to the success of the German economy. Mittelstand firms, thereof numerous Hidden Champions, significantly contribute to Germany’s economic performance, innovation, and export strength. However, the advancing digitalization poses complex challenges for Mittelstand firms. To benefit from the manifold opportunities offered by digital technologies and to defend or even expand existing market positions, Mittelstand firms must transform themselves and their business models. This dissertation uses quantitative methods and contributes to a deeper understanding of the distinct needs and influencing factors of the digital transformation of Mittelstand firms. The results of the empirical analyses of a unique database of 525 mid-sized German manufacturing firms, comprising both firm-related information and survey data, show that organizational capabilities and characteristics significantly influence the digital transformation of Mittelstand firms. The results support the assumption that dynamic capabilities promote the digital transformation of such firms and underline the important role of ownership structure, especially regarding family influence, for the digital transformation of the business model and the pursuit of growth goals with digitalization. In addition to the digital transformation of German Mittelstand firms, this dissertation examines the economic success and regional impact of Hidden Champions and hence, contributes to a better understanding of the Hidden Champion phenomenon. Using quantitative methods, it can be empirically proven that Hidden Champions outperform other mid-sized firms in financial terms and promote regional development. Consequently, the results of this dissertation provide valuable research contributions and offer various practical implications for firm managers and owners as well as policy makers.
Every action we perform, no matter how simple or complex, has a cognitive representation. It is commonly assumed that these are organized hierarchically. Thus, the representation of a complex action consists of multiple simpler actions. The representation of a simple action, in turn, consists of stimulus, response, and effect features. These are integrated into one representation upon the execution of an action and can be retrieved if a feature is repeated. Depending on whether retrieved features match or only partially match the current action episode, this might benefit or impair the execution of a subsequent action. This pattern of costs and benefits results in binding effects that indicate the strength of common representation between features. Binding effects occur also in more complex actions: Multiple simple actions seem to form representations on a higher level through the integration and retrieval of sequentially given responses, resulting in so-called response-response binding effects. This dissertation aimed to investigate what factors determine whether simple actions form more complex representations. The first line of research (Articles 1-3) focused on dissecting the internal structure of simple actions. Specifically, I investigated whether the spatial relation of stimuli, responses, or effects, that are part of two different simple actions, influenced whether these simple actions are represented as one more complex action. The second line of research (Articles 2, 4, and 5) investigated the role of context on the formation and strength of more complex action representations. Results suggest that spatial separation of responses as well as context might affect the strength of more complex action representations. In sum, findings help to specify assumptions on the structure of complex action representations. However, it may be important to distinguish factors that influence the strength and structure of action representations from factors that terminate action representations.
This thesis comprises of four research papers on the economics of education and industrial relations, which contribute to the field of empirical economic research. All of the corresponding papers focus on analysing how much time individuals spend on specific activities. The allocation of available time resources is a decision that individuals make throughout their lifetime. In this thesis, we consider individuals at different stages of their lives - students at school, university students, and dependent employees at the workplace.
Part I includes two research studies on student's behaviour in secondary and tertiary education.
Chapter 2 explores whether students who are relatively younger or older within the school year exhibit differential time allocation. Building on previous findings showing that relatively younger students perform worse in school, the study shows that relatively younger students are aware of their poor performance in school and feel more strain as a result. Nevertheless, there are no clear differences to be found in terms of time spent on homework, while relatively younger students spend more time watching television and less time on sports activities. Thus, the results suggest that the lower learning outcomes are not associated with different time allocations between school-related activities and non-school-related activities.
Chapter 3 analyses how individual ability and labour market prospects affect study behaviour. The theoretical modelling predicts that both determinants increase study effort. The empirical investigation is based on cross-sectional data from the National Educational Panel Study (NEPS) and includes thousands of students in Germany. The analyses show that more gifted students exhibit lower subjective effort levels and invest less time in self-study. In contrast, very good labour market prospects lead to more effort exerted by the student, both qualitatively and quantitatively. The potential endogeneity problem is taken into account by using regional unemployment data as an instrumental variable.
Part II includes two labour economic studies on determinants of overtime. Both studies belong to the field of industrial relations, as they focus on union membership on the one hand and the interplay of works councils and collective bargaining coverage on the other.
Chapter 4 shows that union members work less overtime than non-members do. The econometric approach takes the problem of unobserved heterogeneity into account; but provides no evidence that this issue affects the results. Different channels that could lead to this relationship are analysed by examining relevant subgroups separately. For example, this effect of union membership can also be observed in establishments with works councils and for workers who are very likely to be covered by collective bargaining agreements. The study concludes that the observed effect is due to the fact that union membership can protect workers from corresponding increased working time demands by employers.
Chapter 5 builds on previous studies showing a negative effect of works councils on overtime. In addition to co-determination by works councils at the firm level, collective bargaining coverage is an important factor in the German industrial relations system. Corresponding data was not available in the SOEP for quite some time. Therefore, the study uses recent SOEP data, which also contains information on collective bargaining coverage. A cross-sectional analysis is conducted to examine the effects of works councils in establishments with and without collective bargaining coverage. Similar to studies analysing other outcome variables, the results show that the effect of works councils exists only for employees covered by a collective bargaining agreement.
Computer simulation has become established in a two-fold way: As a tool for planning, analyzing, and optimizing complex systems but also as a method for the scientific instigation of theories and thus for the generation of knowledge. Generated results often serve as a basis for investment decisions, e.g., road construction and factory planning, or provide evidence for scientific theory-building processes. To ensure the generation of credible and reproducible results, it is indispensable to conduct systematic and methodologically sound simulation studies. A variety of procedure models exist that structure and predetermine the process of a study. As a result, experimenters are often required to repetitively but thoroughly carry out a large number of experiments. Moreover, the process is not sufficiently specified and many important design decisions still have to be made by the experimenter, which might result in an unintentional bias of the results.
To facilitate the conducting of simulation studies and to improve both replicability and reproducibility of the generated results, this thesis proposes a procedure model for carrying out Hypothesis-Driven Simulation Studies, an approach that assists the experimenter during the design, execution, and analysis of simulation experiments. In contrast to existing approaches, a formally specified hypothesis becomes the key element of the study so that each step of the study can be adapted and executed to directly contribute to the verification of the hypothesis. To this end, the FITS language is presented, which enables the specification of hypotheses as assumptions regarding the influence specific input values have on the observable behavior of the model. The proposed procedure model systematically designs relevant simulation experiments, runs, and iterations that must be executed to provide evidence for the verification of the hypothesis. Generated outputs are then aggregated for each defined performance measure to allow for the application of statistical hypothesis testing approaches. Hence, the proposed assistance only requires the experimenter to provide an executable simulation model and a corresponding hypothesis to conduct a sound simulation study. With respect to the implementation of the proposed assistance system, this thesis presents an abstract architecture and provides formal specifications of all required services.
To evaluate the concept of Hypothesis-Driven Simulation Studies, two case studies are presented from the manufacturing domain. The introduced approach is applied to a NetLogo simulation model of a four-tiered supply chain. Two scenarios as well as corresponding assumptions about the model behavior are presented to investigate conditions for the occurrence of the bullwhip effect. Starting from the formal specification of the hypothesis, each step of a Hypothesis-Driven Simulation Study is presented in detail, with specific design decisions outlined, and generated inter- mediate data as well as final results illustrated. With respect to the comparability of the results, a conventional simulation study is conducted which serves as reference data. The approach that is proposed in this thesis is beneficial for both practitioners and scientists. The presented assistance system allows for a more effortless and simplified execution of simulation experiments while the efficient generation of credible results is ensured.
The following dissertation contains three studies examining academic boredom development in five high-track German secondary schools (AVG-project data; Study 1: N = 1,432; Study 2: N = 1,861; Study 3: N = 1,428). The investigation period spanned 3.5 years, with four waves of measurement from grades 5 to 8 (T1: 5th grade, after transition to secondary school; T2: 5th grade, after mid-term evaluations; T3: 6th grade, after mid-term evaluations; T4: 8th grade, after mid-term evaluations). All three studies featured cross-sectional and longitudinal analyses, separating, and comparing the subject domains of mathematics and German.
Study 1 provided an investigation of academic boredom’s factorial structure alongside correlational and reciprocal relations of different forms of boredom and academic self-concept. Analyses included reciprocal effects models and latent correlation analyses. Results indicated separability of boredom intensity, boredom due to underchallenge and boredom due to overchallenge, as separate, correlated factors. Evidence for reciprocal relations between boredom and academic self-concept was limited.
Study 2 examined the effectiveness and efficacy of full-time ability grouping for as a boredom intervention directed at the intellectually gifted. Analyses included propensity score matching, and latent growth curve modelling. Results pointed to limited effectiveness and efficacy for full-time ability grouping regarding boredom reduction.
Study 3 explored gender differences in academic boredom development, mediated by academic interest, academic self-concept, and previous academic achievement. Analyses included measurement invariance testing, and multiple-indicator-multi-cause-models. Results showed one-sided gender differences, with boys reporting less favorable boredom development compared to girls, even beyond the inclusion of relevant mediators.
Findings from all three studies were embedded into the theoretical framework of control-value theory (Pekrun, 2006; 2019; Pekrun et al., 2023). Limitations, directions for future research, and practical implications were acknowledged and discussed.
Overall, this dissertation yielded important insights into boredom’s conceptual complexity. This concerned factorial structure, developmental trajectories, interrelations to other learning variables, individual differences, and domain specificities.
Keywords: Academic boredom, boredom intensity, boredom due to underchallenge, boredom due to overchallenge, ability grouping, gender differences, longitudinal data analysis, control-value theory
Startups are essential agents for the evolution of economies and the creative destruction of established market conditions for the benefit of a more effective and efficient economy. Their significance is manifested in their drive for innovation and technological advancements, their creation of new jobs, their contribution to economic growth, and their impact on increased competition and increased market efficiency. By reason of their attributes of newness and smallness, startups often experience a limitation in accessing external financial resources. Extant research on entrepreneurial finance examines the capital structure of startups, various funding tools, financing environments in certain regions, and investor selection criteria among other topics. My dissertation contributes to this research area by examining the becoming increasingly important funding instrument of venture debt. Prior research on venture debt only investigated the business model of venture debt, the concept of venture debt, the selection criteria of venture debt providers, and the role of patents in the venture debt provider’s selection process. Based on qualitative and quantitative methods, the dissertation outlines the emergence of venture debt in Europe as well as the impact of venture debt on startups to open up a better understanding of venture debt.
The results of the qualitative studies indicate that venture debt was formed based on a ‘Kirznerian’ entrepreneurial opportunity and venture debt impacts startups positive and negative in their development via different impact mechanisms.
Based on these results, the dissertation analyzes the empirical impact of venture debt on a startup’s ability to acquire additional financial resources as well as the role of the reputation of venture debt providers. The results suggest that venture debt increases the likelihood of acquiring additional financial resources via subsequent funding rounds and trade sales. In addition, a higher venture debt provider reputation increases the likelihood of acquiring additional financial resources via IPOs.
This cumulative thesis encompass three studies focusing on the Weddell Sea region in the Antarctic. The first study produces and evaluates a high quality data set of wind measurements for this region. The second study produces and evaluates a 15 year regional climate simulation for the Weddell Sea region. And the third study produces and evaluates a climatology of low level jets (LLJs) from the simulation data set. The evaluations were done in the attached three publications and the produced data sets are published online.
In 2015/2016, the RV Polarstern undertook an Antarctic expedition in the Weddell Sea. We operated a Doppler wind lidar on board during that time running different scan patterns. The resulting data was evaluated, corrected, processed and we derived horizontal wind speed and directions for vertical profiles with up to 2 km height. The measurements cover 38 days with a temporal resolution of 10-15 minutes. A comparisons with other radio sounding data showed only minor differences.
The resulting data set was used alongside other measurements to evaluate temperature and wind of simulation data. The simulation data was produced with the regional climate model CCLM for the period of 2002 to 2016 for the Weddell Sea region. Only smaller biases were found except for a strong warm bias during winter near the surface of the Antarctic Plateau. Thus we adapted the model setup and were able to remove the bias in a second simulation.
This new simulation data was then used to derive a climatology of low level jets (LLJs). Statistics of occurrence frequency, height and wind speed of LLJs for the Weddell Sea region are presented along other parameters. Another evaluation with measurements was also performed in the last study.
Do Personality Traits, Trust and Fairness Shape the Stock-Investing Decisions of an Individual?
(2023)
This thesis is comprised of three projects, all of which are fundamentally connected to the choices that individuals make about stock investments. Differences in stock market participation (SMP) across countries are large and difficult to explain. The second chapter focuses on differences between Germany (low SMP) and East Asian countries (mostly high SMP). The study hypothesis is that cultural differences regarding social preferences and attitudes towards inequality lead to different attitudes towards stock markets and subsequently to different SMPs. Using a large-scale survey, it is found that these factors can, indeed, explain a substantial amount of the country differences that other known factors (financial literacy, risk preferences, etc.) could not. This suggests that social preferences should be given a more central role in programs that aim to enhance SMP in countries like Germany. The third chapter documented the importance of trust as well as herding for stock ownership decisions. The findings show that trust as a general concept has no significant contribution to stock investment intention. A thorough examination of general trust elements reveals that in group and out-group trust have an impact on individual stock market investment. Higher out group trust directly influences a person's decision to invest in stocks, whereas higher in-group trust increases herding attitudes in stock investment decisions and thus can potentially increase the likelihood of stock investments as well. The last chapter investigates the significance of personality traits in stock investing and home bias in portfolio selection. Findings show that personality traits do indeed have a significant impact on stock investment and portfolio allocation decisions. Despite the fact that the magnitude and significance of characteristics differ between two groups of investors, inexperienced and experienced, conscientiousness and neuroticism play an important role in stock investments and preferences. Moreover, high conscientiousness scores increase stock investment desire and portfolio allocation to risky assets like stocks, discouraging home bias in asset allocation. Regarding neuroticism, a higher-level increases home bias in portfolio selection and decreases willingness to stock investment and portfolio share. Finally, when an investor has no prior experience with portfolio selection, patriotism generates home bias. For experienced investors, having a low neuroticism score and a high conscientiousness and openness score seemed to be a constant factor in deciding to invest in a well-diversified international portfolio
Intensiv diskutierte Aspekte der Politikwissenschaft heben zunehmend die Bedeutung von Strategiefähigkeit zur erfolgreichen Durchführung von Wahlkämpfen für Parteien hervor. Der Widerspruch der mit den Implikationen der modernen Mediengesellschaft eingehergehenden unterstellten Akteursfähigkeit der Parteien und ihrer kollektiven heterogenen Interessens- und Organisationsvielfalt bleibt dabei bestehen. Die Fokussierung der Parteien auf das Ziel der Stimmenmaximierung bringt unter den sich wandelnden Rahmenbedingungen Veränderungen der Binnenstrukturen mit sich. So diskutieren Parteienforscher seit Längerem die Notwendigkeit eines vierten Parteitypus als Nachfolger von Kirchheimers Volkspartei (1965). Verschiedene dieser Ansätze berücksichtigen primär die Wahlkampffokussierung der Parteien, während andere vor allem auf den gesteigerten Strategiebedarf abzielen. Auch die Wechselwirkungen mit den Erfordernissen der Mediengesellschaft sowie Auswirkungen des gesellschaftlichen Wandels stehen im Vordergrund zahlreicher Untersuchungen. Die Arbeit von Uwe Jun (2004), der mit dem Modell der professionalisierten Medienkommunikationspartei auch die organisatorischen und programmatischen Transformationsaspekte des Parteiwandels beleuchtet, liefert einen bemerkenswerten Beitrag zur Party-Change-Debatte und bietet durch die angeschlossene vergleichende exemplarische Fallstudie eine praxisnahe Einordnung. Die geringe empirische Relevanz, die Jun seinem Parteityp anhand der Untersuchung von SPD und New Labor zwischen 1995 und 2005 bestätigt, soll in dieser Arbeit versucht werden zu relativieren, in dem der Parteiwandel der deutschen Großparteien seit der Wiedervereinigung durch die Untersuchung ihrer Wahlkampffähigkeit aufgezeigt wird. Anhand eines längsschnittlichen Vergleiches der Bundestagswahlkämpfe von SPD und CDU zwischen 1990 und 2013 soll die Plausibilität dieses vierten Parteitypus überprüft werden. Hierdurch soll die Entwicklung der Strategie- und Wahlkampffähigkeit beider Großparteien in den Bundestagswahlkämpfen seit 1990 untersucht und die Ergebnisse miteinander verglichen und in Bezug auf den Parteiwandel eingeordnet werden.
Dass sich Parteien genau wie ihre gesellschaftliche und politische Umwelt im Wandel befinden, ist nicht zu bestreiten und seit Langem viel diskutierter Gegenstand der Parteienforschung. „Niedergangsdiskussion“, Mitgliederschwund, Nicht- und Wechselwähler, Politik- und Parteienverdrossenheit, Kartellisierung und Institutionalisierung von Parteien sind nur einige der in diesem Kontext geläufigen Schlagwörter. Prozesse der Individualisierung, Globalisierung und Mediatisierung führen zu veränderten Rahmenbedingungen, unter denen Parteien sich behaupten müssen. Diese Veränderungen in der äußeren Umwelt wirken sich nachhaltig auf das parteipolitische Binnenleben, auf Organisationsstrukturen und Programmatik aus. Die Parteienforschung hat daher schon vor zwanzig Jahren begonnen, ein typologisches Nachfolgemodell der Volkspartei zu diskutieren, das diesen Wandel berücksichtigt. Verschiedene typologische Konstruktionen von z. B. Panebianco (1988), Katz und Mair (1995) oder von Beyme erfassen (2000) wichtige Facetten des Strukturwandels politischer Parteien und stellen mehrheitlich plausible typologische Konzepte vor, die die Parteien in ihrem Streben nach Wählerstimmen und Regierungsmacht zutreffend charakterisieren. Die Parteienforschung stimmt bezüglich des Endes der Volksparteiära mehrheitlich überein. Bezüglich der Nachfolge konnte sich unter den neueren vorgeschlagenen Typen jedoch kein vierter Typ als verbindliches Leitmodell etablieren. Bei genauerer Betrachtung weichen die in den verschiedenen Ansätzen für einen vierten Parteitypen hervorgehobenen Merkmale (namentlich Professionalisierung des Parteiapparates, die Berufspolitikerdominanz, Verstaatlichung und Kartellbildung sowie die Fixierung auf die Medien) wenig von jüngeren Modellvorschlägen ab und bedürfen daher mehr einer Ergänzung. Die in der Regel mehrdimensionalen entwicklungstypologischen Verlaufstypen setzten seit den 1980er Jahren unterschiedliche Schwerpunkte und warten mit vielen Vorschlägen der Einordnung auf. Einer der jüngsten Ansätze von Uwe Jun aus dem Jahr 2004, der das typologische Konzept der professionalisierten Medienkommunikationspartei einführt, macht deutlich, dass die Diskussion um Gestalt und Ausprägungen des vierten Parteityps noch in vollem Gang und für weitere Vorschläge offen ist – der „richtige“ Typ also noch nicht gefunden wurde. Jun bleibt in seiner Untersuchung den zentralen Transformationsleitfragen nach der Ausgestaltung der Parteiorganisation, der ideologisch-programmatischen Orientierung und der strategisch-elektoralen Wählerorientierung verhaftet und setzt diese Elemente in den Fokus sich wandelnder Kommunikationsstrategien. Die bisher in parteitypologischen Arbeiten mitunter vernachlässigte Komponente der strukturellen Strategiefähigkeit als Grundlage zur Entwicklung ebensolcher Reaktionsstrategien wird bei Jun angestoßen und soll in dieser Arbeit aufgegriffen und vertieft werden.
Der aktuellen Partychange-Diskussion zum Trotz scheint die Annahme, dass Parteien, die sich verstärkt der Handlungslogik der Massenmedien unterwerfen, deren strategischen Anforderungen durch interne Adaptionsverfahren auch dauerhaft gerecht zu werden vermögen, nicht immer zutreffend. Die Veränderungen der Kommunikationsstrategien als Reaktion auf gesamtgesellschaftliche Wandlungsprozesse stehen zwar im Zentrum der Professionalisierungsbemühungen der politischen Akteure, bleiben aber in ihrer Wirkung eingeschränkt. Wenngleich das Wissen in den Parteien um die Notwendigkeiten (medialer) Strategiefähigkeit besteht und die Parteien hierauf mit Professionalisierung, organisatorischen und programmatischen Anpassungsleistungen und der Herausbildung strategischer Zentren reagieren, so ist mediengerechtes strategisches Agieren noch lange keine natürliche Kernkompetenz der Parteien. Vor allem in Wahlkampfzeiten, die aufgrund abnehmender Parteibindungen und zunehmender Wählervolatilität für die Parteien zum eigentlich zentralen Moment der Parteiendemokratie werden, wird mediengerechtes Handeln zum wesentlichen Erfolgsfaktor. Strategiefähigkeit wird hierbei zur entscheidenden Voraussetzung und scheint zudem in diesen Phasen von den Parteien erfolgreicher umgesetzt zu werden als im normalen politischen Alltag. Die wahlstrategische Komponente findet in Juns typologischer Konstruktion wenig Beachtung und soll in dieser Arbeit daher als ergänzendes Element hinzugefügt werden. Arbeitshypothese Die beiden deutschen Großparteien berufen sich auf unterschiedliche Entstehungsgeschichten, die sich bis in die Gegenwart auf die Mitglieder-, Issue- und Organisationsstrukturen von SPD und CDU auswirken und die Parteien in ihren Anpassungsleistungen an die sich wandelnde Gesellschaft beeinflussen. Beide Parteien versuchen, auf die veränderten sozialen und politischen Rahmenbedingungen und den daraus resultierenden Bedeutungszuwachs von politischer Kommunikationsplanung mit einem erhöhten Maß an Strategiefähigkeit und kommunikativer Kompetenz zu reagieren. Diese Entwicklung tritt seit der deutschen Wiedervereinigung umso stärker in Augenschein, als dass nach 1990 die Bindekraft der Volksparteien nochmals nachließ, sodass die Parteien sich zunehmend gezwungen sehen, die „lose verkoppelten Anarchien“ in wahlstrategische Medienkommunikationsparteien zu transformieren. Diesen vierten Parteityp kennzeichnet vor allem die zunehmende Bemühung um Strategiefähigkeit, die mittels Organisationsstrukturen und programmatischer Anpassungsleistungen die Effizienz der elektoralen Ausrichtung verbessern soll. Insgesamt geht die Party-Change-Forschung davon aus, dass die Parteien sich zunehmend angleichen. Dies gilt es in dieser Studie zu überprüfen. Unter Berücksichtigung unterschiedlicher Entwicklungspfade kann vermutet werden, dass auch die Transformationsprozesse bei SPD und CDU in unterschiedlicher Weise verlaufen. Wenngleich die SPD über einen höheren Strategiebedarf und die größere Innovationsbereitschaft zu verfügen scheint, werden auf Seiten der Union potentiell strategiefähigere Strukturen vermutet, die die erfolgreiche Umsetzung von Wahlkampfstrategien erleichtern. Die historische Entwicklung und der Aspekt der Historizität spielen in diesem Kontext eine Rolle.
Zusätzlich spielen individuelle Führungspersönlichkeiten eine zentrale Rolle in innerparteilichen Transformationsprozessen, welche für die Ausprägung strategiefähiger Strukturen oftmals von größerer Bedeutung sind als institutionalisierte Strukturen. Im Vordergrund steht die Untersuchung des Parteiwandels anhand der Veränderung der Kommunikationsstrategien der Parteien im Allgemeinen sowie der Strategiefähigkeit in Wahlkämpfen im Besonderen, da diese als zentrale Merkmale für den vierten Parteityp in Anlehnung an die Professionelle Medienkommunikationspartei (Jun 2004) gewertet werden sollen. Strategiefähigkeit soll dabei anhand der Kriterien des Umgangs der Parteien mit Programmatik, Organisation und externen Einflussfaktoren in Wahlkämpfen operationalisiert werden. Die Analyse untersucht sowohl das Handeln einzelner Personen wie auch die Rolle der Partei als Gesamtorganisation. Die Arbeit besteht aus zehn Kapiteln und gliedert sich in zwei Blöcke: einen theoretisch konzeptionellen Teil, der die in der Perspektive dieser Arbeit zentralen Grundlagen und Rahmenbedingungen zusammenführt sowie die sich daran anschließende Untersuchung der Konzeption und Implementation von Kommunikationskampagnen im Wahlkampf seit 1990. Das aktuell in die politikwissenschaftliche Diskussion eingebrachte Feld der politischen Strategiefähigkeit (Raschke/Tils 2007) wird in ausführlicher theoretischer Grundlegung bisher zwar mit den Implikationen der Medienkommunikation und damit einhergehend auch den organisatorischen und programmatischen Strukturmerkmalen der Parteien verknüpft, diese erfolgte allerdings oft ohne vertiefte Berücksichtigung des Parteiwandels. Dies soll in diesem Beitrag daher versucht werden. Der Diskursanalyse des Strategiebegriffes in Wahlkampfsituationen folgt die detaillierte Darstellung der drei Operationalisierungsparameter, die in die Festlegung des Parteityps münden. Die Diskussion idealtypischer Wahlkampfmodelle als theoretischer Bezugsrahmen für die Bewertung der Wahlkampagnen ergänzt den theoretisch-konzeptionellen Bezugsrahmen. Die insgesamt in der Literatur in ihren Ausführungen oftmals normativ gestalteten Darstellungen idealtypischer politischer Strategie sollen im letzten Teil der Arbeit auf ihre Umsetzbarkeit im parteipolitischen Alltag überprüft werden und dies nicht nur anhand einzelner, mit einander nicht in Zusammenhang stehender Ereignisse, sondern anhand der sich periodisch unter vergleichbaren Bedingungen wiederholenden Wahlkämpfe. Dafür werden die jeweiligen Ausgangs- und Rahmenbedingungen der einzelnen Wahlkämpfe sowie die zuvor dargelegten Elemente professionalisierter Wahlkampagnen für die Wahlkampagnen von SPD und CDU seit 1990 dargestellt. Aus diesen Gegenüberstellungen soll im Anschluss der längsschnittliche Vergleich der Strategiefähigkeit und Kommunikationskompetenz von SPD und CDU abgeleitet werden
Das Ziel dynamischer Mikrosimulationen ist es, die Entwicklung von Systemen über das Verhalten der einzelnen enthaltenen Bestandteile zu simulieren, um umfassende szenariobasierte Analysen zu ermöglichen. Im Bereich der Wirtschafts- und Sozialwissenschaften wird der Fokus üblicherweise auf Populationen bestehend aus Personen und Haushalten gelegt. Da politische und wirtschaftliche Entscheidungsprozesse meist auf lokaler Ebene getroffen werden, bedarf es zudem kleinräumiger Informationen, um gezielte Handlungsempfehlungen ableiten zu können. Das stellt Forschende wiederum vor große Herausforderungen im Erstellungsprozess regionalisierter Simulationsmodelle. Dieser Prozess reicht von der Generierung geeigneter Ausgangsdatensätze über die Erfassung und Umsetzung der dynamischen Komponenten bis hin zur Auswertung der Ergebnisse und Quantifizierung von Unsicherheiten. Im Rahmen dieser Arbeit werden ausgewählte Komponenten, die für regionalisierte Mikrosimulationen von besonderer Relevanz sind, beschrieben und systematisch analysiert.
Zunächst werden in Kapitel 2 theoretische und methodische Aspekte von Mikrosimulationen vorgestellt, um einen umfassenden Überblick über verschiedene Arten und Möglichkeiten der Umsetzung dynamischer Modellierungen zu geben. Im Fokus stehen dabei die Grundlagen der Erfassung und Simulation von Zuständen und Zustandsänderungen sowie die damit verbundenen strukturellen Aspekte im Simulationsprozess.
Sowohl für die Simulation von Zustandsänderungen als auch für die Erweiterung der Datenbasis werden primär logistische Regressionsmodelle zur Erfassung und anschließenden wahrscheinlichkeitsbasierten Vorhersage der Bevölkerungsstrukturen auf Mikroebene herangezogen. Die Schätzung beruht insbesondere auf Stichprobendaten, die in der Regel neben einem eingeschränktem Stichprobenumfang keine oder nur unzureichende regionale Differenzierungen zulassen. Daher können bei der Vorhersage von Wahrscheinlichkeiten erhebliche Differenzen zu bekannten Totalwerten entstehen. Um eine Harmonisierung mit den Totalwerten zu erhalten, lassen sich Methoden zur Anpassung von Wahrscheinlichkeiten – sogenannte Alignmentmethoden – anwenden. In der Literatur werden zwar unterschiedliche Möglichkeiten beschrieben, über die Auswirkungen dieser Verfahren auf die Güte der Modelle ist jedoch kaum etwas bekannt. Zur Beurteilung verschiedener Techniken werden diese im Rahmen von Kapitel 3 in umfassenden Simulationsstudien unter verschiedenen Szenarien umgesetzt. Hierbei kann gezeigt werden, dass durch die Einbindung zusätzlicher Informationen im Modellierungsprozess deutliche Verbesserungen sowohl bei der Schätzung der Parameter als auch bei der Vorhersage der Wahrscheinlichkeiten erzielt werden können. Zudem lassen sich dadurch auch bei fehlenden regionalen Identifikatoren in den Modellierungsdaten kleinräumige Wahrscheinlichkeiten erzeugen. Insbesondere die Maximierung der Likelihood des zugrundeliegenden Regressionsmodells unter der Nebenbedingung, dass die bekannten Totalwerte eingehalten werden, weist in allen Simulationsstudien überaus gute Ergebnisse auf.
Als eine der einflussreichsten Komponenten in regionalisierten Mikrosimulationen erweist sich die Umsetzung regionaler Mobilität. Gleichzeitig finden Wanderungen in vielen Mikrosimulationsmodellen keine oder nur unzureichende Beachtung. Durch den unmittelbaren Einfluss auf die gesamte Bevölkerungsstruktur führt ein Ignorieren jedoch bereits bei einem kurzen Simulationshorizont zu starken Verzerrungen. Während für globale Modelle die Integration von Wanderungsbewegungen über Landesgrenzen ausreicht, müssen in regionalisierten Modellen auch Binnenwanderungsbewegungen möglichst umfassend nachgebildet werden. Zu diesem Zweck werden in Kapitel 4 Konzepte für Wanderungsmodule erstellt, die zum einen eine unabhängige Simulation auf regionalen Subpopulationen und zum anderen eine umfassende Nachbildung von Wanderungsbewegungen innerhalb der gesamten Population zulassen. Um eine Berücksichtigung von Haushaltsstrukturen zu ermöglichen und die Plausibilität der Daten zu gewährleisten, wird ein Algorithmus zur Kalibrierung von Haushaltswahrscheinlichkeiten vorgeschlagen, der die Einhaltung von Benchmarks auf Individualebene ermöglicht. Über die retrospektive Evaluation der simulierten Migrationsbewegungen wird die Funktionalität der Wanderdungskonzepte verdeutlicht. Darüber hinaus werden über die Fortschreibung der Population in zukünftige Perioden divergente Entwicklungen der Einwohnerzahlen durch verschiedene Konzepte der Wanderungen analysiert.
Eine besondere Herausforderung in dynamischen Mikrosimulationen stellt die Erfassung von Unsicherheiten dar. Durch die Komplexität der gesamten Struktur und die Heterogenität der Komponenten ist die Anwendung klassischer Methoden zur Messung von Unsicherheiten oft nicht mehr möglich. Zur Quantifizierung verschiedener Einflussfaktoren werden in Kapitel 5 varianzbasierte Sensitivitätsanalysen vorgeschlagen, die aufgrund ihrer enormen Flexibilität auch direkte Vergleiche zwischen unterschiedlichsten Komponenten ermöglichen. Dabei erweisen sich Sensitivitätsanalysen nicht nur für die Erfassung von Unsicherheiten, sondern auch für die direkte Analyse verschiedener Szenarien, insbesondere zur Evaluation gemeinsamer Effekte, als überaus geeignet. In Simulationsstudien wird die Anwendung im konkreten Kontext dynamischer Modelle veranschaulicht. Dadurch wird deutlich, dass zum einen große Unterschiede hinsichtlich verschiedener Zielwerte und Simulationsperioden auftreten, zum anderen aber auch immer der Grad an regionaler Differenzierung berücksichtigt werden muss.
Kapitel 6 fasst die Erkenntnisse der vorliegenden Arbeit zusammen und gibt einen Ausblick auf zukünftige Forschungspotentiale.
Even though proper research on Cauchy transforms has been done, there are still a lot of open questions. For example, in the case of representation theorems, i.e. the question when a function can be represented as a Cauchy transform, there is 'still no completely satisfactory answer' ([9], p. 84). There are characterizations for measures on the circle as presented in the monograph [7] and for general compactly supported measures on the complex plane as presented in [27]. However, there seems to exist no systematic treatise of the Cauchy transform as an operator on $L_p$ spaces and weighted $L_p$ spaces on the real axis.
This is the point where this thesis draws on and we are interested in developing several characterizations for the representability of a function by Cauchy transforms of $L_p$ functions. Moreover, we will attack the issue of integrability of Cauchy transforms of functions and measures, a topic which is only partly explored (see [43]). We will develop different approaches involving Fourier transforms and potential theory and investigate into sufficient conditions and characterizations.
For our purposes, we shall need some notation and the concept of Hardy spaces which will be part of the preliminary Chapter 1. Moreover, we introduce Fourier transforms and their complex analogue, namely Fourier-Laplace transforms. This will be of extraordinary usage due to the close connection of Cauchy and Fourier(-Laplace) transforms.
In the second chapter we shall begin our research with a discussion of the Cauchy transformation on the classical (unweighted) $L_p$ spaces. Therefore, we start with the boundary behavior of Cauchy transforms including an adapted version of the Sokhotski-Plemelj formula. This result will turn out helpful for the determination of the image of the Cauchy transformation under $L_p(\R)$ for $p\in(1,\infty).$ The cases $p=1$ and $p=\infty$ are playing special roles here which justifies a treatise in separate sections. For $p=1$ we will involve the real Hardy space $H_{1}(\R)$ whereas the case $p=\infty$ shall be attacked by an approach incorporating intersections of Hardy spaces and certain subspaces of $L_{\infty}(\R).$
The third chapter prepares ourselves for the study of the Cauchy transformation on subspaces of $L_{p}(\R).$ We shall give a short overview of the basic facts about Cauchy transforms of measures and then proceed to Cauchy transforms of functions with support in a closed set $X\subset\R.$ Our goal is to build up the main theory on which we can fall back in the subsequent chapters.
The fourth chapter deals with Cauchy transforms of functions and measures supported by an unbounded interval which is not the entire real axis. For convenience we restrict ourselves to the interval $[0,\infty).$ Bringing once again the Fourier-Laplace transform into play, we deduce complex characterizations for the Cauchy transforms of functions in $L_{2}(0,\infty).$ Moreover, we analyze the behavior of Cauchy transform on several half-planes and shall use these results for a fairly general geometric characterization. In the second section of this chapter, we focus on Cauchy transforms of measures with support in $[0,\infty).$ In this context, we shall derive a reconstruction formula for these Cauchy transforms holding under pretty general conditions as well as results on the behaviur on the left half-plane. We close this chapter by rather technical real-type conditions and characterizations for Cauchy transforms of functions in $L_p(0,\infty)$ basing on an approach in [82].
The most common case of Cauchy transforms, those of compactly supported functions or measures, is the subject of Chapter 5. After complex and geometric characterizations originating from similar ideas as in the fourth chapter, we adapt a functional-analytic approach in [27] to special measures, namely those with densities to a given complex measure $\mu.$ The chapter is closed with a study of the Cauchy transformation on weighted $L_p$ spaces. Here, we choose an ansatz through the finite Hilbert transform on $(-1,1).$
The sixth chapter is devoted to the issue of integrability of Cauchy transforms. Since this topic has no comprehensive treatise in literature yet, we start with an introduction of weighted Bergman spaces and general results on the interaction of the Cauchy transformation in these spaces. Afterwards, we combine the theory of Zen spaces with Cauchy transforms by using once again their connection with Fourier transforms. Here, we shall encounter general Paley-Wiener theorems of the recent past. Lastly, we attack the issue of integrability of Cauchy transforms by means of potential theory. Therefore, we derive a Fourier integral formula for the logarithmic energy in one and multiple dimensions and give applications to Fourier and hence Cauchy transforms.
Two appendices are annexed to this thesis. The first one covers important definitions and results from measure theory with a special focus on complex measures. The second appendix contains Cauchy transforms of frequently used measures and functions with detailed calculations.
Die Dissertation beschäftigt sich mit einer neuartigen Art von Branch-and-Bound Algorithmen, deren Unterschied zu klassischen Branch-and-Bound Algorithmen darin besteht, dass
das Branching durch die Addition von nicht-negativen Straftermen zur Zielfunktion erfolgt
anstatt durch das Hinzufügen weiterer Nebenbedingungen. Die Arbeit zeigt die theoretische Korrektheit des Algorithmusprinzips für verschiedene allgemeine Klassen von Problemen und evaluiert die Methode für verschiedene konkrete Problemklassen. Für diese Problemklassen, genauer Monotone und Nicht-Monotone Gemischtganzzahlige Lineare Komplementaritätsprobleme und Gemischtganzzahlige Lineare Probleme, präsentiert die Arbeit
verschiedene problemspezifische Verbesserungsmöglichkeiten und evaluiert diese numerisch.
Weiterhin vergleicht die Arbeit die neue Methode mit verschiedenen Benchmark-Methoden
mit größtenteils guten Ergebnissen und gibt einen Ausblick auf weitere Anwendungsgebiete
und zu beantwortende Forschungsfragen.
Allocating scarce resources efficiently is a major task in mechanism design. One of the most fundamental problems in mechanism design theory is the problem of selling a single indivisible item to bidders with private valuations for the item. In this setting, the classic Vickrey auction of~\citet{vickrey1961} describes a simple mechanism to implement a social welfare maximizing allocation.
The Vickrey auction for a single item asks every buyer to report its valuation and allocates the item to the highest bidder for a price of the second highest bid. This auction features some desirable properties, e.g., buyers cannot benefit from misreporting their true value for the item (incentive compatibility) and the auction can be executed in polynomial time.
However, when there is more than one item for sale and buyers' valuations for sets of items are not additive or the set of feasible allocations is constrained, then constructing mechanisms that implement efficient allocations and have polynomial runtime might be very challenging. Consider a single seller selling $n\in \N$ heterogeneous indivisible items to several bidders. The Vickrey-Clarke-Groves auction generalizes the idea of the Vickrey auction to this multi-item setting. Naturally, every bidder has an intrinsic value for every subset of items. As in in the Vickrey auction, bidders report their valuations (Now, for every subset of items!). Then, the auctioneer computes a social welfare maximizing allocation according to the submitted bids and charges buyers the social cost of their winning that is incurred by the rest of the buyers. (This is the analogue to charging the second highest bid to the winning bidder in the single item Vickrey auction.) It turns out that the Vickrey-Clarke-Groves auction is also incentive compatible but it poses some problems: In fact, say for $n=40$, bidders would have to submit $2^{40}-1$ values (one value for each nonempty subset of the ground set) in total. Thus, asking every bidder for its valuation might be impossible due to time complexity issues. Therefore, even though the Vickrey-Clarke-Groves auction implements a social welfare maximizing allocation in this multi-item setting it might be impractical and there is need for alternative approaches to implement social welfare maximizing allocations.
This dissertation represents the results of three independent research papers all of them tackling the problem of implementing efficient allocations in different combinatorial settings.
Formulations of macrocyclic lactone anthelmintics such as moxidectin are regularly administered to sheep to combat parasites. A disadvantage of these pharmaceuticals are their side effects on non-target organisms when entering the environment. Little is known about anthelmintic effects on plant reproduction and whether the effects depend on environmental factors. For ecological and methodological reasons, we aimed at testing whether temperature affects the efficacy of a common moxidectin-based formulation on seed germination. We carried out a germination experiment including three typical species of temperate European grasslands (Centaurea jacea, Galium mollugo, Plantago lanceolata). We applied three temperature regimes (15/5, 20/10, 30/20°C), and a four-level dilution series (1:100–1:800) of formulated moxidectin (i.e., Cydectin oral drench). These solutions represent seed-anthelmintic contacts in the digestive tract of sheep shortly after deworming. In addition, a control was carried out with purified water only. We regularly counted emerging seedlings and calculated final germination percentage, mean germination time and synchrony of germination. Formulated moxidectin significantly reduced percentage, speed and synchrony of germination. A 1:100 dilution of the formulation reduced germination percentage by a quarter and increased mean germination time by six days compared to the control. Temperature moderated effects of the anthelmintic drug on germination in all response variables and all species, but in different patterns and magnitudes (significant anthelmintic x temperature x species interactions). In all response variables, the two more extreme temperature regimes (15/5, 30/20°C) led to the strongest effects of formulated moxidectin. With respect to germination percentage, G. mollugo was more sensitive to formulated moxidectin at the warmest temperature regime, whereas P. lanceolata showed the highest sensitivity at the coldest regime. This study shows that it is important to consider temperature dependencies of the effects of pharmaceuticals on seed germination when conducting standardised germination experiments.
The argan woodlands of South Morocco represent an open-canopy dryland forest with traditional silvopastoral usage that includes browsing by goats, sheep and camels, oil production as well as agricultural use. In the past, these forests have undergone extensive clearing, but are now protected by the state. However, the remaining argan woodlands are still under pressure from intensive grazing and illegal firewood collection. Although the argan-forest area seems to be overall decreasing due to large forest clearings for intensive agriculture, little quantitative data is available on the dynamics and overall state of the remaining argan forest. To determine how the argan woodlands in the High Atlas and the Anti-Atlas had changed in tree-crown cover from 1972 to 2018 we used historical black and white HEXAGON satellite images as well as recent WorldView satellite images (see Part A of our study). Because tree shadows can oftentimes not be separated from the tree crown on panchromatic satellite images, individual trees were mapped in three size categories to determine if trees were unchanged, had decreased/increased in crown size or had disappeared or newly grown. The current state of the argan trees was evaluated by mapping tree architectures in the field. Tree-cover changes varied highly between the test sites. Trees that remained unchanged between 1972 and 2018 were in the majority, while tree mortality and tree establishment were nearly even. Small unchanged trees made up 48.4% of all remaining trees, of these 51% showed degraded tree architectures. 40% of small (re-) grown trees were so overbrowsed that they only appeared as bushes, while medium (3–7 m crown diameter) and large trees (>7 m) showed less degraded trees regardless if they had changed or not. Approaches like grazing exclusion or cereal cultivation lead to a positive influence on tree architecture and less tree-cover decrease. Although the woodland was found to be mostly unchanged 1972–2018, the analysis of tree architecture reveals that a lot of (mostly small) trees remained stable but in a degraded state. This stability might be the result of the small trees’ high degradation status and shows the heavy pressure on the argan forest.
With the start of the Coronavirus (COVID-19) pandemic, the global education system has a faced immense challenges and disruptions resulting in and the necessity for an immediate redesign of teaching and learning in the school context. Face-to-face classroom instruction had to be replaced by ‘emergency remote teaching’, requiring teacher to adapt their daily routines to a new and unprecedented educational reality. Researchers and policymakers worldwide have agreed that, despite the fact that efforts were made to immediately adapt to emergency remote teaching, disadvantaged and vulnerable students may be especially at risk in emergency remote teaching. Given the differences in schooling organization across countries during the COVID-19 pandemic it can be expected that teachers performed inclusive instructional practices significantly different. Against the unpredictable situation, cross-country research has been urgently required to provide data that could inform education policy. Thus, this study explored teachers’ perceptions of supporting at risk students during the first COVID-19 school closures, as well as examining teachers’ inclusive teaching practices in three countries: Germany, Austria and Portugal. ANOVA results revealed important country differences. In general, it appears that teachers in Germany and Austria reported to have implemented less practices to address vulnerable and at-risk students compared to Portuguese teachers. Implications of the results, as well as further lines of research are outlined.
Objective: Attunement is a novel measure of nonverbal synchrony reflecting the duration of the present moment shared by two interaction partners. This study examined its association with early change in outpatient psychotherapy.
Methods: Automated video analysis based on motion energy analysis (MEA) and cross-correlation of the movement time-series of patient and therapist was conducted to calculate movement synchrony for N = 161 outpatients. Movement-based attunement was defined as the range of connected time lags with significant synchrony. Latent change classes in the HSCL-11 were identified with growth mixture modeling (GMM) and predicted by pre-treatment covariates and attunement using multilevel multinomial regression.
Results: GMM identified four latent classes: high impairment, no change (Class 1); high impairment, early response (Class 2); moderate impairment (Class 3); and low impairment (Class 4). Class 2 showed the strongest attunement, the largest early response, and the best outcome. Stronger attunement was associated with a higher likelihood of membership in Class 2 (b = 0.313, p = .007), Class 3 (b = 0.251, p = .033), and Class 4 (b = 0.275, p = .043) compared to Class 1. For highly impaired patients, the probability of no early change (Class 1) decreased and the probability of early response (Class 2) increased as a function of attunement.
Conclusions: Among patients with high impairment, stronger patient-therapist attunement was associated with early response, which predicted a better treatment outcome. Video-based assessment of attunement might provide new information for therapists not available from self-report questionnaires and support therapists in their clinical decision-making.
Der Beitrag diskutiert das Verhältnis von allgemeiner und beruflicher Bildung mithilfe einer pädagogischen Theorie der Organisation. Allgemeine Erwachsenen- und berufliche Bildung treffen sich darin, Prozess und Ergebnis der wechselseitigen Auseinandersetzung zwischen Organisationen und Individuen zu sein. Organisationen sind nicht nur Kontext von Bildung, sie sind selber gebildet und wirken pädagogisch. Der Beitrag erschließt die Funktion und Eigenlogik der Organisation in ihrem Verhältnis zu den Grundbegriffen der Erziehungswissenschaft. Der Vorschlag ermöglicht es, Lernen, Sozialisation, Erziehung, Bildung und Kompetenz in ihrer Geltung für die ineinander verflochtene Strukturdynamik von Organisation und Subjekt zu differenzieren. Als Alternative zur geläufigen Organisationsvergessenheit der Pädagogik integriert die differenzierende Ordnung dieser Begriffe eine erziehungswissenschaftliche Perspektive auf Individuum und Organisation.
In den letzten Jahren hat die Alternative für Deutschland (AfD) das Thema Wohnen zunehmend in ihren Wahlprogrammen aufgegriffen und für die eigene politische Profilierung genutzt. Der Beitrag zeigt, inwiefern die Thematisierung des Wohnens bei der AfD so formuliert ist, dass es erhebliche Anknüpfungspunkte an rechtes Gedankengut herstellt. Der Beitrag thematisiert die damit verbundenen Herausforderungen für eine kritische Stadtforschung sowie für progressive soziale Bewegungen. Er plädiert dafür, sich der Gefahr der rechten Vereinnahmung des Themas Wohnen bewusst zu sein und sich deutlich von den rechten Übernahmeversuchen abzugrenzen sowie Gegenstrategien zu entwickeln.
Background: Hyperhidrosis (excessive sweating, OMIM %114110) is a complex disorder with multifactorial causes. Emotional strains and social stress increase symptoms and lead to a vicious circle. Previously, we showed significantly higher depression scores, and normal cortisol awakening responses in patients with primary focal hyperhidrosis (PFH). Stress reactivity in response to a (virtual) Trier Social Stress Test (TSST-VR) has not been studied so far. Therefore, we measured sweat secretion, salivary cortisol and alpha amylase (sAA) concentrations, and subjective stress ratings in affected and non-affected subjects in response to a TSST-VR.
Method: In this pilot study, we conducted TSST-VRs and performed general linear models with repeated measurements for salivary cortisol and sAA levels, heart rate, axillary sweat and subjective stress ratings for two groups (diagnosed PFH (n = 11), healthy controls (n = 16)).
Results: PFH patients showed significantly heightened sweat secretion over time compared to controls (p = 0.006), with highest quantities during the TSST-VR. In both groups, sweating (p < 0.001), maximum cortisol levels (p = 0.002), feelings of stress (p < 0.001), and heart rate (p < 0.001) but not sAA (p = 0.068) increased significantly in response to the TSST-VR. However, no differences were detected in subjective ratings, cortisol concentrations and heart rate between PFH patients and controls (pall > 0.131).
Conclusion: Patients with diagnosed PFH showed stress-induced higher sweat secretion compared to healthy controls but did not differ in the stress reactivity with regard to endocrine or subjective markers. This pilot study is in need of replication to elucidate the role of the sympathetic nervous system as a potential pathway involved in the stress-induced emotional sweating of PFH patients.
Influence of Ozone and Drought on Tree Growth under Field Conditions in a 22 Year Time Series
(2022)
Studying the effect of surface ozone (O3) and water stress on tree growth is important for planning sustainable forest management and forest ecology. In the present study, a 22-year long time series (1998–2019) on basal area increment (BAI) and fructification severity of European beech (Fagus sylvatica L.) and Norway spruce (Picea abies (L.) H.Karst.) at five forest sites in Western Germany (Rhineland Palatinate) was investigated to evaluate how it correlates with drought and stomatal O3 fluxes (PODY) with an hourly threshold of uptake (Y) to represent the detoxification capacity of trees (POD1, with Y = 1 nmol O3 m−2 s−1). Between 1998 and 2019, POD1 declined over time by on average 0.31 mmol m−2 year−1. The BAI showed no significant trend at all sites, except in Leisel where a slight decline was observed over time (−0.37 cm2 per year, p < 0.05). A random forest analysis showed that the soil water content and daytime O3 mean concentration were the best predictors of BAI at all sites. The highest mean score of fructification was observed during the dry years, while low level or no fructification was observed in most humid years. Combined effects of drought and O3 pollution mostly influence tree growth decline for European beech and Norway spruce.
Reconstructing invisible deviating events: A conformance checking approach for recurring events
(2022)
Conformance checking enables organizations to determine whether their executed processes are compliant with the intended process. However, if the processes contain recurring activities, state-of-the-art approaches unfortunately have difficulties calculating the conformance. The occurrence of complex temporal rules can further increase the complexity of the problem. Identifying this limitation, this paper presents a novel approach towards dealing with recurring activities in conformance checking. The core idea of the approach is to reconstruct the missing events in the event log using defined rules while incorporating specified temporal event characteristics. This approach then enables the use of native conformance checking algorithms. The paper illustrates the algorithmic approach and defines the required temporal event characteristics. Furthermore, the approach is applied and evaluated in a case study on an event log for melanoma surveillance.
List-method directed forgetting (LMDF) is the demonstration that people can intentionally forget previously studied information when they are asked to forget what they have previously learned and remember new information instead. In addition, recent research demonstrated that people can selectively forget when cued to forget only a subset of the previously studied information. Both forms of forgetting are typically observed in recall tests, in which the to-be-forgotten and to-be-remembered information is tested independent of original cuing. Thereby, both LMDF and selective directed forgetting (SDF) have been studied mostly with unrelated item materials (e.g., word lists). The present study examined whether LMDF and SDF generalize to prose material. Participants learned three prose passages, which they were cued to remember or forget after the study of each passage. At the time of testing, participants were asked to recall the three prose passages regardless of original cuing. The results showed no significant differences in recall of the three lists as a function of cuing condition. The findings suggest that LMDF and SDF do not occur with prose material. Future research is needed to replicate and extend these findings with (other) complex and meaningful materials before drawing firm conclusions. If the null effect proves to be robust, this would have implications regarding the ecological validity and generalizability of current LMDF and SDF findings.