Refine
Year of publication
Document Type
- Doctoral Thesis (898) (remove)
Language
- German (505)
- English (382)
- Multiple languages (7)
- French (4)
Keywords
- Deutschland (38)
- Stress (37)
- Optimierung (22)
- Modellierung (19)
- Fernerkundung (17)
- Hydrocortison (16)
- stress (16)
- Motivation (12)
- Stressreaktion (12)
- cortisol (12)
Institute
- Psychologie (182)
- Raum- und Umweltwissenschaften (148)
- Fachbereich 4 (72)
- Mathematik (64)
- Wirtschaftswissenschaften (61)
- Fachbereich 1 (34)
- Geschichte, mittlere und neuere (28)
- Informatik (28)
- Germanistik (26)
- Fachbereich 6 (22)
- Kunstgeschichte (22)
- Politikwissenschaft (18)
- Anglistik (17)
- Fachbereich 2 (16)
- Soziologie (16)
- Fachbereich 3 (12)
- Philosophie (9)
- Romanistik (9)
- Computerlinguistik und Digital Humanities (7)
- Medienwissenschaft (6)
- Geschichte, alte (5)
- Allgemeine Sprach- und Literaturwissenschaft (4)
- Fachbereich 5 (4)
- Klassische Philologie (4)
- Pädagogik (4)
- Ethnologie (3)
- Japanologie (3)
- Sinologie (3)
- Archäologie (2)
- Rechtswissenschaft (2)
- Bodenkunde (1)
- Phonetik (1)
- Slavistik (1)
- Universitätsbibliothek (1)
Semantic-Aware Coordinated Multiple Views for the Interactive Analysis of Neural Activity Data
(2024)
Visualizing brain simulation data is in many aspects a challenging task. For one, data used in brain simulations and the resulting datasets is heterogeneous and insight is derived by relating all different kinds of it. Second, the analysis process is rapidly changing while creating hypotheses about the results. Third, the scale of data entities in these heterogeneous datasets is manifold, reaching from single neurons to brain areas interconnecting millions. Fourth, the heterogeneous data consists of a variety of modalities, e.g.: from time series data to connectivity data, from single parameters to a set of parameters spanning parameter spaces with multiple possible and biological meaningful solutions; from geometrical data to hierarchies and textual descriptions, all on mostly different scales. Fifth, visualizing includes finding suitable representations and providing real-time interaction while supporting varying analysis workflows. To this end, this thesis presents a scalable and flexible software architecture for visualizing, integrating and interacting with brain simulations data. The scalability and flexibility is achieved by interconnected services forming in a series of Coordinated Multiple View (CMV) systems. Multiple use cases are presented, introducing views leveraging this architecture, extending its ecosystem and resulting in a Problem Solving Environment (PSE) from which custom-tailored CMV systems can be build. The construction of such CMV system is assisted by semantic reasoning hence the term semantic-aware CMVs.
Nachdem er in den 1750er und 1760er Jahren graphische Bildsatiren zu aktuellen innen- und außenpolitischen Themen veröffentlich hatte, wurde William Hogarth selbst in zahlreichen Karikaturen verspottet und verleumdet. Ausgehend von dieser Beobachtung fragt die vorliegende Dissertation, welche Haltung sich den politischen Blättern des Künstlers entnehmen lässt und mit welchen künstlerischen Mitteln er dieser Ausdruck verlieh. Durch Analyse der politischen Ikonographie lassen sich die Themen und Akteure beschreiben. Mit der rezeptionsästhetischen Methode unter Hinzunahme der Sprech- und Bildakttheorie und der Propaganda Studies werden ihre tendenziösen Aussagen und manipulative Absichten entschlüsselt.
In ihrer Regierungsaffinität unterscheidet sich Hogarths politische Kunst maßgeblich von der oppositionellen Bildsatire Londons. Die Differenz spiegelt sich v. a. in den persönlichen Angriffen, mit denen zeitgenössische Satiriker Hogarth kritisierten. Als erstes reagierte Paul Sandby („The Painter’s March from Finchly“, 1753) auf Hogarths Darstellung des Jakobitischen Aufstandes 1745, womit er eine Begründung für die von William Augustus, Duke of Cumberland angestrebte Militärreform lieferte („March of the Guards to Finchley“, 1751); Für seine Gin Act-Kampagne („Gin Lane“ und „Beer Street“, 1750/51) erweiterte er die Pro-Gin-Ikonographie der 1730er Jahre (Anonymous: „The lamentable Fall of Madam Geneva”, 1736, Anonymous: „To those melancholly Sufferers the Destillers […] The Funeral Procession of Madam Geneva“, 1751), um sich für die staatliche Reglementierung der Destillen auszusprechen. In seinen Publikationen zum Siebenjährigen Krieg, mit denen er die Politik der jeweiligen Regierungen unter Thomas Pellham-Holles, Duke of Newcastle und William Pitt (the Elder) („The Invasion“, 1756) oder John Stuart, Earl of Bute („The Times Pl. 1“, 1763) unterstützte, zeigt sich Hogarths Opportunismus. Letztlich wurde seine Fürsprache für die unbeliebte Tory-Regierung und seine Kritik an William Pitt Anlass für Hogarths Herabwürdigung durch die Whig-treue Satire. Nach diesem Bruch publizierten beide Seiten verunglimpfende Portraitkarikaturen, die auf Rufmord des Gegners durch Kriminalisierung, Deformation und Dämonisierung setzten (William Hogarth: „John Wilkes Esqr.“, 1763, Anonymous „Tit for Tat“, 1763, Anonymous: „An Answer to the Print of John Wilkes Esqr. by WM Hogarth“, 1763, Anonymous: „Pug the snarling cur chastised Or a Cure for the Mange“, 1763).
Die Bildvergleiche zwischen Hogarths politischen Werken und den Reaktionen, die sie hervorriefen, zeigen, dass der Unterschied nicht im Bildgegenstand oder der politischen Ikonographie liegt, sondern in der Ausrichtung ihrer politischen Einflussnahme. Dabei ist vor allem Hogarths regierungsloyale Haltung hervorzuheben. Folglich muss die Forschungsmeinung von einer grundsätzlich kritischen Haltung Hogarths redigiert werden, da er sich nachweislich konservativ positioniert und dem Regierungshandeln und Machterhalt der Eliten Vorschub leistete.
Das vorliegende Dissertationsvorhaben untersucht die propagandistische Qualität der Werke Hogarths im Vergleich zu den zeitgenössischen Satirikern und macht die unterschiedliche politische Stoßrichtung sichtbar. Aufschluss gibt dabei die Anwendung künstlerischer und karikaturesker Mittel (das „Wie“) zum Zweck der burlesque (Posse/Parodie), des ridicule (Lächerlichmachung/Spott) bis bin zur Agitation, sowohl in Hogarths Werken als auch in den Karikaturen, die gegen ihn gerichtet waren. Da William Hogarth diese Stilmittel maßgeblich prägte und ihre Entwicklung forcierte, werden sie in der vorliegenden Arbeit unter dem Begriff Hogarthian Wit summiert. Mithilfe der Methode und Begriffe der Propaganda Studies lassen sich Intention und Zweck (das „Was“) als Bildakte beschreiben: Während es sich bei den Werken grundsätzlich um bias handelte, die basierend auf einer Ideologie die öffentliche Meinung beeinflusste, nahm ihre Schlagkraft in den 1760er Jahren stark zu; auf verrätselte Stellungnahmen folgte persönliche und offene Kritik an öffentlichen Personen, bis hin zum Rufmord. Dabei rezipierten sich die Künstler gegenseitig und bildeten Thesen und Antithesen aus. Hogarths einseitige Darstellungen wurden korrigiert und ergänzt, seine politische Kunst als Propaganda enttarnt. Schließlich wurden ihm Lügen und üble Nachrede vorgeworfen. Indem sie ihn anklagten oder durch Sekundärstigmatisierung eine Bestrafung in effigie vornahmen, forderten die Werke vom Rezipienten ein strafendes Urteil. Zu den künstlerischen Mitteln, die dabei zur Anwendung kommen, gehören eine politische Ikonographie und stereotype Feindbilder sowie nationale Konstruktionen, rezeptionsästhetische Mittel wie Juxtapositionen, Rezeptions- und Identifikationsfiguren sowie rhetorische und Mittel des Sprechakts, bis hin zu Perlokutionen. Die Werke lassen sich als Propaganda und somit als hierarchische Kommunikation beschreiben, die manipulative Bildstrategien nutzten, welche nicht nur der Beeinflussung der öffentlichen Meinung dienten, sondern politische Handlungen forcierten. Bezeichnend ist, dass beide Seiten dieselben Ikonographie, Stil-, Kompositions- und Kommunikationsmittel anwendeten, unabhängig von ihrer politischen Aussage, wodurch der Hogarthian Wit gefestigt und stetig weiterentwickelt wurde.
In machine learning, classification is the task of predicting a label for each point within a data set. When the class of each point in the labeled subset is already known, this information is used to recognize patterns and make predictions about the points in the remainder of the set, referred to as the unlabeled set. This scenario falls in the field of supervised learning.
However, the number of labeled points can be restricted, because, e.g., it is expensive to obtain this information. Besides, this subset may be biased, such as in the case of self-selection in a survey. Consequently, the classification performance for unlabeled points may be limited. To improve the reliability of the results, semi-supervised learning tackles the setting of labeled and unlabeled data. Moreover, in many cases, additional information about the size of each class can be available from undisclosed sources.
This cumulative thesis presents different studies to combine this external cardinality constraint information within three important algorithms for binary classification in the supervised context: support vector machines (SVM), classification trees, and random forests. From a mathematical point of view, we focus on mixed-integer programming (MIP) models for semi-supervised approaches that consider a cardinality constraint for each class for each algorithm.
Furthermore, since the proposed MIP models are computationally challenging, we also present techniques that simplify the process of solving these problems. In the SVM setting, we introduce a re-clustering method and further computational techniques to reduce the computational cost. In the context of classification trees, we provide correct values for certain bounds that play a crucial role for the solver performance. For the random forest model, we develop preprocessing techniques and an intuitive branching rule to reduce the solution time. For all three methods, our numerical results show that our approaches have better statistical performances for biased samples than the standard approach.
Mixed-Integer Optimization Techniques for Robust Bilevel Problems with Here-and-Now Followers
(2025)
In bilevel optimization, some of the variables of an optimization problem have to be an optimal solution to another nested optimization problem. This specific structure renders bilevel optimization a powerful tool for modeling hierarchical decision-making processes, which arise in various real-world applications such as in critical infrastructure defense, transportation, or energy. Due to their nested structure, however, bilevel problems are also inherently hard to solve—both in theory and in practice. Further challenges arise if, e.g., bilevel problems under uncertainty are considered.
In this dissertation, we address different types of uncertainties in bilevel optimization using techniques from robust optimization. We study mixed-integer linear bilevel problems with lower-level objective uncertainty, which we tackle using the notion of Gamma-robustness. We present two exact branch-and-cut approaches to solve these Gamma-robust bilevel problems, along with cuts tailored to the important class of monotone interdiction problems. Given the overall hardness of the considered problems, we additionally propose heuristic approaches for mixed-integer, linear, and Gamma-robust bilevel problems. The latter rely on solving a linear number of deterministic bilevel problems so that no problem-specific tailoring is required. We assess the performance of both the exact and the heuristic approaches through extensive computational studies.
In addition, we study the problem of determining optimal tolls in a traffic network in which the network users hedge against uncertain travel costs in a robust way. The overall toll-setting problem can be seen as a single-leader multi-follower problem with multiple robustified followers. We model this setting as a mathematical problem with equilibrium constraints, for which we present a mixed-integer, nonlinear, and nonconvex reformulation that can be tackled using state-of-the-art general-purpose solvers. We further illustrate the impact of considering robustified followers on the toll-setting policies through a case study.
Finally, we highlight that the sources of uncertainty in bilevel optimization are much richer compared to single-level optimization. To this end, we study two aspects related to so-called decision uncertainty. First, we propose a strictly robust approach in which the follower hedges against erroneous observations of the leader's decision. Second, we consider an exemplary bilevel problem with a continuous but nonconvex lower level in which algorithmic necessities prevent the follower from making a globally optimal decision in an exact sense. The example illustrates that even very small deviations in the follower's decision may lead to arbitrarily large discrepancies between exact and computationally obtained bilevel solutions.
This dissertation examines the relevance of regimes for stock markets. In three research articles, we cover the identification and predictability of regimes and their relationships to macroeconomic and financial variables in the United States.
The initial two chapters contribute to the debate on the predictability of stock markets. While various approaches can demonstrate in-sample predictability, their predictive power diminishes substantially in out-of-sample studies. Parameter instability and model uncertainty are the primary challenges. However, certain methods have demonstrated efficacy in addressing these issues. In Chapter 1 and 2, we present frameworks that combine these methods meaningfully. Chapter 3 focuses on the role of regimes in explaining macro-financial relationships and examines the state-dependent effects of macroeconomic expectations on cross-sectional stock returns. Although it is common to capture the variation in stock returns using factor models, their macroeconomic risk sources are unclear. According to macro-financial asset pricing, expectations about state variables may be viable candidates to explain these sources. We examine their usefulness in explaining factor premia and assess their suitability for pricing stock portfolios.
In summary, this dissertation improves our understanding of stock market regimes in three ways. First, we show that it is worthwhile to exploit the regime dependence of stock markets. Markov-switching models and their extensions are valuable tools for filtering the stock market dynamics and identifying and predicting regimes in real-time. Moreover, accounting for regime-dependent relationships helps to examine the dynamic impact of macroeconomic shocks on stock returns. Second, we emphasize the usefulness of macro-financial variables for the stock market. Regime identification and forecasting benefit from their inclusion. This is particularly true in periods of high uncertainty when information processing in financial markets is less efficient. Finally, we recommend to address parameter instability, estimation risk, and model uncertainty in empirical models. Because it is difficult to find a single approach that meets all of these challenges simultaneously, it is advisable to combine appropriate methods in a meaningful way. The framework should be as complex as necessary but as parsimonious as possible to mitigate additional estimation risk. This is especially recommended when working with financial market data with a typically low signal-to-noise ratio.
Veterinärantibiotika werden weltweit in großem Umfang zur Behandlung von Tierkrankheiten eingesetzt. Aufgrund der schlechten Resorption der Mittel im Darm der Tiere gelangen sie zum Großteil unverändert über Ausscheidungen auf landwirtschaftliche Nutzflächen. Dort können sie von Nichtzielorganismen, wie Gefäßpflanzen, aufgenommen werden und deren frühe Entwicklung bedrohen. In diesem Kontext wurde bisher vor allem der Einfluss auf Kulturpflanzen untersucht, während Wildpflanzenarten des ökologisch bedeutsamen Kulturgraslandes, die vor allem durch Gülleausbringung in Kontakt mit Antibiotikastoffen kommen, deutlich weniger fokussiert wurden. Deshalb wurde in dieser Arbeit der Einfluss realistischer Konzentrationen (0,1 - 20 mg/L) zweier häufig verwendeter Veterinärantibiotika, Tetracyclin und Sulfamethazin, auf die Keimung und das frühe Wachstum von typischen Arten des temperaten Kulturgraslandes untersucht. Da in der Natur oft mehrere Stressoren gleichzeitig auf einen Organismus einwirken, wurden auch zwei Multistressszenarien, nämlich Pharmazeutikamischungen und das Zusammenspiel von pharmazeutischem Wirkstoff mit abiotischen Bedingungen (Trockenstress) untersucht. In vier Themenblöcken wurden sowohl standardisierte Laborversuche als auch naturnähere Topf- und Feldversuche durchgeführt.
Die Ergebnisse zeigten, dass sowohl die Keimung als auch das frühe Wachstum durch beide Wirkstoffe, jedoch häufiger durch Tetracyclin, beeinträchtigt wurden. Während die Keimung uneinheitlich in Bezug auf die Effektrichtung beeinflusst wurde, zeigte sich eine starke, antibiotika- und konzentrationsabhängige Reduktion der Wurzellänge vor allem durch Tetracyclin, in den Petrischalenversuchen (20 mg/L bis 96 %, bei Dactylis glomerata). Das oberirdische Wachstum (Blattlänge, Wuchshöhe, Biomasse) wurde geringer beinflusst, und dabei oft wachstumsfördernd. In der gesamten Arbeit zeigten sich immer wieder Hormesis- Effekte, d.h. geringe Konzentrationen, die stimulierend wirkten, während höhere Konzentrationen toxisch wirkten. Die betrachteten Kombinationen verschiedener Faktoren führten entgegen der Erwartung nicht eindeutig zu stärkeren oder alleinigen Einflüssen. In einzelnen Fällen zeigten sich solche Muster, jedoch wurden auch Verluste von Einzeleffekten bei den Kombinationen beobachtet oder Einzeleffekte, die sich dort erneut abbildeten.
Es zeigten sich, wenn auch uneinheitlich, signifikante Einflüsse auf die frühen Entwicklungsstadien von typischen Wildpflanzenarten, die bereits durch andere Faktoren einen Rückgang erfahren. Gerade im Hinblick auf die wiederholte Ausbringung von Gülle und die potenzielle Akkumulation dieser hoch persistenten Stoffe stellen Veterinärantibiotika einen weiteren wichtigen Einflussfaktor dar, der die Biodiversität und Artzusammensetzung gefährdet, weshalb zu einem umweltbewussten Umgang mit ihnen geraten wird.
Partial differential equations are not always suited to model all physical phenomena, especially, if long-range interactions are involved or if the actual solution might not satisfy the regularity requirements associated with the partial differential equation. One remedy to this problem are nonlocal operators, which typically consist of integrals that incorporate interactions between two separated points in space and the corresponding solutions to nonlocal equations have to satisfy less regularity conditions.
In PDE-constrained shape optimization the goal is to minimize or maximize an objective functional that is dependent on the shape of a certain domain and on the solution to a partial differential equation, which is usually also influenced by the shape of this domain. Moreover, parameters associated with the nonlocal model are oftentimes domain dependent and thus it is a natural next step to now consider shape optimization problems that are governed by nonlocal equations.
Therefore, an interface identification problem constrained by nonlocal equations is thoroughly investigated in this thesis. Here, we focus on rigorously developing the first and second shape derivative of the associated reduced functional. In addition, we study first- and second-order shape optimization algorithms in multiple numerical experiments.
Moreover, we also propose Schwarz methods for nonlocal Dirichlet problems as well as regularized nonlocal Neumann problems. Particularly, we investigate the convergence of the multiplicative Schwarz approach and we conduct a number of numerical experiments, which illustrate various aspects of the Schwarz method applied to nonlocal equations.
Since applying the finite element method to solve nonlocal problems numerically can be quite costly, Local-to-Nonlocal couplings emerged, which combine the accuracy of nonlocal models on one part of the domain with the fast computation of partial differential equations on the remaining area. Therefore, we also examine the interface identification problem governed by an energy-based Local-to-Nonlocal coupling, which can be numerically computed by making use of the Schwarz method. Here, we again present a formula for the shape derivative of the associated reduced functional and investigate a gradient based shape optimization method.
Based on data collected from two surveys conducted in Germany and Taiwan, my first paper (Chapter 2) examines the impact of culture through language priming (Chinese vs. German or English) on individuals’ price fairness perception and attitudes towards government intervention and economic policy involving inequality. We document large cross-language differences: in both surveys, subjects who were asked and answered in Chinese demonstrated significantly higher perceived price fairness in a free market mechanism than their counterparts who completed the survey in German or English language. They were also more inclined to accept a Pareto improvement policy which increases social and economic inequality. In the second survey, Chinese language induced also a lower readiness to accept government intervention in markets with price limits compared to English language. Since language functions as a cultural mindset prime, our findings imply that culture plays an important role in fairness perception and preferences regarding social and economic inequality.
Chapter 3 of this work deals with patriotism priming. By conducting two online experimental studies conducted in Germany and China, we tested three different kinds of priming methods for constructive and blind patriotism respectively. Subjects were randomly distributed to one of three treatments motivated by previous studies in different countries: a constructive patriotism priming treatment, a blind patriotism priming treatment and a non-priming baseline. While the first experiment had a between-subject design, the second one enabled both a between-subject and within-subject comparison, since the level of patriotism of individuals was measured before and after priming respectively. The design of the second survey also enabled a comparison among the three priming methods for constructive and blind patriotism. The results showed that the tested methods, especially the national achievements as a priming mechanism, functioned well overall for constructive patriotism.
Surprisingly, the priming for blind patriotism did not work in either Germany or China and the opposite results were observed. Discussion and implications for future studies are provided at the end of the chapter.
Using data from the same studies as in Chapter 3, Chapter 4 examines the impact of patriotism on individuals’ fairness perception and preferences regarding inequality and on their attitudes toward economic policy involving inequality. Across surveys and countries, a positive and significant effect of blind patriotism on economic individualism was found. For China, we also found a significant relationship between blind patriotism and the agreement to unequal economic policy. In contrast to blind patriotism, we did not find an association of constructive patriotism to economic individualism and to attitudes toward economic policy involving inequality. Political and economic implications based on the results are discussed.
The last chapter (Chapter 5) studies the self-serving bias (when an individual’s perception about fairness is biased by self-interest) in the context of price setting and profit distribution. By analyzing data from four surveys conducted in six countries, we found that the stated appropriate product price and the fair allocation of profit was significantly higher, when the outcome was favorable to oneself. This self-serving bias in price fairness perception, however, differed across countries significantly and was significantly higher in Germany, Taiwan and China than in Vietnam, Estonia and Japan.
Although economic individualism and masculinity were found to have a significant negative effect on self-interest bias in price fairness judgment, they did not sufficiently explain the differences in self-interest bias between countries. Furthermore, we also observed an increase of self-interest bias in profit allocation over time in time-series data for one country (Germany) with data from 2011 to 2023.
The four papers are all co-authored with Prof. Marc Oliver Rieger, and the first paper has been accepted for publications in Review of Behavioral Economics.
The gender wage gap in labor market outcomes has been intensively investigated for decades, yet it remains a relevant and innovative research topic in labor economics. Chapter 2 of this dissertation explores the pressing issue of gender wage disparity in Ethiopia. By applying various empirical methodologies and measures of occupational segregation, this chapter aims to analyze the role of female occupational segregation in explaining the gender wage gap across the pay distribution. The findings reveal a significant difference in monthly wages, with women consistently earning lower wages across the wage distribution.
Importantly, the result indicates a negative association between female occupational segregation and the average earnings of both men and women. Furthermore, the estimation result shows that female occupational segregation partially explains the gender wage gap at the bottom of the wage distribution. I find that the magnitude of the gender wage gap in the private sector is higher than in the public sector.
In Chapter 3, the Ethiopian Demography and Health Survey data are leveraged to explore the causal relationship between female labor force participation and domestic violence. Domestic violence against women is a pervasive public health concern, particularly in Africa, including Ethiopia, where a significant proportion of women endure various forms of domestic violence perpetrated by intimate partners. Economic empowerment of women through increased participation in the labor market can be one of the mechanisms for mitigating the risk of domestic violence.
This study seeks to provide empirical evidence supporting this hypothesis. Using the employment rate of women at the community level as an instrumental variable, the finding suggests that employment significantly reduces the risk of domestic violence against women. More precisely, the result shows that women’s employment status significantly reduces domestic violence by about 15 percentage points. This finding is robust for different dimensions of domestic violence, such as physical, sexual, and emotional violence.
By examining the employment outcomes of immigrants in the labor market, Chapter 4 extends the dissertation's inquiry to the dynamics of immigrant economic integration into the destination country. Drawing on data from the German Socio-Economic Panel, the chapter scrutinizes the employment gap between native-born individuals and two distinct groups of first-generation immigrants: refugees and other migrants. Through rigorous analysis, Chapter 4 aims to identify the factors contributing to disparities in employment outcomes among these groups. In this chapter, I aim to disentangle the heterogeneity characteristic of refugees and other immigrants in the labor market, thereby contributing to a deeper understanding of immigrant labor market integration in Germany.
The results show that refugees and other migrants are less likely to find employment than comparable natives. The refugee-native employment gap is much wider than other migrant-native employment gap. Moreover, the findings vary by gender and migration categories. While other migrant men do not differ from native men in the probability of being employed, refugee women are the most disadvantaged group compared to other migrant women and native women in the probability of being employed. The study suggests that German language proficiency and permanent resident permits partially explain the lower employment probability of refugees in the German labor market.
Chapter 5 (co-authored with Uwe Jirjahn) utilizes the same dataset to explore the immigrant-native trade union membership gap, focusing on the role of integration in the workplace and into society. The integration of immigrants into society and the workplace is vital not only to improve migrant's performance in the labor market but also to actively participate in institutions such as trade unions. In this study, we argue that the incomplete integration of immigrants into the workplace and society implies that immigrants are less likely to be union members than natives. Our findings show that first-generation immigrants are less likely to be trade union members than natives. Notably, the analysis shows that the immigrant-native gap in union membership depends on immigrants’ integration into the workplace and society. The gap is smaller for immigrants working in firms with a works council and having social contacts with Germans. Moreover, the results reveal that the immigrant-native union membership gap is decreasing in the year since arrival in Germany.
In dieser Dissertation wird der Workflow der Erstellung einer Augmented Reality App für das Projekt „ARmob” auf Androidgeräten beschrieben. Diese App positioniert durch SfM-Technik erstellte, nach dem neuesten Stand der Forschung rekonstruierte 3D-Objekte an ihren ursprünglichen Standort in der Realität. Die virtuellen Objekte werden jeweils vom Standpunkt und Blickwinkel des Betrachters passend in die reale Welt eingeblendet, so dass der Eindruck entsteht, die Objekte seien Teil der Realität. Die lagegenaue Darstellung ist abhängig von der Satellitenerreichbarkeit der GNSS und der Genauigkeit der weiteren Sensoren. Die App soll als Grundlage und Framework für weitere Apps zur Erforschung der Raumwahrnehmung im Bereich der Kartographie dienen.
Convex Duality in Consumption-Portfolio Choice Problems with Epstein-Zin Recursive Preferences
(2025)
This thesis deals with consumption-investment allocation problems with Epstein-Zin recursive utility, building upon the dualization procedure introduced by [Matoussi and Xing, 2018]. While their work exclusively focuses on truly recursive utility, we extend their procedure to include time-additive utility using results from general convex analysis. The dual problem is expressed in terms of a backward stochastic differential equation (BSDE), for which existence and uniqueness results are established. In this regard, we close a gap left open in previous works, by extending results restricted to specific subsets of parameters to cover all parameter constellations within our duality setting.
Using duality theory, we analyze the utility loss of an investor with recursive preferences, that is, her difference in utility between acting suboptimally in a given market, compared to her best possible (optimal) consumption-investment behaviour. In particular, we derive universal power utility bounds, presenting a novel and tractable approximation of the investors’ optimal utility and her welfare loss associated to specific investment-consumption choices. To address quantitative shortcomings of those power utility bounds, we additionally introduce one-sided variational bounds that offer a more effective approximation for recursive utilities. The theoretical value of our power utility bounds is demonstrated through their application in a new existence and uniqueness result for the BSDE characterizing the dual problem.
Moreover, we propose two approximation approaches for consumption-investment optimization problems with Epstein-Zin recursive preferences. The first approach directly formalizes the classical concept of least favorable completion, providing an analytic approximation fully characterized by a system of ordinary differential equations. In the special case of power utility, this approach can be interpreted as a variation of the well-known Campbell-Shiller approximation, improving some of its qualitative shortcomings with respect to state dependence of the resulting approximate strategies. The second approach introduces a PDE-iteration scheme, by reinterpreting artificial completion as a dynamic game, where the investor and a dual opponent interact until reaching an equilibrium that corresponds to an approximate solution of the investors optimization problem. Despite the need for additional approximations within each iteration, this scheme is shown to be quantitatively and qualitatively accurate. Moreover, it is capable of approximating high dimensional optimization problems, essentially avoiding the curse of dimensionality and providing analytical results.
Globalization significantly transforms labor markets. Advances in production technologies, transportation, and political integration reshape how and where goods and services are produced. Local economic conditions and diverse policy responses create varying speeds of change, affecting regions' attractiveness for living and working -- and promoting mobility.
Competition for talent necessitates a deep understanding of why individuals choose specific destinations, how to ensure their effective labor market integration, and what workplace factors affect workers' well-being.
This thesis focuses on two crucial aspects of labor market change -- Migration and workplace technological change. It contributes to our understanding of the determinants of labor mobility, the factors facilitating migrant integration, and the role of workplace automation for worker well-being.
Chapter 2 investigates the relationship between minimum wages (MWs) and regional worker mobility in the EU. EU citizens are free to work anywhere in the common market, which allows them to take advantage of the significant variation in MWs across the EU. However, although MWs are set at the national level, it is also their local relevance that varies substantially -- depending on factors such as the share of affected workers or the extent to which they shift local compensation levels. These variations may attract workers from elsewhere, from within a country or from abroad.
Analyzing regional variations in the Kaitz index, a measure of local MW impact, reveals that higher MWs can significantly increase inflows of low-skilled EU workers, particularly in central Europe.
Chapter 3 examines the inequality in returns to skills experienced by immigrants, focusing on the role of linguistic proximity between migrants' origin and destination countries. Harmonized individual-level data from nine linguistically diverse migrant-hosting economies allows for an analysis of the wage gaps faced by immigrants from various origins, implicitly indicating how well they and their skills are integrated into the local labor markets. The analysis reveals that greater linguistic distance is associated with a higher wage penalty for highly skilled immigrants and a lower position in the wage distribution for those without tertiary education.
Chapter 4 investigates an institutional factor potentially relevant for the integration of immigrants -- the labor market impact of Confucius Institutes (CIs), Chinese government-sponsored institutions that promote Chinese language and culture abroad. CIs have been found to foster trade and cultural exchange, indicating their potential relevance in shaping attitudes and trust of natives towards China and Chinese individuals. Examining the relationship between local CI presence and the wages of Chinese immigrants in local labor markets of the United States, the analysis reveals that CIs associate with significantly reduced wages for nearby residing Chinese immigrants. An event study demonstrates that the mere announcement of a new CI negatively impacts local wages for Chinese immigrants, independent of the CI's actual opening.
Chapter 5 explores how working in automatable jobs affects life satisfaction in Germany. Following earlier literature, we classify occupations by potential for automation, and define the top third of occupations in this metric as \textit{automatable jobs}. We find workers in highly automatable jobs reporting a lower life satisfaction. Moreover, we detect a non-linearity, where workers in moderately automatable jobs (the second third of the distribution) experience a positive association with life satisfaction. Overall, the negative relationship of automation is most pronounced among younger and blue-collar workers, irrespective of the non-linearity.
Entrepreneurship is recognized as an important discipline to achieve sustainable development and to address sustainability goals without losing sight of economic aspects. However, entrepreneurship rates are rather low in many industrialized countries with high income levels. Research clearly shows that there is a gap in the entrepreneurial process between intentions and subsequent actions. This means that not everyone with entrepreneurial ambitions also follows through and implements actions. This gap also exists for aspects of sustainability. As a result, there is a need to better understand the traditional and sustainability-focused entrepreneurial process in order to increase corresponding actions. This dissertation offers such a comprehensive perspective and sheds light on individual and contextual predictors for traditional and sustainability-focused behavior of entrepreneurs and self-employed across four studies.
The first three studies focus on individual predictors. By providing a systematic literature review with 107 articles, Chapter 2 highlights the ambivalent role of religion for the entrepreneurial process. Relying on the theory of planned behavior (TPB) as theoretical basis, religion can have positive effects on entrepreneurial attitudes and behavioral control, but also negative consequences for other aspects of behavioral control and subjective norms due to religious restrictions.
The quantitative empirical study in Chapter 3 similarly relies on the TPB and sheds light on individual perceptual factors influencing the sustainability-related intention-action gap in entrepreneurship. Using data from the 2021 Global Entrepreneurship Monitor (GEM) Adult Population Survey (APS) including 22,008 early-stage entrepreneurs from 44 countries worldwide, the results support our theoretical reasoning that sustainability-focused intentions are positively related to social entrepreneurial actions. In addition, it is demonstrated that positive perceptual moderators such as self-efficacy and knowing other entrepreneurs as role models strengthen this relationship while a negative perception such as fear of failure restricts social actions in early-stage entrepreneurship.
The next quantitative empirical study in Chapter 4 examines the behavioral consequences of well-being at a sample of 6,955 German self-employed during COVID-19. This chapter builds on two complementary behavioral perspectives to predict how reductions in financial and non-financial well-being relate to investments in venture development. In this regard, reductions in financial well-being are positively related to time investments, supporting the performance feedback perspective in terms of higher search efforts under negative performance. In contrast, reductions in non-financial well-being are negatively related to time and monetary investments, yielding support for the broadening-and-build perspective indicating that negative psychological experiences narrow the thought-action repertoire and hinder resource deployment. The insights across these first three studies about individual predictors indicate that many different, subjective beliefs, perceptions and emotional states can influence the entrepreneurial process making entrepreneurship and self-employment highly individualized disciplines.
The last quantitative empirical study provides an explorative view on a large number of contextual predictors for social and ecological considerations in entrepreneurial actions. Combining GEM data from 2021 on country level with further information from the World Bank and the OECD, a machine learning approach is employed on a sample of 84 countries worldwide. The results suggest that governmental and regulatory as well as cultural factors are relevant to predict social and ecological considerations. Moreover, market-related aspects are shown to be relevant predictors, especially socio-economic factors for social considerations and economic factors for ecological considerations. Overall, the four studies in this dissertation highlight the complexity of the entrepreneurial process being determined by many different individual and contextual factors. Due to the multitude of potential predictors, this dissertation can only give an initial overview of a selection of factors with many more aspects and interdependencies still to be examined by future research.
Small and medium-sized enterprises (SMEs) and mid-sized companies are vital contributors to the global economy, driving employment growth, fostering innovation, and enhancing international competiveness. However, in the aftermath of the Great Financial Crisis (GFC) and the collapse of the large finance company CIT Group, which provided 60% loans to US middle-market firms, banks reduced their lending activities. Thus, it became challenging for firms to obtain long-term loans. The financing gap has increased further due to high interest rates, the COVID-19 pandemic, the unstable situation in the real estate market as well as higher costs due to the adoption of digital infrastructure and sustainability goals. Therefore, the search for alternative financing solutions outside bank lending and public markets became unavoidable for SMEs and mid-sized companies. Private debt funds entered the market, and, since the GFC, they have played a crucial role in offering alternative financing for firms globally. Private debt fund managers raise capital commitments through closed-end funds (like private equity) and make senior loans (like banks) directly to, mostly, middlemarket firms. The private debt market has experienced rapid growth in recent decades. The private debt funds assets under management (AuM) increased by 380% from 2008 to 2022, reaching $1.5 trillion AuM in 2022 . The high growth of private debt shows great interest from investors in this alternative asset class and lucrative investment opportunities.
Despite its substantial and growing size, the private debt market is relatively understudied. This dissertation introduces private debt as an important alternative financing source, provides an overview of private debt strategies, seniority, and structure, discusses the legal considerations concerning private debt, and briefly compares the two most mature private debt markets: Europe and the U.S. Moreover, it assesses the size of the European private debt market and compares its development in different European regions. Furthermore, it examines in detail the business model of private debt funds based on a survey of 191 European and U.S. private debt managers with private debt assets under management of over $390 billion. Finally, it delves deeper into the relationship between private debt and private equity funds and their role in buyouts.
To sum up, this dissertation provides a basis and inspiration for future research to expand upon and dive deeper into the world of private debt funds, their business model, and their impact on portfolio companies and the economy as a whole.
In this dissertation, I analyze how large players in financial markets exert influence on smaller players and how this affects the decisions of the large ones. I focus on how the large players process information in an uncertain environment, form expectations and communicate these to smaller players through their actions. I examine these relationships empirically in the foreign exchange market and in the context of a game-theoretic model of an investment project.
In Chapter 2, I investigate the relationship between the foreign exchange trading activity of large US-based market participants and the volatility of the nominal spot exchange rate. Using a novel dataset, I utilize the weekly growth rate of aggregate foreign currency positions of major market participants to proxy trading activity in the foreign exchange market. By estimating the heterogeneous autoregressive model of realized volatility (HAR-RV), I find evidence of a positive relationship between trading activity and volatility, which is mainly driven by unexpected changes in trading activity and is asymmetric for some of the currencies considered. My results contribute to the understanding of the drivers of exchange rate volatility and the role of large players in the flow of information in financial markets.
In Chapters 3 and 4, I consider a sequential global game of an investment project to examine how a large creditor influences the decisions of small creditors with her lending decision. I pay particular attention to the timing of the large player’s decision, i.e. whether she makes her decision to roll over a credit before or after the small players. I show that she faces a trade-off between signaling to and learning from small creditors. By being a focal point for coordination, her actions have a substantial impact on the probability of coordination failure and the failure of the investment project. I investigate the sensitivity of the equilibrium by comparing settings with perfect and imperfect learning. The results highlight the importance of signaling and provide a new perspective on the idea of catalytic finance and the influence of a lender-of-last-resort in self-fulfilling debt crises.
There is a wide range of methodologies for policy evaluation and socio-economic impact assessment. A fundamental distinction can be made between micro and macro approaches. In contrast to micro models, which focus on the micro-unit, macro models are used to analyze aggregate variables. The ability of microsimulation models to capture interactions occurring at the micro-level makes them particularly suitable for modeling complex real-world phenomena. The inclusion of a behavioral component into microsimulation models provides a framework for assessing the behavioral effects of policy changes.
The labor market is a primary area of interest for both economists and policy makers. The projection of labor-related variables is particularly important for assessing economic and social development needs, as it provides insight into the potential trajectory of these variables and can be used to design effective policy responses. As a result, the analysis of labor market behavior is a primary area of application for behavioral microsimulation models. Behavioral microsimulation models allow for the study of second-round effects, including changes in hours worked and participation rates resulting from policy reforms. It is important to note, however, that most microsimulation models do not consider the demand side of the labor market.
The combination of micro and macro models offers a possible solution as it constitutes a promising way to integrate the strengths of both models. Of particular relevance is the combination of microsimulation models with general equilibrium models, especially computable general equilibrium (CGE) models. CGE models are classified as structural macroeconomic models, which are defined by their basis in economic theory. Another important category of macroeconomic models are time series models. This thesis examines the potential for linking micro and macro models. The different types of microsimulation models are presented, with special emphasis on discrete-time dynamic microsimulation models. The concept of behavioral microsimulation is introduced to demonstrate the integration of a behavioral element into microsimulation models. For this reason, the concept of utility is introduced and the random utility approach is described in detail. In addition, a brief overview of macro models is given with a focus on general equilibrium models and time series models. Various approaches for linking micro and macro models, which can either be categorized as sequential approaches or integrated approaches, are presented. Furthermore, the concept of link variables is introduced, which play a central role in combining both models. The focus is on the most complex sequential approach, i.e., the bi-directional linking of behavioral microsimulation models with general equilibrium macro models.
In den letzten Jahren hat die Nutzung von Drohnen deutlich zugenommen. Dies liegt unter anderem an der Leistungssteigerung, der guten Verfügbarkeit und an dem einfachen Einsatz von Drohnen. Damit sind auch Anwendungen in der Forschung möglich geworden, die zuvor unmöglich oder mit hohen Kosten verbunden waren. Als Sensor zur Datenaufzeichnung findet im Bereich der Forschung häufig eine Kamera Verwendung. Zusammen mit einer Drohne können Bereiche einfach und kostengünstig überflogen und dabei erkundet, beobachtet oder überwacht werden. Neben der Kamera als Sensor werden auch häufig Multispektralkameras und Lidar eingesetzt. Dagegen findet Radar im Bereich von kleinen Drohnen kaum Anwendung. Ziel dieser Forschungsarbeit war es zu untersuchen, ob neuste Radartechnik einen Mehrwert in der Fernerkundung mit kleinen Drohnen bieten kann.
Hierfür wurden moderne Radarsensoren aus dem Automobilbereich ausgewählt. Als Drohnen wurden sowohl Quadrocopter als auch eine Starrflügler-Drohne eingesetzt. Für die Analyse, Berechnung und Auswertung der Daten wurde MATLAB verwendet. Der erste Ansatz beruhte auf einer Starrflügler-Drohne, die sich durch ihren freien Zugriff auf die Steuerung auszeichnet. Dadurch können auch spezielle Anforderungen an die Flugregelung berücksichtigt werden. Allerdings können mit einer Starrflügler-Drohne keine langsamen oder sogar statische Luftaufnahmen erstellt werden, um Erfahrung mit den Radardaten zu erlangen. Aus diesem Grund wurde anschließend ein Radar-Messsystem entworfen, das unabhängig von der Drohne eingesetzt werden kann. Zusammen mit einem Quadrocopter konnten so statische Radarmessungen durchgeführt werden, um die Verwendbarkeit der Radardaten in der Fernerkundung zu bestätigen. Das Messsystem konnte so aber nur für 2-dimensionale Anwendungen eingesetzt werden. In der weiteren Forschungsarbeit wurde untersucht, ob es möglich ist, mit einem Radarsensor der nur in 2-dimensionen misst eine 3-dimensionale Aufzeichnungen zu erstellen. Als Versuchsobjekt wurde eine Hütte gewählt, die Anhand der Radardaten dargestellt werden sollte. Dafür wurde ein Prozess zur Datenverarbeitung mit elf Schritten entworfen, womit die Hütte auf 0,6 Meter genau rekonstruiert werden konnte. Im letzten Teil der Forschungsarbeit wurde untersucht, ob sich die Genauigkeit des Messsystems erhöhen lässt, um noch mehr Anwendungsfälle bedienen zu können. Dafür wurde ein neuer Radarsensor eingesetzt, der eine höhere Genauigkeit besitzt. Die Forschungsarbeit konzentrierte sich darauf, die Abhängigkeit der Radardaten zum ungenauen Lagesensor aufzulösen. Dabei wurde die Fluglage über die Radardaten selbst berechnet, womit die Fluglage genauer bestimmt werden kann als allein über den Lagesensor. Erst damit kann die höhere Genauigkeit des neuen Radarsensors auch tatsächlich ausgenutzt werden.
Mit den Ergebnissen der Forschungsarbeit sowie den vorgestellten Radarsensoren, stehen der Fernerkundung mit kleinen Drohnen, neben den klassischen Sensoren, zukünftig auch Radarsensoren zur Verfügung. Mit dem Messsystem und den Erkenntnissen aus der Forschungsarbeit werden bereits erste spezifische Anwendungen in Forschungsprojekten untersucht. Darüber hinaus konnten auch Anwendungsfälle außerhalb der Fernerkundung identifiziert werden. Die Weiterentwicklung im Bereich des autonomen Fahrens wird für Leistungssteigerungen bei Radarsensoren sorgen. Damit stehen auch der Fernerkundung zukünftig noch bessere Radarsensoren zur Verfügung.
Within this thesis the hedging behaviour of airlines from 2005 to 2019 is analysed by using an unbalanced panel dataset consisting of a total of 78 airlines from 39 countries. The focus of the analysis is on financial and operational hedging as well as the influence of both on CO2 emissions and the development of emitted CO2 emissions. For the analysis Probit models with random effects and OLS models with fixed effects were used.
The results regarding the relationship between leverage and financial hedging indicate a negative relationship between everage and financial fuel hedging and a non-linear convex relationship for highly leveraged airlines, which is contrary to the theory of financial distress.
In addition, the study provides evidence that airlines using other types of derivatives, such as interest rate derivatives, engage in more fuel hedging.
In terms of operational hedging, the analysis suggests that operating a diversified fleet is a complement to, rather than a substitute for, financial hedging. With regard to alliance membership, the results do not show that alliance membership is a substitute for financial hedging, as members of alliances are more likely to engage in hedging transactions and to a greater extent.
The analysis shows that the relative CO2 emissions fall in the period under review, but this does not apply to the absolute amount. No general statement can be made about the influence of financial and operational hedging on CO2 emissions, as the results are mixed.
Zirkularität und zirkulare Geschäftsmodelle in der Holzindustrie: eine empirische Untersuchung
(2025)
Der ökologische Zustand der Erde befindet sich infolge von Umweltverschmutzung, Abfallaufkommen und CO₂-bedingtem Klimawandel in einem kritischen Zustand. Mit rund 40 % trägt der Bau- und Gebäudesektor erheblich zu den globalen Treibhausgasemissionen bei. Holz gilt als klimafreundliche Alternative zu Beton und Stahl, bedarf jedoch ebenfalls einer nachhaltigen Nutzung. Die Kreislaufwirtschaft bietet mit der Wiederverwendung ein zukunftsweisendes Konzept: So sind etwa 45% des beim Rückbau von Gebäuden anfallenden Holzes potenziell als Rohstoff nutzbar. Dadurch werden alternative Rohstoffquellen erschlossen und das Abfallaufkommen reduziert.
Trotz dieses Potenzials liegt der Zirkularitätsgrad der Weltwirtschaft derzeit nur bei 7,2 %. Vor diesem Hintergrund untersucht die Dissertation, welche Wettbewerbsstrategien und welche organisationalen Fähigkeiten die Entwicklung zirkulärer Geschäftsmodelle fördern. Der Fokus liegt auf der Holzindustrie der DACH-Region, die historisch durch forstwirtschaftliche Nachhaltigkeit geprägt ist, jedoch bislang überwiegend linearen Strukturen folgt.
Die Arbeit kombiniert theoretische Fundierung, eine vierjährige Literaturrecherche, Experteninterviews sowie im Zentrum eine quantitative Unternehmensbefragung (n = 200). Daraus wurde eine aktivitätsorientierte Skala zur Bewertung der Zirkularität eines Geschäftsmodells entwickelt. Analysiert wurden drei Perspektiven: Fähigkeiten, Strategien und Stakeholder.
Im Kontext der Fähigkeitsperspektive wurde ermittelt, dass die dynamischen Fähigkeiten positive Implikationen auf die Umsetzung von Zirkularität haben. Im Forschungsfeld der Strategieperspektive wurde deutlich, dass die Innovationsführerschaft positive Effekte auf die Umsetzung der Kreislaufwirtschaft besitzt. Zudem weisen sowohl die Innovationsführerschaft als auch die Qualitätsführerschaft einen positiven indirekten Effekt über die dynamischen Fähigkeiten auf die Entwicklung zirkulärer Geschäftsmodelle auf. Im Rahmen der Stakeholderperspektive wurde eruiert, dass der Stakeholder-Druck im Zusammenwirken mit einem grünen Unternehmensimage eine Katalysator-Wirkung besitzt. Der Einfluss der Interessengruppen führt dazu, dass die Unternehmen ein grünes Image in eine substanzielle Umsetzungsphase überführen. Darüber hinaus wurde ersichtlich, dass der Stakeholder-Druck als zentraler Veränderungsfaktor wirkt. Während die direkten Auswirkungen der dynamischen Fähigkeiten durch den Druck zurückgehen, nehmen die indirekten Effekte auf die Erreichung von Zirkularität zu. Abschließend werden Handlungsempfehlungen für Unternehmen sowie wissenschaftliche Implikationen und zukünftige Forschungsmöglichkeiten abgeleitet.
Case-Based Reasoning (CBR) is a symbolic Artificial Intelligence (AI) approach that has been successfully applied across various domains, including medical diagnosis, product configuration, and customer support, to solve problems based on experiential knowledge and analogy. A key aspect of CBR is its problem-solving procedure, where new solutions are created by referencing similar experiences, which makes CBR explainable and effective even with small amounts of data. However, one of the most significant challenges in CBR lies in defining and computing meaningful similarities between new and past problems, which heavily relies on domain-specific knowledge. This knowledge, typically only available through human experts, must be manually acquired, leading to what is commonly known as the knowledge-acquisition bottleneck.
One way to mitigate the knowledge-acquisition bottleneck is through a hybrid approach that combines the symbolic reasoning strengths of CBR with the learning capabilities of Deep Learning (DL), a sub-symbolic AI method. DL, which utilizes deep neural networks, has gained immense popularity due to its ability to automatically learn from raw data to solve complex AI problems such as object detection, question answering, and machine translation. While DL minimizes manual knowledge acquisition by automatically training models from data, it comes with its own limitations, such as requiring large datasets, and being difficult to explain, often functioning as a "black box". By bringing together the symbolic nature of CBR and the data-driven learning abilities of DL, a neuro-symbolic, hybrid AI approach can potentially overcome the limitations of both methods, resulting in systems that are both explainable and capable of learning from data.
The focus of this thesis is on integrating DL into the core task of similarity assessment within CBR, specifically in the domain of process management. Processes are fundamental to numerous industries and sectors, with process management techniques, particularly Business Process Management (BPM), being widely applied to optimize organizational workflows. Process-Oriented Case-Based Reasoning (POCBR) extends traditional CBR to handle procedural data, enabling applications such as adaptive manufacturing, where past processes are analyzed to find alternative solutions when problems arise. However, applying CBR to process management introduces additional complexity, as procedural cases are typically represented as semantically annotated graphs, increasing the knowledge-acquisition effort for both case modeling and similarity assessment.
The key contributions of this thesis are as follows: It presents a method for preparing procedural cases, represented as semantic graphs, to be used as input for neural networks. Handling such complex, structured data represents a significant challenge, particularly given the scarcity of available process data in most organizations. To overcome the issue of data scarcity, the thesis proposes data augmentation techniques to artificially expand the process datasets, enabling more effective training of DL models. Moreover, it explores several deep learning architectures and training setups for learning similarity measures between procedural cases in POCBR applications. This includes the use of experience-based Hyperparameter Optimization (HPO) methods to fine-tune the deep learning models.
Additionally, the thesis addresses the computational challenges posed by graph-based similarity assessments in CBR. The traditional method of determining similarity through subgraph isomorphism checks, which compare nodes and edges across graphs, is computationally expensive. To alleviate this issue, the hybrid approach seeks to use DL models to approximate these similarity calculations more efficiently, thus reducing the computational complexity involved in graph matching.
The experimental evaluations of the corresponding contributions provide consistent results that indicate the benefits of using DL-based similarity measures and case retrieval methods in POCBR applications. The comparison with existing methods, e.g., based on subgraph isomorphism, shows several advantages but also some disadvantages of the compared methods. In summary, the methods and contributions outlined in this work enable more efficient and robust applications of hybrid CBR and DL in process management applications.
When natural phenomena and data-based relations are driven by dynamics which are not purely local, they cannot be described satisfactorily by partial differential equations. As a consequence, mathematical models governed by nonlocal operators are of interest. This thesis is concerned with nonlocal operators of the form
$\mathcal{L}u(x) = PV \int_{\mathbb{R}^d} (u(x)-u(y)) K(x,dy), x \in \mathbb{R}^d$,
which are determined through a family of Borel measures $K=(K(x, \cdot))_{x \in \mathbb{R}^d}$ on $\mathbb{R}^d$ and which act on the vector space of Borel measurable functions $u: \mathbb{R}^d \rightarrow \mathbb{R}$. For a large class of families $K$, namely those where $K$ is a symmetric transition kernel satisfying a specific non-degeneracy condition, a variational theory for nonlocal equations of the type $\mathcal{L}u=f$ is established which builds upon gadgets from both measure theory and classical analysis. While measure theory is used to provide a nonlocal integration by parts formula that allows to set up a reasonable variational formulation of the above equation in dependency of the particular boundary condition (Dirichlet, Robin, Neumann) considered, Hilbert space theory and fixed-point approaches are utilized to develop sufficient conditions for the existence of variational solutions. This theory is then applied to two specific realizations of $\mathcal{L}$ of interest before a weak maximum principle is established which is finally used to study overlapping domain decomposition methods for the nonlocal and homogeneous Dirichlet problem.
Die Abteilung Kunstschutz der deutschen Wehrmacht im besetzten Griechenland (1941-1944) bestand aus wehrpflichtigen deutschen Archäologen. Sie waren zunächst Stipendiaten oder Mitarbeiter des Archäologischen Instituts des Deutschen Reiches (AIDR) unter den Bedingungen des Nationalsozialismus, bevor sie im Zweiten Weltkrieg in der Uniform der Wehrmacht zurückkehrten. Ihre Biografien im Kontext der Abteilung Athen, deren Direktor Georg Karo bis 1936 war, sowie der Zentrale der Instituts, unter dem von 1932 bis 1936 amtierenden Präsidenten Theodor Wiegand, sind ein Untersuchungsgegenstand. Die außenpolitische Legitimation des NS-Regimes durch die Olympischen Spiele und der wichtigste wissenschaftspolitische Erfolg des Institutes, die Wiederaufnahme der Olympiagrabung, die Wiegand und Karo seit 1933 anstrebten und durch ihre politischen Netzwerke 1936 erreichten, werden in der Dissertation in ihrer wechselseitigen Bedingtheit aufgezeigt. Diese Anpassungsleistungen an das NS-Regime prägten den eigenen archäologischen Nachwuchs aber auch die griechische Gesellschaft.
Schutzmaßnahmen waren nur ein kleiner Tätigkeitsbereich der Kunstschützer aber ein wichtiger Teil der Wehrmachtspropaganda. Der Institutspräsident Martin Schede (1937 bis 1945) forderte Mitarbeitern vor allem für zwei AIDR-Projekte an: die Erstellung von Flugbildern von möglichst ganz Griechenland und Ausgrabungen auf Kreta. Bereits diese Zwischenergebnisse berechtigen zu dem Titel „Kunstschutz als Alibi“.
Die Dissertation versucht, die Frage zu beantworten, warum der archäologische Kunstschutz nicht mehr als ein Alibi sein konnte. Dies geschieht vor allem unter Berücksichtigung der politischen aber auch der militärischen Traditionslinien deutscher Archäologie in Griechenland und Deutschland.
The goal of this work is to compare operators that are defined on probably varying Hilbert spaces. Distance concepts for operators as well as convergence concepts for such operators are explained and examined. For distance concepts we present three main notions. All have in common that they use space-linking operators that connect the spaces. At first, we look at unitary maps and compare the unitary orbits of the operators. Then, we consider isometric embeddings, which is based on a concept of Joachim Weidmann. Then we look at contractions but with more norm equations in comparison. The latter idea is based on a concept of Olaf Post called quasi-unitary equivalence. Our main result is that the unitary and isometric distances are equal provided the operators are both self-adjoint and have 0 in their essential spectra. In the third chapter, we focus specifically on the investigation of these distance terms for compact operators or operators in p-Schatten classes. In this case, the interpretation of the spectra as null sequences allows further distance investigation. Chapter four deals mainly with convergence terms of operators on varying Hilbert spaces. The analyses in this work deal exclusively with concepts of norm resolvent convergence. The main conclusion of the chapter is that the generalisation for norm resolvent convergence of Joachim Weidmann and the generalisation of Olaf Post, called quasi-unitary equivalence, are equivalent to each other. In addition, we specify error bounds and deal with the convergence speed of both concepts. Two important implications of these convergence notions are that the approximation is spectrally exact, i.e., the spectra converge suitably, and that the convergence is transferred to the functional calculus of the bounded functions vanishing at infinity.
The new millennium has been characterized by rising digitalization, the proliferation of shadow banking, and significant advancements in machine learning and natural language processing. These trends present both challenges and opportunities, which my dissertation addresses. This cumulative dissertation investigates critical aspects of financial stability, monetary policy, and the transition towards cashless economies through three distinct but interrelated studies.
The first paper examines the risk-taking channel of monetary policy transmission within the euro area, focusing on shadow banks. Through vector autoregressive models, it assesses the impact of conventional and unconventional monetary policy shocks on shadow banks' asset growth and risk asset ratios. The results indicate that lower interest rates lead to a portfolio reallocation towards riskier assets and a general expansion of assets in shadow banks. In the case of conventional monetary policy shocks, both effects last three times as long as in the case of unconventional monetary policy shocks. Country-specific as well as sector-specific estimations confirm these findings. This study bridges gaps in the existing literature, especially in the eurozone, by highlighting the significant role shadow banks play in monetary policy transmission, suggesting implications for financial regulation and stability.
The second paper explores the influence of financial stability considerations on US monetary policy, particularly during the Great Recession. Utilizing natural language processing and machine learning techniques on congressional hearings, this study constructs indicators for financial stability sentiment expressed by the Federal Reserve Chairs. Empirical analysis is conducted using Taylor-rule models, revealing that negative financial stability sentiment is associated with a more accommodative monetary policy stance, even before the Great Recession. This work provides new insights into the integration of financial stability concerns into monetary policy frameworks, demonstrating the need for a balanced approach to economic stability. The article suggests that under a dual mandate, such as that of the Federal Reserve, financial stability can, to some extent, already be factored into monetary policy deliberations.
The third paper sheds new light on ``cash paradox'' by uncovering the factors of the cashless transition that has not been entirely understood so far. Using a comprehensive dataset across 65 countries, the study employs panel data models to explain the paradox (increasing demand for central bank money despite soaring digitalization), especially among technologically advanced countries, e.g., Japan. Empirical evidence suggests that digitalization is not significantly associated with higher reliance on physical cash. It uncovers a unique non-linear relationship between trust and cash usage (``Arch of Trust'') which holds after addressing potential endogeneity issues using 2SLS estimation. Opposed to the widespread misinterpretations of Keynes' (1937) reasons for holding cash, the findings highlight that distrust is the key factor unlocking two distinct puzzles in economics, linking cash hoarding with ``missing'' funds on capital markets and slower shift toward digital payments in low-trust societies. A key insight is the role of trust as a (social) insurance, cushion or safety net, dampening the perception of risk and reducing precautionary and transactionary demand for physical cash, while encouraging a shift towards riskier alternatives. This, in turn, is connected to the third puzzle, the ``paradox of prudence.'' A shift from riskier investments to safer assets, cash, may be prudent at the individual level but risky for the overall economy, a concern for macroprudential policymakers. Additionally, the research highlights the critical role of culture in driving the global movement towards cashless economies. Moreover, cultures that are more self-expression-oriented (which is the main cultural dimension) and culturally closer to Sweden are associated with less cash-intensive economies. These insights are vital for macroprudential regulators as well as for policymakers designing payment systems and CBDC in culturally diverse regions like the Eurozone.
Collectively, these papers contribute to a deeper understanding of monetary policy, financial stability, and the transition from cash-based to (nearly) cashless societies, offering significant theoretical and practical implications for academics, regulators and central bankers.
Although universality has fascinated over the last decades, there are still numerous open questions in this field that require further investigation. In this work, we will mainly focus on classes of functions whose Fourier series are universal in the sense that they allow us to approximate uniformly any continuous function defined on a suitable subset of the unit circle.
The structure of this thesis is as follows. In the first chapter, we will initially introduce the most important notation which is needed for our following discussion. Subsequently, after recalling the notion of universality in a general context, we will revisit significant results concerning universality of Taylor series. The focus here is particularly on universality with respect to uniform convergence and convergence in measure. By a result of Menshov, we will transition to universality of Fourier series which is the central object of study in this work.
In the second chapter, we recall spaces of holomorphic functions which are characterized by the growth of their coefficients. In this context, we will derive a relationship to functions on the unit circle via an application of the Fourier transform.
In the second part of the chapter, our attention is devoted to the $\mathcal{D}_{\textup{harm}}^p$ spaces which can be viewed as the set of harmonic functions contained in the $W^{1,p}(\D)$ Sobolev spaces. In this context, we will also recall the Bergman projection. Thanks to the intensive study of the latter in relation to Sobolev spaces, we can derive a decomposition of $\mathcal{D}_{\textup{harm}}^p$ spaces which may be seen as analogous to the Riesz projection for $L^p$ spaces. Owing to this result, we are able to provide a link between $\mathcal{D}_{\textup{harm}}^p$ spaces and spaces of holomorphic functions on $\mathbb{C}_\infty \setminus \s$ which turns out to be a crucial step in determining the dual of $\mathcal{D}_{\textup{harm}}^p$ spaces.
The last section of this chapter deals with the Cauchy dual which has a close connection to the Fantappié transform. As an application, we will determine the Cauchy dual of the spaces $D_\alpha$ and $D_{\textup{harm}}^p$, two results that will prove to be very helpful later on. Finally, we will provide a useful criterion that establishes a connection between the density of a set in the direct sum $X \oplus Y$ and the Cauchy dual of the intersection of the respective spaces.
The subsequent chapter will delve into the theory of capacities and, consequently, potential theory which will prove to be essential in formulating our universality results. In addition to introducing further necessary terminologies, we will define capacities in the first section following [16], however in the frame of separable metric spaces, and revisit the most important results about them.
Simultaneously, we make preparations that allow us to define the $\mathrm{Li}_\alpha$-capacity which will turn out to be equivalent to the classical Riesz $\alpha$-capacity. The $\mathrm{Li}_\alpha$-capacity proves to be more adapted to the $D_\alpha$ spaces. It becomes apparent in the course of our discussion that the $\mathrm{Li}_\alpha$-capacity is essential to prove uniqueness results for the class $D_\alpha$. This leads to the centerpiece of this chapter which forms the energy formula for the $\mathrm{Li}_\alpha$-capacity on the unit circle. More precisely, this identity establishes a connection between the energy of a measure and its corresponding Fourier coefficients. We will briefly deal with the complement-equivalence of capacities before we revisit the concept of Bessel and Riesz capacities, this time, however, in a much more general context, where we will mainly rely on [1]. Since we defined capacities on separable metric spaces in the first section, we can draw a connection between Bessel capacities and $\mathrm{Li}_\alpha$-capacities. To conclude this chapter, we would like to take a closer look at the geometric meaning of capacities. Here, we will point out a connection between the Hausdorff dimension and the polarity of a set, and transfer it to the $\mathrm{Li}_\alpha$-capacity. Another aspect will be the comparison of Bessel capacities across different dimensions, in which the theory of Wolff potentials crystallizes as a crucial auxiliary tool.
In the fourth chapter of this thesis, we will turn our focus to the theory of sets of uniqueness, a subject within the broader field of harmonic analysis. This theory has a close relationship with sets of universality, a connection that will be further elucidated in the upcoming chapter.
The initial section of this chapter will be dedicated to the notion of sets of uniqueness that is specifically adapted to our current context. Building on this concept, we will recall some of the fundamental results of this theory.
In the subsequent section, we will primarily rely on techniques from previous chapters to determine the closed sets of uniqueness for the class $\mathcal{D}_{\alpha}$. The proofs we will discuss are largely influenced by [16, p.\ 178] and [9, pp.\ 82].
One more time, it will become evident that the introduction of the $\mathrm{Li}_\alpha$-capacity in the third chapter and the closely associated energy formula on the unit circle, were the pivotal factors that enabled us to carry out these proofs.
In the final chapter of our discourse, we will present our results on universality. To begin, we will recall a version of the universality criterion which traces back to the work of Grosse-Erdmann (see [26]). Coupled with an outcome from the second chapter, we will prove a result that allows us to obtain the universality of a class using the technique of simultaneous approximation. This tool will play a key role in the proof of our universality results which will follow hereafter.
Our attention will first be directed toward the class $D_\alpha$ with $\alpha$ in the interval $(0,1]$. Here, we summarize that universality with respect to uniform convergence occurs on closed and $\alpha$-polar sets $E \subset \s$. Thanks to results of Carleson and further considerations, which particularly rely on the favorable behavior of the $\mathrm{Li}_\alpha$-kernel, we also find that this result is sharp. In particular, it may be seen as a generalization of the universality result for the harmonic Dirichlet space.
Following this, we will investigate the same class, however, this time for $\alpha \in [-1,0)$. In this case, it turns out that universality with respect to uniform convergence occurs on closed and $(-\alpha)$-complement-polar sets $E \subset \s$. In particular, these sets of universality can have positive arc measure. In the final section, we will focus on the class $D_{\textup{harm}}^p$. Here, we manage to prove that universality occurs on closed and $(1,p)$-polar sets $E \subset \s$. Through results of Twomey [68] combined with an observation by Girela and Pélaez [23], as well as the decomposition of $D_{\textup{harm}}^p$, we can deduce that the closed sets of universality with respect to uniform convergence of the class $D_{\textup{harm}}^p$ are characterized by $(1,p)$-polarity. We conclude our work with an application of the latter result to the class $D^p$. We will show that the closed sets of divergence for the class $D^p$ are given by the $(1,p)$-polar sets.
Ensuring fairness in machine learning models is crucial for ethical and unbiased automated decision-making. Classifications from fair machine learning models should not discriminate against sensitive variables such as sexual orientation and ethnicity. However, achieving fairness is complicated by biases inherent in training data, particularly when data is collected through group sampling, like stratified or cluster sampling as often occurs in social surveys. Unlike the standard assumption of independent observations in machine learning, clustered data introduces correlations that can amplify biases, especially when cluster assignment is linked to the target variable.
To address these challenges, this cumulative thesis focuses on developing methods to mitigate unfairness in machine learning models. We propose a fair mixed effects support vector machine algorithm, a Cluster-Regularized Logistic Regression and a fair Generalized Linear Mixed Model based on boosting, all of them are capable of handling both grouped data and fairness constraints simultaneously. Additionally, we introduce a Julia package, FairML.jl, which provides a comprehensive framework for addressing fairness issues. This package offers a preprocessing technique, based on resampling methods, to mitigate biases in the data, as well as a post-processing method, that seeks for a optimal cut-off selection.
To improve fairness in classifications both processes can be incorporated in any classification method available in the MLJ.jl package. Furthermore, FairML.jl incorporates in-processing approaches, such as optimization-based techniques for logistic regression and support vector machine, to directly address fairness during model training in regular and mixed models.
By accounting for data complexities and implementing various fairness-enhancing strategies, our work aims to contribute to the development of more equitable and reliable machine learning models.
This dissertation addresses the measurement and evaluation of the energy and resource efficiency of software systems. Studies show that the environmental impact of Information and Communications Technologies (ICT) is steadily increasing and is already estimated to be responsible for 3 % of the total greenhouse gas (GHG) emissions. Although it is the hardware that consumes natural resources and energy through its production, use, and disposal, software controls the hardware and therefore has a considerable influence on the used capacities. Accordingly, it should also be attributed a share of the environmental impact. To address this softwareinduced impact, the focus is on the continued development of a measurement and assessment model for energy and resource-efficient software. Furthermore, measurement and assessment methods from international research and practitioner communities were compared in order to develop a generic reference model for software resource and energy measurements. The next step was to derive a methodology and to define and operationalize criteria for evaluating and improving the environmental impact of software products. In addition, a key objective is to transfer the developed methodology and models to software systems that cause high consumption or offer optimization potential through economies of scale. These include, e. g., Cyber-Physical Systems (CPS) and mobile apps, as well as applications with high demands on computing power or data volumes, such as distributed systems and especially Artificial Intelligence (AI) systems.
In particular, factors influencing the consumption of software along its life cycle are considered. These factors include the location (cloud, edge, embedded) where the computing and storage services are provided, the role of the stakeholders, application scenarios, the configuration of the systems, the used data, its representation and transmission, or the design of the software architecture. Based on existing literature and previous experiments, distinct use cases were selected that address these factors. Comparative use cases include the implementation of a scenario in different programming languages, using varying algorithms, libraries, data structures, protocols, model topologies, hardware and software setups, etc. From the selection, experimental scenarios were devised for the use cases to compare the methods to be analyzed. During their execution, the energy and resource consumption was measured, and the results were assessed. Subtracting baseline measurements of the hardware setup without the software running from the scenario measurements makes the software-induced consumption measurable and thus transparent. Comparing the scenario measurements with each other allows the identification of the more energyefficient setup for the use case and, in turn, the improvement/optimization of the system as a whole. The calculated metrics were then also structured as indicators in a criteria catalog. These indicators represent empirically determinable variables that provide information about a matter that cannot be measured directly, such as the environmental impact of the software. Together with verification criteria that must be complied with and confirmed by the producers of the software, this creates a model with which the comparability of software systems can be established.
The gained knowledge from the experiments and assessments can then be used to forecast and optimize the energy and resource efficiency of software products. This enables developers, but also students, scientists and all other stakeholders involved in the life cycleof software, to continuously monitor and optimize the impact of their software on energy and resource consumption. The developed models, methods, and criteria were evaluated and validated by the scientific community at conferences and workshops. The central outcomes of this thesis, including a measurement reference model and the criteria catalog, were disseminated in academic journals. Furthermore, the transfer to society has been driven forward, e. g., through the publication of two book chapters, the development and presentation of exemplary best practices at developer conferences, collaboration with industry, and the establishment of the eco-label “Blue Angel” for resource and energy-efficient software products. In the long term, the objective is to effect a change in societal attitudes and ultimately to achieve significant resource savings through economies of scale by applying the methods in the development of software in general and AI systems in particular.
In most textbooks optimal sample allocation is tailored to rather theoretical examples. However, in practice we often face large-scale surveys with conflicting objectives and many restrictions on the quality and cost at population and subpopulation levels. This multiobjectiveness results in a multitude of efficient sample allocations, each giving different weight to a single survey purpose. Additionally, since the input data to the allocation problem often relies on supplementary information derived from estimation, historical data, or expert knowledge, allocations might be inefficient when specified for sampling.
This doctoral thesis presents a framework for optimal allocation to standard sampling schemes that allows for specifying the tradeoff between different objectives and analyzing their sensitivity to other problem components, aiming to support a decision-maker in identifying an at-most preferred sample allocation. It dedicates a full chapter to each of the following core questions: 1) How to efficiently incorporate quality and cost constraints for large-scale surveys, say, for thousands of strata with hundreds of precision and cost constraints? 2) How to handle vector-valued objectives with their components addressing different, possibly conflicting survey purposes? 3) How to consider uncertainty in the input data?
The techniques presented can be used separately or in combination as a general problem-solving framework for constrained multivariate and multidomain, possibly uncertain, sample allocation. The main problem is formulated in a way that highlights the different components of optimal sample allocation and can be taken as a gateway to develop solution strategies to each of the questions above, while shifting the focus between different problem aspects. The first question is addressed through a conic quadratic reformulation, which can be efficiently solved for large problem instances using interior-point methods. Based on this the second question is tackled using a weighted Chebyshev minimization, which provides insight into the sensitivity of the problem and enables a stepwise procedure for considering nonlinear decision functionals. Lastly, uncertainty in the input data is addressed through regularization, chance constraints and robust problem formulations.
Biodiversity is threatened by a wide range of anthropogenic activities. Monitoring offers critical insights into how and why biodiversity is changing, helping to identify effective measures for maintaining biodiversity and its ecosystem services. However, conventional biodiversity monitoring methods are labor-intensive, and standardized long-term monitoring series are scarce. DNA-based approaches like metabarcoding environmental DNA (eDNA) promise rapid, cost-efficient, and highly resolved community data. At the same time, scientists are looking for alternative data sources that can compensate for the lack of long-term monitoring data to study past biodiversity changes. This work explores the potential of the German Environmental Specimen Bank (ESB), a pollution monitoring archive, which appears particularly promising for retrospective biodiversity monitoring. Biota samples from different ecosystems across the country are collected and archived at an exceptional level of standardization. Sampling species act as natural eDNA samplers, accumulating genetic traces from surrounding organisms. The cryogenic storage at the ESB preserves any eDNA in the samples in its original state. In this thesis, Chapter I serves as an introductory chapter, outlining the general chances and challenges of metabarcoding for assessing biodiversity. Chapter II focuses on primer design and testing the utility of ESB sampling species like mussels and macroalgae for characterizing the surrounding community. Both chapters form the basis for Chapters III to V, which report the use of ESB time series to uncover sample-associated communities and the changes they undergo. Chapter III illustrates the value of these time series by revealing the invasion trajectory of an alien barnacle into German coastal waters and linking the process to climate change. Chapter IV forms the core of this thesis by presenting an expanded measurement of biodiversity change in ESB time series across different taxonomic groups and ecosystem types. Here, a gradual compositional change (turnover) is reported from bacterial, fungal, microeukaryotic, and metazoan communities tending to either spatial homogenization or differentiation. Observed trends are tested for significance using a dynamic model of community ecology based on the equilibrium theory of island biogeography. The model reveals significantly accelerated turnover rates across all taxonomic groups and ecosystems investigated, suggesting a common, anthropogenically induced driver of biodiversity change. Since these analyses most likely include DNA derived from dead as well as from living organisms, Chapter V aims to separate both groups by metabarcoding both DNA and less stable ribosomal RNA from mussel samples. Contrary to the hypothesis, RNA is detectable from both living endobionts and dietary taxa. However, it outcompetes DNA in detecting microeukaryotic biodiversity. In summary, this thesis demonstrates the outstanding potential of archived ESB samples for retrospective biodiversity monitoring, a resource that offers many further untapped opportunities for future biodiversity research at multiple scales.
The present dissertation deals with variable stress patterns in English complex adjectives such as celebratory, identifiable or imaginative. This variation is usually described in terms of retaining the stress from the embedded base (idéntify -> idéntifiable) or deviating from the stress of the embedded base (idéntify -> identifíable). While several accounts have explored this variation, none of them have been able to identify a plausible reason for why it occurs. Additionally, the role of individual speaker differences has been disregarded in the discussion. This dissertation therefore explores the empirically observable extent of the variation and investigates possible causes of it with a special focus on individual differences between speakers. It uses data from a complex online experiment that included five different tasks to assess speakers' stress production, perception, morphological processing, vocabulary size and other factors. It furthermore tests the predictions of previous accounts on the large set of authentic utterances from speakers collected using this online experiment. The data show that individual differences in vocabulary size between speakers are a significant predictor of a speaker's tendency to retain the stress of the embedded base.
Biotic communities experienced significant changes in recent decades. Climate change, the overexploitation of natural resources and the immigration of invasive species are major drivers for this change and present unknown challenges for communities worldwide. To assess the impact of these drivers, standardised long-term studies are required, which are currently lacking for many species and ecosystems. Analysing environmental samples and the DNA of associated organisms using metabarcoding and high-throughput sequencing provides a cost-efficient and rapid way to generate the high-resolution biodiversity data which is so direly needed.
In this thesis, I demonstrate the great potential of using samples from the German Environ- mental Specimen Bank (ESB), a long-term monitoring archive that has been collecting and cryogenically storing highly standardised environmental samples since 1985. Modern analytical methods enable retrospective long-term biodiversity monitoring using these samples. In the first chapter, I illustrate metabarcoding as a central method, discussing its strengths and drawbacks, how to avoid them, and new application approaches. This chapter provides the methodological basis for the following studies.
In subsequent chapters, I present time series analyses of communities associated with these environmental samples. While for Chapter two the focus is on terrestrial arthropod communities, in Chapter three aquatic and terrestrial communities across the tree of life are analysed. A null model was developed for this survey for robust conclusions. The studies covered the last three decades and revealed substantial compositional changes across all ecosystems. These changes deviated significantly from the model, indicating that the changes are occurring faster than expected. Moreover, a trend toward homogenization in many terrestrial communities was uncovered. Climate change and the immigration of invasive species in combination with the loss of site-specific species are suspected to be the main drivers for this. In a follow-up study, changes of arthropod communities in German and South Korean terrestrial ecosystems were compared using ESB leaf samples from these two countries. Since both ESBs are harmonised in sample collection and processing, comparative analyses were applicable. This research covered the last decade and revealed substantial declines in species richness in Korea. Abiotic and biotic factors are discussed as potential drivers of these results.
Finally, the possibility of assessing tree health by analysing changes in functional fungal groups using German ESB samples was investigated. The results indicate that increasing infestation of specific functional groups is a proxy for declining tree health, with further analyses planned. In this dissertation, I present the great potential of samples from long-term monitoring archives to conduct retrospective biodiversity trend analyses across the tree of life. As technologies evolve, these samples will help to understand past and predict future ecosystem changes.
Knapp 90 Jahre nach Erscheinen des Buchs von Paul Graindor zu den „Bustes et Statues-Portraits d'Egypte Romaine“ widmet sich mit der vorliegenden Dissertation erstmals wieder eine monographische Studie der marmornen Bildnisplastik der römischen Provinz Aegyptus von ihrer Gründung im Jahr 30 v. Chr. bis zum Ende des 3. Jhs. n. Chr. Basierend auf einer umfassenden Zusammenstellung bekannter, aber auch bislang unpublizierter Portraits sowie einer Neudokumentation zahlreicher Objekte gelingt erstmalig eine belastbare chronologische und typologische Auswertung dieser Bildnisse. Zwar bilden dabei die Darstellungen aus weißem Marmor die zentrale und auch quantitativ bei weitem größte Materialgruppe, doch es finden auch Bildnisse aus anderen Werkstoffen wie Bronze, Kalkstein, Gips oder Alabaster Berücksichtigung. Da die Provinz aufgrund geringer eigener Marmorvorkommen fast ausschließlich auf Importe angewiesen war, sind die Marmorbildnisse ein exzellentes Forschungsobjekt, um nicht nur den Handel von Marmor nach Ägypten und seine Distribution und Weiterverarbeitung in der Provinz zu untersuchen, sondern auch damit verbundene handwerkliche Besonderheiten, wie die häufig zu beobachtenden Ergänzungen mit Stuck- oder Steinelementen. Darüber hinaus werden auch Überlegungen zur Semantik des Materials sowie der Herkunft und dem Selbstverständnis der dargestellten Personen angestellt.
Einige Forschungsergebnisse zeigen, dass emotionale Empfindungen kognitive Bereiche beeinflussen oder mit diesen im Zusammenhang stehen. Aufbauend auf den Ergebnissen wurden zwei Studien konzipiert. In Studie 1 wurde der Zusammenhang zwischen den Valenzen der dispositionalen emotionalen Empfindungen und der globalen Selbstbewertung des Gedächtnisses (Metagedächtnis) bei Lehramtsstudierenden (N = 218) untersucht. Die dispositionalen Empfindungen wurden mittels des deutschen Positive and Negativ Affect Schedule (PANAS) (Krohne, Egloff, Kohlmann & Tausch, 1996) und die globale Selbstbewertung des Gedächtnisses mit dem deutschen Squire Subjective Memory Questionnaire (SSMQ) (Wolf, 2017) erfasst. Angenommen wurde, dass die positive Valenz im Gegensatz zu der negativen Valenz im positiven Zusammenhang mit der höheren Gedächtniseinschätzung stehen. Die Ergebnisse bestätigen die Hypothesen. In Studie 2 wurde die aktuelle Valenz mittels des Open Affective Standardized Image Set (OASIS) (Kurdi, Lozano & Banaji, 2017) induziert, um Veränderungen des Metagedächtnisses und der tatsächlichen Gedächtnisleistung bei Lehramtsstudierenden (N = 44) zu untersuchen. Angenommen wurde, dass die positive Valenz positiv, die negative Valenz negativ und die neutrale Valenz nicht auf das Metagedächtnis und die Gedächtnisleistung wirkt. Weitere Zusammenhänge zwischen dem Metagedächtnis und der Gedächtnisleistung sowie der induzierten Valenz und der Gedächtnisleistung wurden angenommen. Die Messinstrumente aus Studie 1 blieben dieselben. Die Gedächtnisleistung wurde mittels eines sinnarmen Silbentests nach Ebbinghaus (1885) operationalisiert. Die Ergebnisse bestätigen die Hypothesen nicht. Die Emotionsinduktion hatte keinen Erfolg. Die Ergebnisse können damit nicht auf eine veränderte Valenz bezogen werden. Wie in Studie 1 zeigte sich ein Zusammenhang zwischen den dispositionalen Empfindungen und dem Metagedächtnis. Weitere explorative Ergebnisse, vor allem im Bezug auf das Geschlecht, wurden dargestellt. Die Ergebnisse sind bedeutsam für die Professionalisierung von Lehramtsstudierenden.
Extracellular enzymes in microbial communities play a central role in nutrient cycling and the degradation of (pollutant) substances in various natural and anthropogenic systems. Bound in aquatic biofilms, sludge aggregates, or even unbound at their interfaces, they are of great importance for both the environment and human health. In particular, in wastewater treatment plants and inland waters, hydrolytic activities influence the wide-reaching efficiency of nutrient removal and self-purification, thus contributing significantly to overall water quality.
The main goal of this dissertation project was to investigate the factors that influence enzymatic activity and the health of microbial communities in activated sludge and river systems, particularly in relation to anthropogenic influences and natural environmental conditions. The aim was to contribute to a better understanding of the sensitivity of our freshwater ecosystems and to support the long-term preservation of water quality and ecological stability. The development and optimization of appropriate methods, as well as their testing and applicability, were the focal points.
For this purpose, a fluorometric microplate assay was developed and adapted to determine both extracellular enzyme activities (EEAs) in activated sludge samples and in intact biofilms. Its suitability for field studies was subsequently tested. Inhibition and activity of selected hydrolases under different conditions were investigated to better understand the mechanisms and potential environmental risks posed by anthropogenic influences and seasonal fluctuations of hydrochemical and climatic parameters.
The first phase of the doctoral thesis involved studies on the inhibition of alkaline phosphatase in activated sludge by oxyanions. Using the fluorometric microplate assay, the inhibitory effect was sensitively detected over a pH range of 7.0 to 8.5. IC50- and IC20-concentrations were calculated from modeled dose-response functions. It was found that vanadate and tungstate caused strong inhibitory effects, while molybdate moderately inhibited the enzyme. An increasing pH led to a reduction in the inhibitory effect of tungstate and molybdate. The inhibition effects of vanadate were not significantly affected by the pH. In municipal wastewater, the concentrations of such metal ions are usually low, but industrial wastewater may have pollutant loads that can significantly impact the removal of phosphorus-containing compounds, and thus the efficiency of treatment plants.
In the second phase, an attempt was made to further adapt the developed methodology to investigate EEA and kinetics in intact freshwater biofilms. Four different types of bead materials (lava, glass, sintered quartz, and ceramics) fitting into a 96-well microplate were tested as carriers for biofilms on both the laboratory and field scale. The analysis included a total of seven hydrolases as representatives of key nutrient cycles such as phosphorus, carbon, and nitrogen: phosphatases, glucosidases, peptidases (two different types), and sulfatase. Experiments with increasing substrate concentrations led to classical kinetic profiles according to the Michaelis-Menten mechanism. This allowed for the prediction of the biofilm enzymes’ response to different substrate concentrations. Parameters such as Vmax and Km could be derived from the modeled curves.
Ceramic beads are particularly suitable for long-term studies due to their high stability, while sintered quartz beads should be preferred for the use in stagnant media (material loss under turbulent conditions). Lava and glass beads, on the other and, proved suboptimal for uniform biofilm development due to their surface properties. The potential use of this fast and sensitive test for ecotoxicological or even human-toxicological studies was demonstrated by the effects of caffeine on the activity of PDE. The result of this part of the research represents a powerful tool for assessing environmental pollution and monitoring water quality.
The high application potential was clearly highlighted in the final phase of the project. The goal here was to deepen the understanding of interactions between seasonal factors, anthropogenic influences, and biofilm processes in rivers by investigating EEA and biofilm parameters such as biomass and relating them to hydrochemical and climatic factors. Ceramic beads were exposed both upstream and downstream of a wastewater treatment plant discharge and sampled over a period of seven months. EEAs and biomass varied depending on the season and location, with higher microbial activity observed upstream in winter. Winter conditions led to the dilution of most nutrients as well as in an increse of dissolved oxygen. Nutrient concentrations analyzed downstream were significantly higher in the summer. Accumulation of nutrient or pollutants during the summer months cannot be excluded, which may have led to a general reduction in enzyme activities.
Potential causes could be inhibitory effects on the enzymes, or a reduced enzyme activity due to a sufficiently high nutrient supply. In general, the sampling site upstream showed a more pronounced seasonal dynamics, with a significant proportion of the variance in biological parameters (activity and biomass) attributable to seasonal factors. A secondary component, likely reflecting the impact of the treatment plant discharge, explained another portion of the data variance. Regardless of the season, high correlations between biological parameters were observed upstream, while downstream the data were more decorrelated. This could be because the biofilms, under chronic stress, respond less dynamically to seasonal fluctuations.
This dissertation illustrates that in addition to anthropogenic stress factors, seasonal fluctuations of hydrochemical and climatic parameters should also be considered in "stress downstream the pipe" studies. The selected methods are recommended for explaining and considering the data variance, as they highlight the complex interplay between microbial enzymatic activity, environmental factors, and pollutants in the activated sludge of wastewater treatment plants and also in aquatic systems. The novel bead assay could pave the way for the future standardization of effect-oriented studies on intact aquatic biofilms.
Measuring the economic activity of a country requires high-quality data of businesses. In the case of Germany, this is not only required at national level, but also at federal state level and for different economic sectors. Important sources for high-quality business data are the business register and, among others, also 14 business surveys which are conducted by the Federal Statistical Office of Germany. However, the quality requirements of the Federal Statistical Office are in contrast to the interests of the businesses themselves. For them, answering to a survey's questionnaire is an additional cost factor, also known as response burden. A high response burden should be avoided, since it can have a negative impact on the quality of the businesses' responses to the surveys. Therefore, sample coordination can be used as a method to control the distribution of response burden while securing high-quality data.
When applying already existing business survey coordination systems, developed by different statistical institutes, legal and administrative standards of German official statistics have to be taken into account. These standards consider different sampling fractions, rotation fractions, periodicity, and stratification of the aforementioned 14 business surveys. Therefore, the aim of this doctoral thesis is to check the existing business survey coordination systems for their applicability in the context of German official statistics and, if necessary, to modify them accordingly. These modifications include the introduction of individual burden indicators which aim to take the individual perception of response burden into account.
For this purpose, several synthetic data sets have been created to test the application of the modified versions of the different business survey coordination systems through Monte Carlo simulation studies. These data sets include a large panel data set, reflecting the landscape of businesses in Rhineland-Palatinate and three smaller, synthetic data sets. The latter have been created with the help of the R package BuSuCo which has been developed within the scope of this thesis. The above mentioned simulation studies are evaluated based on different measures for estimation quality as well as for the concentration and distribution of response burden.
Modellierung von o-PO4- Einträgen in saarländische Oberflächenwasserkörper im Trockenwetterfall
(2025)
Die Verfügbarkeit von ortho-Phosphat (o-PO₄) trägt wesentlich zur Eutrophierung von Fließgewässern bei und gefährdet damit das Erreichen des „guten ökologischen Zustands“ gemäß der EU-Wasserrahmenrichtlinie. Da die kommunalen Kläranlagen zentrale Eintragsquellen darstellen, gewinnt die Reduktion von o-PO₄ an dieser Stelle an Bedeutung. Neben der chemischen Phosphorelimination bietet insbesondere die vierte Reinigungsstufe, primär zur Entfernung von Mikroschadstoffen konzipiert, einen Synergieeffekt mit potenziellen Phosphorentfernungsraten von bis zu 85 %.
Zur Bewertung des Einflusses einer solchen Reinigungsstufe wurde ein Modell für ausgewählte saarländische Oberflächenwasserkörper (OWK) entwickelt, das den Trockenwetterfall als eutrophierungsrelevantes Szenario abbildet. Ein zentraler Bestandteil ist ein neu erarbeiteter Retentionsansatz, der biochemische und physikalische Prozesse wie Adsorption, Sedimentation und biologische Assimilation berücksichtigt. Auf Basis der Differenz zwischen emissionsseitig bilanziertem und gemessenem o-PO₄-Gehalt wurden für jeden OWK Verminderungsraten je Fließmeter abgeleitet und schließlich eine Gleichung zur Abschätzung der Retention in Abhängigkeit der Einzugsgebietsgröße formuliert. Die Validierung zeigt hinreichende Modellgenauigkeit, wenngleich negative Frachtdifferenzen in einigen Gewässern auf zusätzliche, nicht eindeutig quantifizierbare Einträge – etwa aus Landwirtschaft oder Kanalverlusten – hindeuten.
Die Szenarienanalyse belegt, dass eine vierte Reinigungsstufe grundsätzlich zur Reduktion von o-PO₄ an den Messstellen beiträgt. Eine Unterschreitung des geltenden Orientierungswertes wird jedoch nur erreicht, wenn sämtliche Kläranlagen eines OWK nachgerüstet werden – und auch dann nur in einigen Fällen. Damit stellt die vierte Reinigungsstufe allein keine ausreichende Alternative zu den Maßnahmen des 3. Bewirtschaftungsplans des Saarlandes dar, kann jedoch als ergänzende Strategie zur Verringerung der Phosphoreinträge dienen.
The application of machine learning and deep learning methods to hydrological modelling has advanced significantly in recent years, offering alternatives to traditional conceptual and physically based approaches. Within the numerous algorithms, long short-term memory (LSTM) networks have proven themselves particularly useful for the task of streamflow modelling. This thesis provides a collection of publications that investigate the capabilities, limitations and interpretability of LSTM for the purpose of streamflow modelling and climate change impact assessment within the lowland Ems catchment in Northwest Germany.
Within a comparative performance evaluation, LSTM and its predecessor, the recurrent neural network, demonstrate superior accuracy compared to the conceptual HBV model across various statistical performance metrics. However, a decline in performance was observed during low-flow conditions in certain sub-catchments. The evaluation of the flow duration curve revealed that the ML models more effectively capture the water balance, while HBV better represents streamflow dynamics.
To enhance the interpretability of LSTM, six explainable artificial intelligence techniques were applied. These methods consistently identified seasonal patterns in the temporal relevance of hydroclimatic input data. In combination with an observed correlation between the internal LSTM states and catchment-scale soil moisture dynamics, the findings suggest that LSTM models are capable of implicitly learning the relevant hydrological processes.
Following, the capabilities of LSTM to model climate change impact scenarios, particularly when they extend beyond historically observed climate conditions, are addressed. An ensemble of climate change projections is provided as hydroclimatic input to evaluate the performance of LSTMs and conceptual models. While all models reveal heterogeneous alterations in streamflow under future climate conditions, significant differences emerge based on the model type. Results provide evidence that LSTMs, in combination with the temperature-based Haude formula for estimating potential evaporation, work inadequately under altered climatic regimes, raising concerns about their applicability in long-term projections. The study also indicates the potential need to incorporate physical constraints into LSTM architectures to ensure model robustness and hydrological plausibility beyond the historical training range.
Collectively, this thesis contributes important insights into the applicability and interpretability of LSTM models in streamflow modelling. Despite the presence of a physically realistic representation of soil moisture dynamics of the Ems catchment, no robust change signals for streamflow under climate change can be derived. Those results underscore the potential of LSTM model approaches for accurate streamflow simulation, however, they require us to always critically question LSTM results, particularly when they are applied outside the training range.
Bilevel problems are optimization problems for which parts of the variables
are constrained to be an optimal solution to another nested optimization
problem. This structure renders bilevel problems particularly well-suited for
modeling hierarchical decision-making processes. They are widely applicable
in areas such as energy markets, transportation systems, security planning,
and pricing. However, the hierarchical nature of these problems also makes
them inherently challenging to solve, both in theory and in practice.
In this thesis, we study different nonlinear problem settings for the
nested optimization problem. First, we focus on nonlinear but convex bilevel
problems with purely integer variables. We propose a solution algorithm that
uses a branch-and-cut framework with tailored cutting planes. We prove
correctness and finite termination of the method under suitable assumptions
and put it into context of existing literature. Moreover, we provide an
extensive numerical study to showcase the applicability of our method and
we compare it to the state-of-the-art approach for a less general setting on
suitable instances from the literature. Furthermore, we discuss challenges that
arise when we try to generalize our approach to the mixed-integer setting.
Next, we study mixed-integer bilevel problems for which the nested
problem has a nonconvex and quadratic objective function, linear constraints,
and continuous variables. We state and prove a complexity-theoretical hardness result for this
problem class and develop a lower and upper bounding scheme to solve
these problems. We prove correctness and finite termination of the proposed
method under suitable assumptions and test its applicability in a numerical
study.
Finally, we consider bilevel problems with continuous variables, where
the nested problem has a convex-quadratic objective function and linear
constraints. We reformulate them as single-level optimization problems using
necessary and sufficient optimality conditions for the nested problem. Then,
we explore the family of so-called P-split reformulations for this single-level
problem and test their applicability in a preliminary numerical study.
Spatial microsimulation is an important tool for integrating geographical information into the evaluation of public policies and the analysis of social phenomena in urban regions. These models simulate the behavior and interaction between units of the region, such as individuals, households or firms, under specific conditions that may or not involve projections over time. This requires a representative base data set for their respective units.
In this thesis, we focus on the geo-referencing step of the population in the construction of this data set, where we define the location of the individuals so that the allocation obtained is representative in relation to the population of the region. To do this, we consider the assignment of households to dwellings with specific coordinates by solving a maximum weight matching problem where side constraints are included so that the allocation obtained satisfies statistical structures intrinsic to the considered region.
The model of this problem represents each feasible assignment of household to dwelling as a binary variable, which results in billions of variables for medium-sized municipalities such as the city of Trier, Germany. Therefore, standard solvers for mixed-integer linear optimization are not able to solve it due to their high time and memory consumption. Hence, we develop two approaches capable of producing high-quality allocations using a reasonable amount of computational resources, one based on specific decomposition algorithms, and the other characterized by the application of an approximation algorithm in the framework of Lagrangian relaxation of the side constraints.
We theoretically explore the allocations obtained by both approaches and perform an extensive computational study using synthetic data sets and real-world data sets associated with the city of Trier. The results show that the developed methods are able to obtain near-optimal solutions using significantly less memory and time than the solver Gurobi, which enables them to tackle significantly larger instances, with approximately 100 000 households and dwellings. Furthermore, the allocations obtained for the real-world data sets correspond to a realistic population distribution, which strengthens the practical applicability of our methods.
Building on Social Virtual Reality to Support Flexible Collaboration and Enrich Therapy Sessions
(2025)
Social virtual environments allow their users to meet and collaborate in a shared three-dimensional space, even when far apart from each other in the real world. Within these spaces, the appearance and interaction capabilities of both users and environments can be adapted and changed in a myriad of ways. To enable virtual environments to fulfill their potential of supporting a wide variety of collaboration use-cases, both the impacts of basic interaction design decisions and the individual needs of specific usage areas need to be explored further.
This thesis approaches this topic in two ways. First, the basic building blocks of collaboration in social virtual environments are explored by asking the question: "How can social virtual spaces that allow interaction beyond real-world constraints utilize the potential of mutual assistance and shared workflows between multiple users?". Going into further detail for a serious use-case in which direct collaborative interactions and their effect on the included users are especially important, it then explores the potential of collaborative virtual spaces in the therapy domain by asking "How can the potential of social virtual spaces be utilized to support and improve therapy encounters?"
With regards to the first research question, the thesis presents two theoretical frameworks detailing different aspects of supporting smooth and varied collaboration processes. In addition, several user studies on the topic of collaborative virtual interaction are described, focusing on the role that different users can play during shared interaction and the effects that this distribution of roles and responsibilities has on both the performance and experience of the involved user pairs.
The results presented for this first research question show that social virtual spaces have the potential to provide dedicated support for collaborative workflows. To enable users to adapt their working mode individually and as a team, interaction techniques should complement a team's natural interaction and communication. When presenting novel interactions to users, providing them with a way to support each other can ease their adaptation to these interactions. In these cases, the inclusion of all interested collaborators as active participators should be prioritized in order to let all users benefit from being immersed in a virtual environment.
Addressing the combination of social virtual spaces with therapy in relation to the second research question, this thesis presents the result of a series of interviews with practicing physio- and psychotherapists. Motivated by the recorded expert feedback, it also reports on two more detailed explorations of specific areas of interest. The work presented for the second research question demonstrated the promise of using virtual environments in both exercise- and conversation-based therapy practice. Investigating the potential of shared interactions, the exploration of virtual recordings and the adaptation of virtual appearances, the presented work uncovered several topic areas that could be further explored regarding their possible use in the treatment of patients.
Taken together, the six research articles presented in this thesis show both the value of supporting and understanding shared interactions in virtual spaces and their potential place in serious use-cases like the therapy domain. When introducing shared virtual environments to new user groups, the opportunity for mutual support through shared interaction techniques could be a crucial building block towards making virtual spaces both accessible and attractive to a variety of users.
Perennial crops eliminate soil disturbance and reduce the amount of synthetic chemicals that are applied to the soil, improving soil biodiversity and food web structure. Additionally, perennial cropping is characterised by all year-round surface coverage which benefits soil biota in terms of habitat and food sources. Perennial intermediate wheatgrass (Thinopyrum intermedium, IWG) was domesticated and commercialised by The Land Institute in Kansas as Kernza® and serves as an example for these nature-based solutions. It develops an extensive root system that has a higher nutrient retention, possibly reducing nutrient runoff. It thereby follows a more resource-conservative strategy with improved belowground-oriented resource allocation in its root system. This may reduce the need for excessive fertiliser as the crop has a higher nitrogen efficiency, among other things.
IWG promoted the earthworm community and its diversity, more specifically, the occurrence of epigeic species (litter inhabitants), since those species benefit from the increased soil coverage and elimination of disturbances in the soil. As IWG creates a dense and extensive root system, as shown by the increased occurrence of root-feeding nematodes, endogeic species (horizontal burrowers) are supported through the provision of a reliable food source. IWG was characterised as a mostly undisturbed system with a highly structured food web through nematode analysis, as expressed through the promotion of structure indicators, for example, that are sensitive to disturbances in the soil and are therefore supported under no-till management. The root microbiome is continuously being shaped by the host as the crop regrows from the roots each vegetation period. This creates a symbiotic relationship and a beneficial feedback loop for the crop. Resultantly, the root-endophytic microbiome under IWG had a higher network complexity, connectivity and stability compared to annual wheat. The regrowth from the roots for IWG requires increased nutrient and energy storage, which was indicated by increased starch values. Correspondingly, the longer residence time of the roots in the soil resulted in higher lignin values. Furthermore, the decomposition pathway was dominated by fungivorous nematodes which may correspond to stimulated nutrient cycling and a heterogeneous resource environment, as seen for low input systems.
Overall, perennial wheat cultivation improved soil biodiversity already after an establishment of 3-6 years. As those benefits were present for all three countries, the varying soil and climate conditions do not seem to interfere with the positive effect of perennial wheat on the soil ecosystem, demonstrating a wide transferability and adaptability of the crop onto other study sites as well. Enhanced complexity and connectivity of the food web in comparison to annual wheat may indicate a resistance against abiotic stress, suggesting IWG cultivation as a viable option for a sustainable and resilient agriculture. The improvement in nutrient cycling and the resource-efficient cultivation strategy for IWG could enable cultivation on marginal land where annual crop cultivation is not possible as the soils are susceptible to erosion and nutrient runoff. This opens up new possibilities for agricultural cultivation on previously unused land, thus contributing to food security in the future.
Three-Point Difference Schemes of High Order of Accuracy for Solving the Sturm-Liouville Problem
(2025)
The dissertation is devoted to the construction and justification of three-point difference schemes of high order of accuracy for solving the Sturm-Liouville problem. A new algorithmic realization of the exact three-point difference scheme on a non-uniform grid has been developed. We show that to compute the coefficients of the exact scheme in an arbitrary grid node, it is necessary to solve two auxiliary Cauchy problems for the system of three linear ordinary differential equations of the first order. The coefficient stability of the exact three-point difference scheme is proved. If the Cauchy problems are solved numerically using any one-step method, we obtain the truncated three-point difference scheme. The accuracy estimate of three-point difference schemes was obtained and the algorithm for finding their solution was developed.
We also developed a new algorithmic realization of the exact three-point difference scheme for the Sturm-Liouville problem with singularities at the ends of the interval. As in the case of the classical Sturm-Liouville problem, to find the coefficients of the exact three-point difference scheme, it is necessary to solve two auxiliary Cauchy problems for each grid node. The coefficient stability of the exact three-point difference scheme is proved. Since the Cauchy problems for the first and last grid nodes are singular, the Taylor series method has been developed to solve them. The accuracy estimate of truncated three-point difference schemes was obtained. To solve the difference scheme, the Newton's iterative method is used.
Numerical experiments are presented which confirm the efficiency of the proposed approach.
In Vielfalt geeint? Europäische Identitätskonstruktionen im bundesdeutschen Diskurs seit 1990
(2025)
Die Arbeit untersucht den bundesdeutschen Diskurs zur europäischen Integration seit 1990 aus diskurslinguistischer Perspektive und versteht ihn als Aushandlungsraum europäischer Identitätskonstruktionen. Ausgangspunkt ist die Annahme, dass institutionelle Vertiefung und geografische Erweiterung der EU nicht allein als verrechtlichte Integrationsschritte zu begreifen sind, sondern stets auch identitätspolitische Dimensionen tragen. Ziel der Studie ist es, die sprachliche Konstituierung der EU als identitätspolitisches Referenzsystem sichtbar zu machen und damit eine diskurslinguistische Ergänzung zur interdisziplinären Integrationsforschung zu leisten. Auf Grundlage eines diachronen Korpus, das zentrale integrationspolitische Etappen und Krisenphasen umfasst, wird ein Mixed-Methods-Ansatz entwickelt, der korpusgeleitete Verfahren mit der hermeneutischen Annotation diskurslinguistischer Kategorien verbindet. Analysiert werden nicht nur lexikalisch-semantische Repräsentationen Europas, sondern vor allem diskursive Grundfiguren wie Einheit, Vielfalt, Eigenes und Fremdes sowie deren Verbindung zu politischen Sinnzuschreibungen. Die Ergebnisse zeigen, in welchem Maße sich im deutschen Diskurs ein stabiler identitätspolitischer Bezugspunkt zur EU herausgebildet hat, wie sich normative Leitbilder und funktionale Rationalitäten überlagern und wie europäische Integration sprachlich zwischen symbolischer Aufladung und strategischer Instrumentalisierung verhandelt wird.
Income composition can have a significant impact on workers’ well-being, productivity, and career paths. Wages often include a variety of components, such as unconditional bonuses, profit-sharing payments, and incentives based on the individual performance of employees. Each of these may influence employee labour outcomes differently and the worker composition may matter for managers when designing the salary package. Simultaneously, workers’ employment choices and well-being are influenced by income outside the job, such as inheritances and lottery winnings, as well as by external factors like technological change. This dissertation includes five empirical studies that investigate these issues, yielding new insights on the role of monetary gifts, financial incentives, labour market institutions, and technology disruptions in affecting employees’ labour and well-being outcomes.
Many developed countries, including Germany, face a steady rise in the share of
individuals obtaining higher education. While rising education itself bears a series
of advantages as extensively studied in previous literature, it is also conceptually
linked to a higher likelihood of working in an occupation that does not match
one’s formal qualifications. Previous studies have predominantly evaluated
how demographic or job‐related aspects correlate with the likelihood of being
educationally ﴾mis﴿matched. However, they have largely ignored institutional
facets of the educational system or industrial organization. Moreover, little is
known about how private wealth affects educational mismatch or whether job
satisfaction is homogenously affected among individuals once such a mismatch
occurs. The five projects collected in this thesis aim to answer these open
questions in the literature for Germany, using data from the Socio‐Economic Panel
and employing different time intervals between 1984 and 2022.
Beginning with the educational system in early childhood, Chapter 2 evaluates
the impact of school‐starting age on the likelihood of over‐ and undereducation.
It exploits the exogenous variation in school‐entry rules across federal states
and years in Germany with regression discontinuity designs. The results report
a negative impact of school‐starting age on the likelihood of undereducation,
but no systematic relationship with overeducation.
Subsequently, Chapter 3 explores the variation in education costs by leveraging
the quasi‐experimental setting induced by the time‐limited introduction of tuition
fees in several German federal states between 2006 and 2014. The increase
in education costs among treated graduates results in a significantly higher
likelihood of overeducation, which endures even several years post‐graduation.
Chapter 4 focuses on the industrial relations system and examines the
correlation between trade union membership and the likelihood and extent of
educational ﴾mis﴿match. The results reveal that trade union members report
significantly less overeducation at both the intensive and extensive margin
and also a higher likelihood of being matched compared to non‐members. Furthermore, the heterogeneity analysis provides evidence that this correlation
is driven by improved bargaining power instead of informational advantages.
Chapter 5 focuses on private wealth as a determinant of educational mismatch
by investigating the impact of a wealth shock through inheritances, lottery
winnings or gifts on the likelihood of over‐ and undereducation. Due to
the diminishing marginal returns of wages with increasing windfall gains the
likelihood of undereducation is expected to decrease, while that of overeducation
is expected to increase. Empirically, these suppositions are supported for
overeducation, as its likelihood increases significantly after the windfall gain.
Further analyses reveal that this effect is driven by individuals switching
occupations while increasing their leisure time, and it materializes only for
medium to large windfall gains.
Contrary to the previous chapters, Chapter 6 focuses on educational mismatch,
more precisely on overeducation, as the independent variable. In particular, it
investigates the correlation between overeducation and job satisfaction. The
results align with the previously established negative correlation for private sector
employees exclusively. In contrast, interaction and subsample analyses reveal a
positive correlation for public sector employees. This link is driven by individuals
with a high degree of altruistic motivation and family orientation.
This dissertation examines how individuals unlock their personal power by investigating individual differences in self-regulation, in particular, how situational conditions interact with the personality dispositions of action versus state orientation. Action-oriented individuals are well able to regulate their affective states and to bridge the intention–behavior gap, showing initiative, implementing demanding intentions, and resisting temptations. State-oriented individuals, by contrast, often struggle to regulate affect and experience difficulties enacting intentions, especially under demanding conditions, tending to hesitate and ruminate. While extensive research has highlighted the advantages of action orientation across various domains such as education and health, this thesis challenges the prevailing one-sided perspective that presents action orientation as inherently superior and frames state orientation negatively. Drawing on Personality Systems Interactions theory, the dissertation adopts a dynamic view that understands these dispositions as context-sensitive rather than fixed. The central assumption is that action and state orientation each require different kinds of situational conditions to fully unlock their potential. Across six empirical studies (overall N = 1,067) using a multimethod approach that combines experimental and survey-based research in diverse populations and contextual settings, this dissertation examines (1) action and state orientation as distinct dispositions, (2) their dynamic interaction with situational factors, and (3) ways to support each in mobilizing personal power. Overall, the findings show that each disposition offers unique advantages - they simply require different situational conditions for their potential to unfold.
The role of implicit motives for affective, cognitive and behavioral processes has been a focal part of psychological research for decades. Yet, the majority of research in this field has been concentrated on processes involving implicit motives in adulthood. The systematic investigation of developmental correlates of implicit motives remains largely uncharted. The studies cumulated in this thesis aim to add to the sparse research on implicit motives in childhood and adolescence. Specifically, the development of the implicit power motive in the transition of middle to late childhood as a function of parenting behavior (Chapter 4), the predictive value of the implicit achievement motive for objective swimming performance in children and adolescents (Chapter 5) and the role of motive congruence for successful goal realization in adolescent samples across two cultures (Chapter 6) were investigated. Results of Study 1 (Chapter 4) indicate a negative longitudinal association of authoritarian parenting with the implicit power motive in children that is moderated by children’s perception of psychologically controlling parenting. Study 2 (Chapter 5) extends existing research on the assumed positive association of the implicit achievement motive and sports performance and demonstrates the moderating role of competitive anxiety on this association. Finally, Study 3 (Chapter 6) illustrates a moderating effect of implicit motives on the association of goal commitment and successful goal realization in German and Zambian adolescents, however, this effect was only observed in the domain of power motivation. Findings from all three studies are discussed in the context of the significance of implicit motives for psychological research.
Das Thema des Erlebnisses steht bereits seit langem im Fokus von Anbietern von Dienstleistungen. Dies gilt insbesondere für den Tourismus, einer Branche, deren Produkte zu einem signifikanten Teil aus solchen bestehen. Entsprechend der Prominenz des Themas, vor allem in den Bereichen touristischer Produktentwicklung und Marketing, ist dieses bereits breit in der Forschung diskutiert worden.
Trotz ausgiebiger Publikationsaktivitäten ist der tatsächliche Wissensstand in diesem Thema dennoch auffällig gering. Ein wichtiges Problem liegt darin begründet, dass die Terminologie im Bereich von Erlebnissen noch nicht allgemein akzeptiert und scharf abgegrenzt ist. So muss zwischen Erlebnissen und Erfahrungen unterschieden werden. Erstere treten während des Prozesses der Wahrnehmung einer touristischen Dienstleistung auf und bilden die Basis für Erfahrungen, welche prägend hinsichtlich der Wahrnehmung wirken und im Gesamtkontext der Reise betrachtet werden. Dieser Unterscheidung wird nicht nur in der englischsprachigen Literatur, in der beide Begriffe mit dem Begriff Experience beschrieben werden, sondern auch in der deutschsprachigen Literatur zumeist zu wenig Rechnung getragen, was dazu führt, dass häufig zu Erlebnissen publiziert wird, obwohl eigentlich Erfahrungen beschrieben werden. Problematisch ist dies vor allem, weil damit ein Phänomen untersucht wird, dessen Basis nahezu gänzlich unbekannt ist. Wichtige Fragen, welche zum Verständnis von Erlebnissen und damit auch von Erfahrungen bleiben unbeantwortet:
1) Welche Faktoren werden in der Genese von Erlebnissen wirksam?
2) Wie wirken diese zusammen?
3) Wie wird die Stärke eines Erlebnisses determiniert?
4) Wie werden Erlebnisse stark genug um den Konsum einer touristischen Dienstleistung zu prägen und damit gegebenenfalls zu Erfahrungen zu werden?
In der vorliegenden Arbeit wurden diese Fragen beantwortet, womit ein erster Schritt in Richtung der Füllung einer für die Tourismuswissenschaft nicht unbedeutenden Forschungslücke gelungen ist.
Um Erlebnisse, den Prozess der Genese dieser und deren Bewertung durch den Gast verstehen zu können, wurde ein triangulierter, zweistufiger Forschungsprozess ersonnen und in einem naturtouristischen Setting im Nationalpark Vorpommersche Boddenlandschaft zur Anwendung gebracht. Es handelt sich dabei um einen Mixed-Methods-Ansatz:
1) Induktive-qualitative Studie auf Basis der Grounded Theory
a. Ziel: Identifikation von Wirkkomponenten und deren Zusammenspiel und Generierung eines Modells
b. Methoden: Verdeckte Beobachtung und narrative Interviews
c. Ergebnisse: Modelle der Genese punktueller Erlebnisse und prägender Erlebnisse
2) Deduktive-quantitative Studie
a. Ziel: Überprüfung und Konkretisierung der in 1) generierten Modelle
b. Methoden: Fragebogengestützte, quantitative Befragung und Auswertung mittels multivariater Verfahren
c. Ergebnisse: Zusammenfassung der beiden Modelle zu einem finalen Modell der Erlebnis- und Erfahrungsgenese
Das Ergebnis des Vorgehens ist ein empirisch erarbeitetes und validiertes, detailliertes Modell der Genese von Erlebnissen und der Bewertung dieser durch den Erlebenden in Bezug auf deren Fähigkeit zu Erfahrungen zu werden.
Neben der Aufarbeitung und Konkretisierung dieses Prozesses konnte zusätzlich die in viele Richtungen diskutierte Bedeutung von Erwartungen und Produktzufriedenheit mit Blick auf die Bewertung von Erlebnissen geklärt werden. So konnte empirisch nachgewiesen werden, dass Erlebnisse, die auf Überraschungen, dem Unerwarteten, basierten besonders resistent gegenüber Störfaktoren waren und positive Erlebnisse zwar durchaus im Zusammenhang mit Produktzufriedenheit stehen aber sich vor allem durch eine zumindest temporär gesteigerte Lebenszufriedenheit manifestieren. Damit konnte das Hauptkriterium für die Bewertung von Erlebnissen mit Blick auf ihre Tauglichkeit zu Erfahrungen identifiziert werden.
Für die weitere Forschung kann die vorliegende Arbeit mit dem finalen Modell der Erlebnis- und Erfahrungsgenese einen soliden Ausgangspunkt bilden. So bieten zahlreiche Faktoren im Modell die Möglichkeit zur weiteren Forschung. Auch sollten die Ergebnisse in weiteren touristischen Kontexten überprüft werden.
Für die touristische Praxis kann die vorliegende Arbeit zahlreiche Hinweise geben. So bedeutet die Generierung von Erlebnissen im touristischen Kontext mehr als nur die Erfüllung von Erwartungen. Die widerstandsfähigsten Erlebnisse sind jene, die den Gast zu überraschen vermögen. Ein qualitativ hochwertiges, den Gast zufriedenstellendes Produkt ist dabei nicht mehr als ein Basisfaktor. Wirklich erfolgreich ist ein erlebnisbasierender Ansatz nur dann, wenn dieser es vermag die Lebenszufriedenheit des Gastes zu steigern.