Universitätspublikationen
Refine
Year of publication
Document Type
- Article (10808)
- Doctoral Thesis (1566)
- Preprint (1554)
- Working Paper (1438)
- Part of Periodical (564)
- Conference Proceeding (511)
- Report (299)
- Part of a Book (107)
- Review (92)
- Book (60)
Language
- English (17088) (remove)
Has Fulltext
- yes (17088) (remove)
Keywords
- inflammation (92)
- COVID-19 (89)
- SARS-CoV-2 (62)
- Financial Institutions (47)
- Germany (45)
- climate change (45)
- aging (43)
- ECB (42)
- cancer (42)
- apoptosis (41)
Institute
- Medizin (5096)
- Physik (2984)
- Wirtschaftswissenschaften (1643)
- Frankfurt Institute for Advanced Studies (FIAS) (1575)
- Biowissenschaften (1399)
- Informatik (1249)
- Center for Financial Studies (CFS) (1136)
- Sustainable Architecture for Finance in Europe (SAFE) (1059)
- Biochemie und Chemie (855)
- House of Finance (HoF) (700)
ABCB9 is a peptide transporter belonging to the ATP-binding cassette (ABC) transporter subfamily B. Due to its high sequence identity to the transporter associated with antigen processing (TAP) the protein was named TAP-like (TAPL). The primary aim of this PhD thesis was the functional characterization of the TAPL transport complex. Despite the lack of TAPL function in the classical MHC class I pathway an involvement of TAPL in antigen presentation was still suggested. Apart from the crucial role of TAP for peptide delivery into the ER, TAP-independent translocation pathways in professional antigen presenting cells (pAPC) have been proposed, but not identified so far. Remarkably, TAPL mRNA and protein expression is strongly induced during differentiation of monocytes to immature and mature dendritic cells. This result was confirmed in the promonocytic cell line THP-1, which was used as a model system for monocyte to macrophage differentiation. By using quantitative immunofluorescence microscopy and subcellular fractionation, TAPL was detected in the lysosomal compartment co-localizing with the lysosome associated membrane protein 2 (LAMP-2) thus excluding the ER-localization formerly reported. Furthermore, by in vitro assays, a TAPL-specific and ATPdependent translocation of peptides into isolated lysosomes was demonstrated. Hence, TAPL is a candidate mediating peptide transport in alternative antigen presentation pathways in pAPCs. The presence of an extra N-terminal transmembrane domain (TMD0) lacking sequence homology to any known protein distinguishes TAPL from most other ABC transporters of its subfamily. By dissecting the TAPL translocation complex into its four putative transmembrane helices containing TMD0 and the core complex, distinct functions to the core complex and TMD0 were assigned. The core-TAPL complex composed of six predicted transmembrane helices and the nucleotide-binding domain (NBD) was expressed transiently in HeLa or stably in Raji cells. Crude membranes containing core-TAPL showed the same peptide transport activity as wt-TAPL demonstrating that the six core helices and the NBD are sufficient for peptide transport. This result also shows that the core transport complex is correctly targeted to and assembled in the membrane. Strikingly, in contrast to the wt transporter, the core complex localizes only partially to lysosomes and is mistargeted to the plasma membrane as observed by immunofluorescence microscopy and confirmed biochemically by cell surface biotinylation. Thus, a crucial role for TMD0 in proper subcellular targeting can be postulated. The vast majority of biological processes are mediated by protein complexes, hence characterization of such protein-protein-interactions is essential for understanding protein function on the cellular level. To identify interaction partners of TAPL, the transporter was isolated by tandem affinity purification. By tandem mass spectrometry the membrane proteins LAMP-1 and LAMP-2 were deciphered as specific proteins interacting with wt-TAPL. Notably, core-TAPL lacks these interactions indicating a role for TMD0 in recruiting other proteins. These results were verified for endogenous TAPL by co-immunoprecipitation. Using cells deficient in LAMP-1 and/or in LAMP-2 an escort function for the LAMP proteins was excluded. Very importantly, the physiological function of the LAMP-1and LAMP-2 interaction with TAPL is an increase in stability, since in their absence half-life of TAPL is drastically reduced.
Understanding the dynamics of recurrent neural networks is crucial for explaining how the brain processes information. In the neocortex, a range of different plasticity mechanisms are shaping recurrent networks into effective information processing circuits that learn appropriate representations for time-varying sensory stimuli. However, it has been difficult to mimic these abilities in artificial neural models. In the present thesis, we introduce several recurrent network models of threshold units that combine spike timing dependent plasticity with homeostatic plasticity mechanisms like intrinsic plasticity or synaptic normalization. We investigate how these different forms of plasticity shape the dynamics and computational properties of recurrent networks. The networks receive input sequences composed of different symbols and learn the structure embedded in these sequences in an unsupervised manner. Information is encoded in the form of trajectories through a high-dimensional state space reminiscent of recent biological findings on cortical coding. We find that these self-organizing plastic networks are able to represent and "understand" the spatio-temporal patterns in their inputs while maintaining their dynamics in a healthy regime suitable for learning. The emergent properties are not easily predictable on the basis of the individual plasticity mechanisms at work. Our results underscore the importance of studying the interaction of different forms of plasticity on network behavior.
Top-down and bottom-up approaches are the general methods used to analyse proteomic samples today, however, the bottom-up approach has been dominant in the last decade. Establishing a bottom-up method involves not only the choice of adequate instruments and the optimisation of the experimental parameters, but also choosing the right experimental conditions and sample preparation steps. LC-ESI MS/MS has widely been used in this field due to its advanced automation. The primary objective of the present study was to establish a sensitive high-throughput nLC-MALDI MS/MS method for the identification and characterisation of proteins in biological samples. The method establishment included optimisation and validation of parameters such as the capillaries in the HPLC systems, gradient slopes, column temperature, spotting frequencies or the MS and MS/MS acquisition methods. The optimisation was performed using two HPLC-systems (Agilent 1100 series and Proxeon Easy nLC system), three spotters and the 4800 MALDI-TOF/TOF analyzer. Furthermore, samples preparation protocols were modified to fit to the established nLCMALDI- TOF/TOF-platform. The potentials of this method was demonstrated by the successful analysis of complex protein samples isolated from lipid particles, pre-adipocytes/adipocytes tissues, membrane proteins and proteins pulled-down from protein-proteins interaction studies. Despite the small amount of proteins in the lipid particles or oil bodies, and the challenges encountered in studying such proteins, 41(6 novel + 14 mammal specific + 21 visceral specific) proteins were added to the already existing proteins of the secretome of human subcutaneous (pre)adipocytes and 6 novel proteins localised in the yeast lipid particles. Protein-protein interaction studies present another area of application. Here the analytical challenges are mostly due to the loss of binding partner upon sample clean-up and to differentiate from non-specific background. Novel interaction partners for AF4•MLL and AF4 protein complex were identified. Furthermore, a novel sample protocol for the analysis of membrane proteins, based on the less specific protease, elastase, was established. Compared to trypsin, a higher sequence coverage and higher coverage of the transmembrane domains were achieved. The use of this enzyme in proteomics has been limited because of its non specific cleavage. However, from the results obtained in these studies, elastase was found to cleave preferentially at the C-terminal site of the amino acids AVLIST. The advantage of the established protocol over conventional protocols is that the same enzyme can be used for shaving of the soluble dormains of intact proteins in membranes and the digestion of the hydrophobic domain after solubilisation. Furthermore, the solvents used are compatible with the nLC-MALDI method setup. In addition, it was also shown that for less specific enzymes, a higher mass accuracy is required to reduce the rate of false positive identifications, since current search engines are not perfectly adapted for these types of enzymes. A brief statistical analysis of the MS/MS data obtained from the LC-MALDI TOF/TOF system showed that for less specific enzymes, under high-energy collision conditions, approximately 43 % of the fragment ions could not be matched to the known y- b type ions and their resultant internal fragments. This limitation greatly influenced the search results. However, this limitation can be overcome by modifying the N-terminal amino acids with basic moieties such as TMT. The use of elastase as a digestion enzyme in proteomic workflow further increased the complexity of the sample. Therefore, orthogonal multidimensional separation was necessary. Offgel-IEF was used as the separation technique for the first dimension. Here peptides are separated according to the pI. However, the acquired samples could not be loaded to the nLC due to the high viscosity of the concentrated samples when using the standard protocol. In order to achieve compatibility of the Offgel-IEF to the nLC-MALDI-TOF/TOF-platform, the separation protocol of the Offgel-IEF was modified by omitting the glycerol, which was the cause of the viscous solution. The novel glycerol free protocol is advantageous over the conventional method because the samples could directly be picked-up and loaded onto the pre-column without resulting in an increase in back pressure or a subsequent pre-column clogging. The glycerol free protocol was then assessed using purple membrane and membrane fraction of C. glutamicum. The results obtained were comparable to those applied in published reports. Therefore, the absence of glycerol did not affect the separation efficiency of the Offgel-IEF. In addition the applicability of elastase and the glycerol free Offgel-IEF for quantitation of membrane proteins was assessed. Most of the unique peptides identified were in the acidic region and 85 % were focused only into one fraction and approximately 95 % in only two fractions. These results are in accordance with previously published results (Lengqvist et al., 2007). When compared with theoretical digests of the proteins identified in this study, it can be concluded that basic moiety (TMT) on the peptide backbone, did not affect the separation efficiency of the Offgel-IEF. In an applied study, changes in the protein content of yeast strain grown in two different media were relatively quantified. For example, prominent proteins, such as the hexose tranporter proteins responsible for transporting glucose accross the membrane, were successfully quantified. Last but not least, the nLC-MALDI-TOF/TOF platform also served as a basis for the development of a high-throughput method for the identification of protein phosphorylation. The establishment of such a method using MALDI has been challenging due to the lack of sensitive matrices, such as CHCA for non-modified peptides, which exhibit a homogenous crystallisation and thus yield stable signal intensity over a long period of time in an automated setup. The first step of this method was the establishment of a matrix/matrix mixture with better crystal morphology and higher analyte signal intensity than the matrix of choice, i.e. DHB. From MS and MS/MS measurements of standard phosphopeptides, a combination of FCCA and CHAC in a 3:1 ratio and 3 mM NH4H2PO4 facilitated high analyte signal intensities and good fragmentation behaviour. Combining a custom-packed biphasic column for the enrichment of phosphopeptides, the applicability of the matrix mixture was assessed in anautomated phosphopeptide analysis using standard phosphopeptides spiked to a 20-fold excess BSA digest. These analyses showed that this method is reproducibile and both flow throughs can be analysed. Applying the method to the analysis of 2 standard phosphoproteins, alpha/beta-casein, and a leukemia related protein, ENL, 13 phosphopeptides from both alpha/beta-Casein and 13 phosphopeptides with 6 phosphorylation sites from the ENL were identified. As a general conclusion, it can be stated that the nLC-MALDI-TOF/TOF method established here in various modifications for different analytical purposes is a robust platform for proteomic analyses.
In this study, I investigate the crustal and upper mantle velocity structure beneath the Rwenzori Mountains in western Uganda. This mountain range is situated within the western branch of the East African Rift and reaches altitudes of more than 5000 m. I use four different approaches that belong to the travel-time tomography method. The first approach is based on the isotropic tomographic inversion of local data, which contain information about 2053 earthquakes recorded by a network of up to 35 stations covering an area of 140×90 km2. The LOTOS-09 algorithm described here is used to realize this approach. The second approach is based on the anisotropic tomographic inversion of the same local dataset. This method employs the tomographic code ANITA, developed with my participation, which provides 3D anisotropic P and isotropic S velocity distributions based on P and S travel-times from local seismicity. For the P anisotropic model, four parameters for each parameterization cell are determined. This represents an orthorhombic anisotropy with one vertically-oriented predefined direction. Three of the parameters describe slowness variations along three horizontal orientations with azimuths of 0°, 60° and 120°, and one is a perturbation along the vertical axis. The third approach is based on tomographic inversion of the teleseismic data, which contain information about the traveltimes of P-waves coming from 284 teleseismic events recorded by the seismic network stations. The TELELOTOS code, which is my own modification of the LOTOS-09 algorithm, is used in this approach. The TELELOTOS code is designed to iteratively invert the local and/or teleseismic datasets. Finally, I present the results of the new tomographic approach, which is based on the simultaneous inversion of the joint local and teleseismic data. The simultaneous use of these datasets for the tomographic inversion has several advantages. In this case, the velocity structure in the study area can be resolved as deep as in the teleseismic approach. At the same time, in the upper part of the study volume, the resolution of the obtained models is as good as in the local tomography. The TELELOTOS algorithm is used to perform the joint tomographic inversion. Special attention is paid in this work to synthetic testing. A number of different synthetic and real data tests are performed to estimate the resolution ability and robustness of the obtained models. In particular, synthetic tests have shown that the results of the anisotropic tomographic inversion of the local data have to be considered as unsatisfactory. For all approaches used in this study, I present synthetic models that reproduce the same pattern of anomalies as that obtained by inverting the real data. These models are used to interpret the results and estimate the real amplitudes of the obtained anomalies. The obtained models exhibit a relatively strong negative P anomaly (up to -10%) beneath the Rwenzori Mountains. Low velocities are found in the northeastern part of the array at shallower depths and are most likely related to sedimentary deposits, while higher velocities are found beneath the eastern rift shoulder and are thought to be related to old cratonic crust. The presence of low velocities in the northwestern part of the array may be caused by a magmatic intrusion beneath the Buranga hot springs. Relatively low velocities were observed within the lower crust and upper mantle in the western and southern parts of the study area (beneath the rift valley and the entire length of the Rwenzori range). The higher amplitude of the low-velocity anomaly in the south can be related to the thinner lithosphere in the southern part of the Albertine rift. In the center of the study area, a small negative anomaly is observed, with the intensity increasing with depth. This anomaly is presumably related to a fluids rising up from a plume branch in the deeper part of the mantle. According to the interpretation of the local earthquake distribution, the Rwenzori Mountains are located between two rift valleys with flanks marked by normal faults. The Rwenzori block is bounded by thrust faults that are probably due to compression.
Capoeta damascina (Teleostei: Cyprinidae) is one of the most common freshwater fish species, found throughout the Levant, Mesopotamia, Turkey and Iran. According to the state of knowledge prior to this study, C. damascina, which is distributed over a wide range of isolated water bodies, was not a well-defined species. It was questionable whether it represents a single species or a complex of closely related species with high intraspecific and comparatively low interspecific variability. The goal of this study was to investigate the taxonomy, systematic position of the C. damascina species complex and the phylogenetic relationships among its members, based on morphological features as well as molecular phylogeny. Samples obtained from throughout the geographic range of this species complex were subjected to comparative morphological analyses in order to define, properly diagnose and separate species within the C. damascina complex. To elucidate phylogenetic relationships among members of the C. damascina species complex, samples were subjected to genetic analyses, using two molecular markers targeting the mitochondrial cytochrome oxidase I (COI, n = 103) and the two adjacent divergence regions (D1-D2) of the nuclear 28S rRNA genes (LSU, n = 65). Based on morphological and molecular genetic data, six closely related species were recognized within the C. damascina complex: C. buhsei, C. caelestis, C. damascina, C. saadii, C. umbla and an undescribed species, Capoeta sp.1. Analyses of the morphometric and meristic data obtained in this study revealed phenotypic variability among the various populations within a species and among the different species. Such differences in morphological characters reflect genetic differences, environmentally induced phenotypic variation or both, as the meristic phenotype of fish is sometimes a consequence of environmental parameters acting on the genotype. Based on phylogenetic analyses, two main lineages were identified within the C. damascina species complex: a western lineage represented by C. caelestis, C. damascina and C. umbla and an eastern lineage represented by C. buhsei, C. saadii and Capoeta sp.1. The close phylogenetic relationships between C. damascina and C. umbla and the sharing of same haplotypes between one specimen of C. damascina from Euphrates and another of C. umbla from Tigris reflect one of three possibilites: recent speciation, mitochondrial introgression or a combination of both. The results obtained in this study indicate that speciation of the above-mentioned six taxa is quite recent and that their dispersal and present-day distribution can be related to Pleistocene events. The drying out of the Persian Gulf, probably during one of the first glacials of the Pleistocene, led the ancestor of the C. damascina species complex in Mesopotamia to reach the rivers of the Gulf and of Hormuz basins and differentiate there, giving rise to the eastern lineage (ancestor of C. buhsei, C. saadii and Capoeta sp.1). As connections presumably existed among the different river drainages and basins in Iran during the wet periods of the Pleistocene, the ancestor of C. buhsei, C. saadii and Capoeta sp.1 was subsequently able to colonize the various Iranian drainages and differentiate there, giving rise to C. buhsei, C. saadii and Capoeta sp.1. After the separation from the eastern lineage, the western lineage, represented by the ancestor of C. damascina, C. umbla and C. caelestis, most likely reached the Levant from the Tigris-Euphrates system during the Pleistocene glacials, when river connections existed in the regions of the upper courses of Ceyhan Nehri (southern Turkey) and some western affluents to the Euphrates. From Ceyhan Nehri, it dispersed into other rivers in southern Turkey during Pleistocene periods of low sea levels until it reached Göksu Nehri and evolved into C. caelestis. The sister population differentiated into C. damascina and C. umbla. Based on the results obtained in this study, it is likely that C. damascina colonized the Levant and southern Turkey during the Pleistocene glacials. This is well supported by the low genetic variability among the C. damascina populations. Direct connections existed among the river drainages in the Levant during the Pleistocene periods of low sea level, thus serving as a pathway for the dispersal of C. damascina. The results of this study provide a coherent picture of the taxonomic position, phylogenetic relationships and evolutionary history of the C. damascina species complex and explain present patterns of distribution considering paleogeographic events.
Background: Although literature provides support for cognitive behavioral therapy (CBT) as an efficacious intervention for social phobia, more research is needed to improve treatments for children. Methods: Forty four Caucasian children (ages 8-14) meeting diagnostic criteria of social phobia according to the Diagnostic and Statistical Manual of Mental Disorders (4th ed.; APA, 1994) were randomly allocated to either a newly developed CBT program focusing on cognition according to the model of Clark and Wells (n = 21) or a wait-list control group (n = 23). The primary outcome measure was clinical improvement. Secondary outcomes included improvements in anxiety coping, dysfunctional cognitions, interaction frequency and comorbid symptoms. Outcome measures included child report and clinican completed measures as well as a diagnostic interview. Results: Significant differences between treatment participants (4 dropouts) and controls (2 dropouts) were observed at post test on the German version of the Social Phobia and Anxiety Inventory for Children. Furthermore, in the treatment group, significantly more children were free of diagnosis than in wait-list group at post-test. Additional child completed and clinician completed measures support the results. Discussion: The study is a first step towards investigating whether CBT focusing on cognition is efficacious in treating children with social phobia. Future research will need to compare this treatment to an active treatment group. There remain the questions of whether the effect of the treatment is specific to the disorder and whether the underlying theoretical model is adequate. Conclusion: Preliminary support is provided for the efficacy of the cognitive behavioral treatment focusing on cognition in socially phobic children. Active comparators should be established with other evidence-based CBT programs for anxiety disorders, which differ significantly in their dosage and type of cognitive interventions from those of the manual under evaluation (e.g. Coping Cat).
Using faculty-librarian partnerships to ensure that students become information fluent in the 21st century In the 21st century educators in partnership with librarians must prepare students effectively for productive use of information especially in higher education. Students will need to graduate from universities with appropriate information and technology skills to enable them to become productive citizens in the workplace and in society. Technology is having a major impact on society; in economics e-business is moving to the forefront; in communication e-mail, the Internet and cellular telephones have reformed how people communicate; in the work environment computers and web utilizations are emphasized and in education virtual learning and teaching are becoming more important. These few examples indicate how the 21st century information environment requires future members of the workforce to be information fluent so they will have the ability to locate information efficiently, evaluate information for specific needs, organize information to address issues, apply information skillfully to solve problems, use information to communicate effectively, and use information responsibly to ensure a productive work environment. Individuals can achieve information fluency by acquiring cultural, visual, computer, technology, research and information management skills to enable them to think critically.
Teaching information literacy: substance and process This presentation explores the concept of information literacy within the broader context of higher education. It argues that, certain assertions in the library literature notwithstanding, the concepts associated with information literacy are not new, but rather very closely resemble the qualities traditionally considered to characterize a well-educated person. The presentation also considers the extent to which the higher education system does indeed foster the attributes commonly associated with information literacy. The term information literacy has achieved the immediacy it currently enjoys within the library community with the advent of the so-called "information age" The information age is commonly touted in the literature, both popular and professional, as constituting nothing short of a revolution. Academic librarians and other educators have of course felt called upon to make their teaching reflect both the growing proliferation of information formats and the major transformations affecting the process of information seeking. Faced with so much novelty and uncertainty, it is no surprise that many have felt that these changes call for a revolution in teaching. It is within this context that the concept of information literacy has flourished. It is argued in this presentation, however, that by treating information literacy as an essentially new specialty that owes much of its importance to the plethora of electronic information, we risk obscuring some of the most fundamental and enduring educational values we should be imparting to our students. Much of the literature on information literacy assumes - rather than argues - that recent changes in the way we approach education are indications of progress. Indeed, much of the self-narrative that institutions produce (in bulletins, mission statements, web sites, etc.) endorses an approach to education that will result in lifelong learners who are critical consumers of information. After critically examining the degree to which such statements of educational approach reflect reality, this presentation concludes by considering the effects of certain changes in the culture of higher education. It considers particularly the transformation - at least in North America - of the traditional model of higher education as a public good to a market-driven business model. It poses the question of whether a change of this significance might in fact detract from, rather than promote, the development of information literate students.
Development of a computational method for reaction-driven de novo design of druglike compounds
(2010)
A new method for computer-based de novo design of drug candidate structures is proposed. DOGS (Design of Genuine Structures) features a ligand-based strategy to suggest new molecular structures. The quality of designed compounds is assessed by a graph kernel method measuring the distance of designed molecules to a known reference ligand. Two graph representations of molecules (molecular graph and reduced graph) are implemented to feature different levels of abstraction from the molecular structure. A fully deterministic construction procedure explicitly designed to facilitate synthesizability of proposed structures is realized: DOGS uses readily available synthesis building blocks and established reaction schemes to assemble new molecules. This approach enables the software to propose not only the final compounds, but also to give suggestions for synthesis routes to generate them at the bench. The set of synthesis schemes comprises about 83 chemical reactions. Special focus was put on ring closure reactions forming drug-like substructures. The library of building blocks consists of about 25,000 readily available synthesis building blocks. DOGS builds up new structures in a stepwise process. Each virtual synthesis step adds a fragment to the growing molecule until a stop criterion (upper threshold for molecular mass or number of synthesis steps) is fulfilled. In a theoretical evaluation, a set of ~1,800 molecules proposed by DOGS is analyzed for critical properties of de novo designed compounds. The software is able to suggest drug-like molecules (79% violate less than two of Lipinski’s ‘rule of five’). In addition, a trained classifier for drug-likeness assigns a score >0.8 to 51% of the designed molecules (with 1.0 being the top score). In addition, most of the DOGS molecules are deemed to be synthesizable by a retro-synthesis descriptor (77% of molecules score in the top 10% of the decriptor’s value range). Calculated logP(o/w) values of constructed molecules resemble a unimodal distribution centred close to the mean of logP(o/w) values calculated for the reference compounds. A structural analysis of selected designs reveals that DOGS is capable of constructing molecules reflecting the overall topological arrangement of pharmacophoric features found in the reference ligands. At the same time, the DOGS designs represent innovative compounds being structurally distinct from the references. Synthesis routes for these examples are short and seem feasible in most cases. Some reaction steps might need modification by using protecting groups to avoid unwanted side reactions. Plausible bioisosters for known privileged fragments addressing the S1 pocket of trypsin were proposed by DOGS in a case study. Three of them can be found in known trypsin inhibitors as S1-adressing side chains. The software was also tested in two prospective case studies to design bioactive compounds. DOGS was applied to design ligands for human gamma-secretase and human histamine receptor subtype 4 (hH4R). Two selected designs for gamma-secretase were readily synthesizable as suggested by the software in one-step reactions. Both compounds represent inverse modulators of the target molecule. In a second case study, a ligand candidate selected for hH4R was synthesized exactly following the three-step synthesis plan suggested by DOGS. This compound showed low activity on the target structure. The concept of DOGS is able to deliver synthesizable and bioactive compounds. Suggested synthesis plans of selected compounds were readily pursuable. DOGS can therefore serve as a valuable idea generator for the design of new pharmacological active compounds.
Biodegradation and elimination of industrial wastewater in the context of whole effluent assessment
(2010)
The focus of this thesis is on the assessment of the degradability of indirectly discharged wastewater in municipal treatment plants and on assessing indirectly discharged effluents by coupling the Zahn-Wellens test with effect-based bioassays. With this approach persistent toxicity of an indirectly discharged effluent can be detected and attributed to the respective emission source. In the first study 8 wastewater samples from different industrial sectors were analysed according to the “Whole-Effluent Assessment“ (WEA) approach developed by OSPAR. In another study this concept has been applied with 20 wastewater samples each from paper manufacturing and metal surface treating industry. In the first study generally low to moderate ecotoxic effects of wastewater samples have been determined. One textile wastewater sample was mutagenic in the Ames test and genotoxic in the umu test. The source of these effects could not be identified. After treatment in the Zahn-Wellens test the mutagenicity in the Ames test was eliminated completely while in the umu test genotoxicity could still be observed. Another wastewater sample from chemical industry was mutagenic in the Ames test. The mutagenicity with this wastewater sample was investigated by additional chemical analysis and backtracking. A nitro-aromatic compound (2-methoxy-4-nitroaniline) used for batchwise azo dye synthesis and its transformation products are the probable cause for the mutagenic effects analysed. Testing the mother liquor from dye production confirmed that this partial wastewater stream was mutagenic in the Ames test. The wasteweater samples from paper manufacturing industry of the second study were not toxic or genotoxic in the acute Daphnia test, fish egg test and umu test. In the luminescent bacteria test, moderate toxicity was observed. Wastewater of four paper mills demonstrated elevated or high algae toxicity, which was in line with the results of the Lemna test, which mostly was less sensitive than the algae test. The colouration of the wastewater samples in the visible band did not correlate with algae toxicity and thus is not considered as its primary origin. The algae toxicity in wastewater of the respective paper factory could also not be explained with the thermomechanically produced groundwood pulp (TMP) partial stream. Presumably other raw materials such as biocides might be the source of algae toxicity. In the algae test, often flat dose–response relationships and growth promotion at higher dilution factors have been observed, indicating that several effects are overlapping. The wastewater samples from the printed circuit board and electroplating industries (all indirectly discharged) were biologically pre-treated for 7 days in the Zahn–Wellens test before ecotoxicity testing. Thus, persistent toxicity could be discriminated from non-persistent toxicity caused, e.g. by ammonium or readily biodegradable compounds. With respect to the metal concentrations, all samples were not heavily polluted. The maximum conductivity of the samples was 43,700 micro S cm -1 and indicates that salts might contribute to the overall toxicity. Half of the wastewater samples proved to be biologically well treatable in the Zahn–Wellens test with COD elimination above 80%, whilst the others were insufficiently biodegraded (COD elimination 28–74%). After the pre-treatment in the Zahn–Wellens test, wastewater samples from four companies were extremely ecotoxic especially to algae. Three wastewater samples were genotoxic in the umu test. Applying the rules for salt correction to the test results following the German Wastewater Ordinance, only a small part of toxicity could be attributed to salts. In one factory, the origin of ecotoxicity has been attributed to the organosulphide dimethyldithiocarbamate (DMDTC) used as a water treatment chemical for metal precipitation. The assumption, based on rough calculation of input of the organosulphide into the wastewater, was confirmed in practice by testing its ecotoxicity at the corresponding dilution ratio after pre-treatment in the Zahn–Wellens test. The results show that bioassays are a suitable tool for assessing the ecotoxicological relevance of these complex organic mixtures. The combination of the Zahn–Wellens test followed by the performance of ecotoxicity tests turned out to be a cost-efficient suitable instrument for the evaluation of indirect dischargers and considers the requirements of the IPPC Directive.