Understanding the role of short-interfering RNA (siRNA) in diverse biological processes is of current interest and often approached through small RNA sequencing. However, analysis of these datasets is difficult due to the complexity of biological RNA processing pathways, which differ between species. Several properties like strand specificity, length distribution, and distribution of soft-clipped bases are few parameters known to guide researchers in understanding the role of siRNAs. We present RAPID, a generic eukaryotic siRNA analysis pipeline, which captures information inherent in the datasets and automatically produces numerous visualizations as user-friendly HTML reports, covering multiple categories required for siRNA analysis. RAPID also facilitates an automated comparison of multiple datasets, with one of the normalization techniques dedicated for siRNA knockdown analysis, and integrates differential expression analysis using DESeq2.
Understanding the complexity of transcriptional regulation is a major goal of computational biology. Because experimental linkage of regulatory sites to genes is challenging, computational methods considering epigenomics data have been proposed to create tissue-specific regulatory maps. However, we showed that these approaches are not well suited to account for the variations of the regulatory landscape between cell-types. To overcome these drawbacks, we developed a new method called STITCHIT, that identifies and links putative regulatory sites to genes. Within STITCHIT, we consider the chromatin accessibility signal of all samples jointly to identify regions exhibiting a signal variation related to the expression of a distinct gene. STITCHIT outperforms previous approaches in various validation experiments and was used with a genome-wide CRISPR-Cas9 screen to prioritize novel doxorubicin-resistance genes and their associated non-coding regulatory regions. We believe that our work paves the way for a more refined understanding of transcriptional regulation at the gene-level.
Endothelial cells play a critical role in the adaptation of tissues to injury. Tissue ischemia induced by infarction leads to profound changes in endothelial cell functions and can induce transition to a mesenchymal state. Here we explore the kinetics and individual cellular responses of endothelial cells after myocardial infarction by using single cell RNA sequencing. This study demonstrates a time dependent switch in endothelial cell proliferation and inflammation associated with transient changes in metabolic gene signatures. Trajectory analysis reveals that the majority of endothelial cells 3 to 7 days after myocardial infarction acquire a transient state, characterized by mesenchymal gene expression, which returns to baseline 14 days after injury. Lineage tracing, using the Cdh5-CreERT2;mT/mG mice followed by single cell RNA sequencing, confirms the transient mesenchymal transition and reveals additional hypoxic and inflammatory signatures of endothelial cells during early and late states after injury. These data suggest that endothelial cells undergo a transient mes-enchymal activation concomitant with a metabolic adaptation within the first days after myocardial infarction but do not acquire a long-term mesenchymal fate. This mesenchymal activation may facilitate endothelial cell migration and clonal expansion to regenerate the vascular network.
Electrocardiograms (ECG) record the heart activity and are the most common and reliable method to detect cardiac arrhythmias, such as atrial fibrillation (AFib). Lately, many commercially available devices such as smartwatches are offering ECG monitoring. Therefore, there is increasing demand for designing deep learning models with the perspective to be physically implemented on these small portable devices with limited energy supply. In this paper, a workflow for the design of small, energy-efficient recurrent convolutional neural network (RCNN) architecture for AFib detection is proposed. However, the approach can be well generalized to every type of long time series. In contrast to previous studies, that demand thousands of additional network neurons and millions of extra model parameters, the logical steps for the generation of a CNN with only 114 trainable parameters are described. The model consists of a small segmented CNN in combination with an optimal energy classifier. The architectural decisions are made by using the energy consumption as a metric in an equally important way as the accuracy. The optimization steps are focused on the software which can be embedded afterwards on a physical chip. Finally, a comparison with some previous relevant studies suggests that the widely used huge CNNs for similar tasks are mostly redundant and unessentially computationally expensive.