This growing collection consists of scholarly works authored by ASU-affiliated faculty, staff, and community members, and it contains many open access articles. ASU-affiliated authors are encouraged to Share Your Work in KEEP.

Displaying 1 - 10 of 25
Filtering by

Clear all filters

Description

Linnorm is a novel normalization and transformation method for the analysis of single cell RNA sequencing (scRNA-seq) data. Linnorm is developed to remove technical noises and simultaneously preserve biological variations in scRNA-seq data, such that existing statistical methods can be improved. Using real scRNA-seq data, we compared Linnorm with existing

Linnorm is a novel normalization and transformation method for the analysis of single cell RNA sequencing (scRNA-seq) data. Linnorm is developed to remove technical noises and simultaneously preserve biological variations in scRNA-seq data, such that existing statistical methods can be improved. Using real scRNA-seq data, we compared Linnorm with existing normalization methods, including NODES, SAMstrt, SCnorm, scran, DESeq and TMM. Linnorm shows advantages in speed, technical noise removal and preservation of cell heterogeneity, which can improve existing methods in the discovery of novel subtypes, pseudo-temporal ordering of cells, clustering analysis, etc. Linnorm also performs better than existing DEG analysis methods, including BASiCS, NODES, SAMstrt, Seurat and DESeq2, in false positive rate control and accuracy.

ContributorsYip, Shun H. (Author) / Wang, Panwen (Author) / Kocher, Jean-Pierre A. (Author) / Sham, Pak Chung (Author) / Wang, Junwen (Author) / College of Health Solutions (Contributor)
Created2017-09-18
128573-Thumbnail Image.png
Description

Two distinct monocyte (Mo)/macrophage (Mp) subsets (Ly6Clow and Ly6Chi) orchestrate cardiac recovery process following myocardial infarction (MI). Prostaglandin (PG) E2 is involved in the Mo/Mp-mediated inflammatory response, however, the role of its receptors in Mos/Mps in cardiac healing remains to be determined. Here we show that pharmacological inhibition or gene

Two distinct monocyte (Mo)/macrophage (Mp) subsets (Ly6Clow and Ly6Chi) orchestrate cardiac recovery process following myocardial infarction (MI). Prostaglandin (PG) E2 is involved in the Mo/Mp-mediated inflammatory response, however, the role of its receptors in Mos/Mps in cardiac healing remains to be determined. Here we show that pharmacological inhibition or gene ablation of the Ep3 receptor in mice suppresses accumulation of Ly6Clow Mos/Mps in infarcted hearts. Ep3 deletion in Mos/Mps markedly attenuates healing after MI by reducing neovascularization in peri-infarct zones. Ep3 deficiency diminishes CX3C chemokine receptor 1 (CX3CR1) expression and vascular endothelial growth factor (VEGF) secretion in Mos/Mps by suppressing TGFβ1 signaling and subsequently inhibits Ly6Clow Mos/Mps migration and angiogenesis. Targeted overexpression of Ep3 receptors in Mos/Mps improves wound healing by enhancing angiogenesis. Thus, the PGE2/Ep3 axis promotes cardiac healing after MI by activating reparative Ly6Clow Mos/Mps, indicating that Ep3 receptor activation may be a promising therapeutic target for acute MI.

ContributorsTang, Juan (Author) / Shen, Yujun (Author) / Chen, Guilin (Author) / Wan, Qiangyou (Author) / Wang, Kai (Author) / Zhang, Jian (Author) / Qin, Jing (Author) / Liu, Guizhu (Author) / Zuo, Shengkai (Author) / Tao, Bo (Author) / Yu, Yu (Author) / Wang, Junwen (Author) / Lazarus, Michael (Author) / Yu, Ying (Author) / College of Health Solutions (Contributor)
Created2017-03-03
127886-Thumbnail Image.png
Description

Modeling of transcriptional regulatory networks (TRNs) has been increasingly used to dissect the nature of gene regulation. Inference of regulatory relationships among transcription factors (TFs) and genes, especially among multiple TFs, is still challenging. In this study, we introduced an integrative method, LogicTRN, to decode TF–TF interactions that form TF

Modeling of transcriptional regulatory networks (TRNs) has been increasingly used to dissect the nature of gene regulation. Inference of regulatory relationships among transcription factors (TFs) and genes, especially among multiple TFs, is still challenging. In this study, we introduced an integrative method, LogicTRN, to decode TF–TF interactions that form TF logics in regulating target genes. By combining cis-regulatory logics and transcriptional kinetics into one single model framework, LogicTRN can naturally integrate dynamic gene expression data and TF-DNA-binding signals in order to identify the TF logics and to reconstruct the underlying TRNs. We evaluated the newly developed methodology using simulation, comparison and application studies, and the results not only show their consistence with existing knowledge, but also demonstrate its ability to accurately reconstruct TRNs in biological complex systems.

ContributorsYan, Bin (Author) / Guan, Daogang (Author) / Wang, Chao (Author) / Wang, Junwen (Author) / He, Bing (Author) / Qin, Jing (Author) / Boheler, Kenneth R. (Author) / Lu, Aiping (Author) / Zhang, Ge (Author) / Zhu, Hailong (Author) / College of Health Solutions (Contributor)
Created2017-10-19
127882-Thumbnail Image.png
Description

The estimation of energy demand (by power plants) has traditionally relied on historical energy use data for the region(s) that a plant produces for. Regression analysis, artificial neural network and Bayesian theory are the most common approaches for analysing these data. Such data and techniques do not generate reliable results.

The estimation of energy demand (by power plants) has traditionally relied on historical energy use data for the region(s) that a plant produces for. Regression analysis, artificial neural network and Bayesian theory are the most common approaches for analysing these data. Such data and techniques do not generate reliable results. Consequently, excess energy has to be generated to prevent blackout; causes for energy surge are not easily determined; and potential energy use reduction from energy efficiency solutions is usually not translated into actual energy use reduction. The paper highlights the weaknesses of traditional techniques, and lays out a framework to improve the prediction of energy demand by combining energy use models of equipment, physical systems and buildings, with the proposed data mining algorithms for reverse engineering. The research team first analyses data samples from large complex energy data, and then, presents a set of computationally efficient data mining algorithms for reverse engineering. In order to develop a structural system model for reverse engineering, two focus groups are developed that has direct relation with cause and effect variables. The research findings of this paper includes testing out different sets of reverse engineering algorithms, understand their output patterns and modify algorithms to elevate accuracy of the outputs.

ContributorsNaganathan, Hariharan (Author) / Chong, Oswald (Author) / Ye, Long (Author) / Ira A. Fulton School of Engineering (Contributor)
Created2015-12-09
127878-Thumbnail Image.png
Description

Small and medium office buildings consume a significant parcel of the U.S. building stock energy consumption. Still, owners lack resources and experience to conduct detailed energy audits and retrofit analysis. We present an eight-steps framework for an energy retrofit assessment in small and medium office buildings. Through a bottom-up approach

Small and medium office buildings consume a significant parcel of the U.S. building stock energy consumption. Still, owners lack resources and experience to conduct detailed energy audits and retrofit analysis. We present an eight-steps framework for an energy retrofit assessment in small and medium office buildings. Through a bottom-up approach and a web-based retrofit toolkit tested on a case study in Arizona, this methodology was able to save about 50% of the total energy consumed by the case study building, depending on the adopted measures and invested capital. While the case study presented is a deep energy retrofit, the proposed framework is effective in guiding the decision-making process that precedes any energy retrofit, deep or light.

ContributorsRios, Fernanda (Author) / Parrish, Kristen (Author) / Chong, Oswald (Author) / Ira A. Fulton School of Engineering (Contributor)
Created2016-05-20
127865-Thumbnail Image.png
Description

Commercial buildings’ consumption is driven by multiple factors that include occupancy, system and equipment efficiency, thermal heat transfer, equipment plug loads, maintenance and operational procedures, and outdoor and indoor temperatures. A modern building energy system can be viewed as a complex dynamical system that is interconnected and influenced by external

Commercial buildings’ consumption is driven by multiple factors that include occupancy, system and equipment efficiency, thermal heat transfer, equipment plug loads, maintenance and operational procedures, and outdoor and indoor temperatures. A modern building energy system can be viewed as a complex dynamical system that is interconnected and influenced by external and internal factors. Modern large scale sensor measures some physical signals to monitor real-time system behaviors. Such data has the potentials to detect anomalies, identify consumption patterns, and analyze peak loads. The paper proposes a novel method to detect hidden anomalies in commercial building energy consumption system. The framework is based on Hilbert-Huang transform and instantaneous frequency analysis. The objectives are to develop an automated data pre-processing system that can detect anomalies and provide solutions with real-time consumption database using Ensemble Empirical Mode Decomposition (EEMD) method. The finding of this paper will also include the comparisons of Empirical mode decomposition and Ensemble empirical mode decomposition of three important type of institutional buildings.

ContributorsNaganathan, Hariharan (Author) / Chong, Oswald (Author) / Huang, Zigang (Author) / Cheng, Ying (Author) / Ira A. Fulton School of Engineering (Contributor)
Created2016-05-20
127858-Thumbnail Image.png
Description

Background: While there is ample evidence for health risks associated with heat and other extreme weather events today, little is known about the impact of weather patterns on population health in preindustrial societies.

Objective: To investigate the impact of weather patterns on population health in Sweden before and during industrialization.

Methods: We

Background: While there is ample evidence for health risks associated with heat and other extreme weather events today, little is known about the impact of weather patterns on population health in preindustrial societies.

Objective: To investigate the impact of weather patterns on population health in Sweden before and during industrialization.

Methods: We obtained records of monthly mortality and of monthly mean temperatures and precipitation for Skellefteå parish, northern Sweden, for the period 1800-1950. The associations between monthly total mortality, as well as monthly mortality due to infectious and cardiovascular diseases, and monthly mean temperature and cumulative precipitation were modelled using a time series approach for three separate periods, 1800−1859, 1860-1909, and 1910-1950.

Results: We found higher temperatures and higher amounts of precipitation to be associated with lower mortality both in the medium term (same month and two-months lag) and in the long run (lag of six months up to a year). Similar patterns were found for mortality due to infectious and cardiovascular diseases. Furthermore, the effect of temperature and precipitation decreased over time.

Conclusions: Higher temperature and precipitation amounts were associated with reduced death counts with a lag of up to 12 months. The decreased effect over time may be due to improvements in nutritional status, decreased infant deaths, and other changes in society that occurred in the course of the demographic and epidemiological transition.

Contribution: The study contributes to a better understanding of the complex relationship between weather and mortality and, in particular, historical weather-related mortality.

ContributorsDaniel, Oudin Astrom (Author) / Edvinsson, Soren (Author) / Hondula, David M. (Author) / Rocklov, Joacim (Author) / Schumann, Barbara (Author) / College of Liberal Arts and Sciences (Contributor)
Created2016-10-05
127833-Thumbnail Image.png
Description

There are many data mining and machine learning techniques to manage large sets of complex energy supply and demand data for building, organization and city. As the amount of data continues to grow, new data analysis methods are needed to address the increasing complexity. Using data from the energy loss

There are many data mining and machine learning techniques to manage large sets of complex energy supply and demand data for building, organization and city. As the amount of data continues to grow, new data analysis methods are needed to address the increasing complexity. Using data from the energy loss between the supply (energy production sources) and demand (buildings and cities consumption), this paper proposes a Semi-Supervised Energy Model (SSEM) to analyse different loss factors for a building cluster. This is done by deep machine learning by training machines to semi-supervise the learning, understanding and manage the process of energy losses. Semi-Supervised Energy Model (SSEM) aims at understanding the demand-supply characteristics of a building cluster and utilizes the confident unlabelled data (loss factors) using deep machine learning techniques. The research findings involves sample data from one of the university campuses and presents the output, which provides an estimate of losses that can be reduced. The paper also provides a list of loss factors that contributes to the total losses and suggests a threshold value for each loss factor, which is determined through real time experiments. The conclusion of this paper provides a proposed energy model that can provide accurate numbers on energy demand, which in turn helps the suppliers to adopt such a model to optimize their supply strategies.

ContributorsNaganathan, Hariharan (Author) / Chong, Oswald (Author) / Chen, Xue-wen (Author) / Ira A. Fulton Schools of Engineering (Contributor)
Created2015-09-14
128638-Thumbnail Image.png
Description

It remains challenging to predict regulatory variants in particular tissues or cell types due to highly context-specific gene regulation. By connecting large-scale epigenomic profiles to expression quantitative trait loci (eQTLs) in a wide range of human tissues/cell types, we identify critical chromatin features that predict variant regulatory potential. We present

It remains challenging to predict regulatory variants in particular tissues or cell types due to highly context-specific gene regulation. By connecting large-scale epigenomic profiles to expression quantitative trait loci (eQTLs) in a wide range of human tissues/cell types, we identify critical chromatin features that predict variant regulatory potential. We present cepip, a joint likelihood framework, for estimating a variant’s regulatory probability in a context-dependent manner. Our method exhibits significant GWAS signal enrichment and is superior to existing cell type-specific methods. Furthermore, using phenotypically relevant epigenomes to weight the GWAS single-nucleotide polymorphisms, we improve the statistical power of the gene-based association test.

ContributorsLi, Mulin Jun (Author) / Li, Miaoxin (Author) / Liu, Zipeng (Author) / Yan, Bin (Author) / Pan, Zhicheng (Author) / Huang, Dandan (Author) / Liang, Qian (Author) / Ying, Dingge (Author) / Xu, Feng (Author) / Yao, Hongcheng (Author) / Wang, Panwen (Author) / Kocher, Jean-Pierre A. (Author) / Xia, Zhengyuan (Author) / Sham, Pak Chung (Author) / Liu, Jun S. (Author) / Wang, Junwen (Author) / College of Health Solutions (Contributor)
Created2017-03-16
128972-Thumbnail Image.png
Description

Background: Most excess deaths that occur during extreme hot weather events do not have natural heat recorded as an underlying or contributing cause. This study aims to identify the specific individuals who died because of hot weather using only secondary data. A novel approach was developed in which the expected number

Background: Most excess deaths that occur during extreme hot weather events do not have natural heat recorded as an underlying or contributing cause. This study aims to identify the specific individuals who died because of hot weather using only secondary data. A novel approach was developed in which the expected number of deaths was repeatedly sampled from all deaths that occurred during a hot weather event, and compared with deaths during a control period. The deaths were compared with respect to five factors known to be associated with hot weather mortality. Individuals were ranked by their presence in significant models over 100 trials of 10,000 repetitions. Those with the highest rankings were identified as probable excess deaths. Sensitivity analyses were performed on a range of model combinations. These methods were applied to a 2009 hot weather event in greater Vancouver, Canada.

Results: The excess deaths identified were sensitive to differences in model combinations, particularly between univariate and multivariate approaches. One multivariate and one univariate combination were chosen as the best models for further analyses. The individuals identified by multiple combinations suggest that marginalized populations in greater Vancouver are at higher risk of death during hot weather.

Conclusions: This study proposes novel methods for classifying specific deaths as expected or excess during a hot weather event. Further work is needed to evaluate performance of the methods in simulation studies and against clinically identified cases. If confirmed, these methods could be applied to a wide range of populations and events of interest.

Created2016-11-15