Matching Items (18)
Filtering by

Clear all filters

135355-Thumbnail Image.png
Description
Glioblastoma multiforme (GBM) is a malignant, aggressive and infiltrative cancer of the central nervous system with a median survival of 14.6 months with standard care. Diagnosis of GBM is made using medical imaging such as magnetic resonance imaging (MRI) or computed tomography (CT). Treatment is informed by medical images and

Glioblastoma multiforme (GBM) is a malignant, aggressive and infiltrative cancer of the central nervous system with a median survival of 14.6 months with standard care. Diagnosis of GBM is made using medical imaging such as magnetic resonance imaging (MRI) or computed tomography (CT). Treatment is informed by medical images and includes chemotherapy, radiation therapy, and surgical removal if the tumor is surgically accessible. Treatment seldom results in a significant increase in longevity, partly due to the lack of precise information regarding tumor size and location. This lack of information arises from the physical limitations of MR and CT imaging coupled with the diffusive nature of glioblastoma tumors. GBM tumor cells can migrate far beyond the visible boundaries of the tumor and will result in a recurring tumor if not killed or removed. Since medical images are the only readily available information about the tumor, we aim to improve mathematical models of tumor growth to better estimate the missing information. Particularly, we investigate the effect of random variation in tumor cell behavior (anisotropy) using stochastic parameterizations of an established proliferation-diffusion model of tumor growth. To evaluate the performance of our mathematical model, we use MR images from an animal model consisting of Murine GL261 tumors implanted in immunocompetent mice, which provides consistency in tumor initiation and location, immune response, genetic variation, and treatment. Compared to non-stochastic simulations, stochastic simulations showed improved volume accuracy when proliferation variability was high, but diffusion variability was found to only marginally affect tumor volume estimates. Neither proliferation nor diffusion variability significantly affected the spatial distribution accuracy of the simulations. While certain cases of stochastic parameterizations improved volume accuracy, they failed to significantly improve simulation accuracy overall. Both the non-stochastic and stochastic simulations failed to achieve over 75% spatial distribution accuracy, suggesting that the underlying structure of the model fails to capture one or more biological processes that affect tumor growth. Two biological features that are candidates for further investigation are angiogenesis and anisotropy resulting from differences between white and gray matter. Time-dependent proliferation and diffusion terms could be introduced to model angiogenesis, and diffusion weighed imaging (DTI) could be used to differentiate between white and gray matter, which might allow for improved estimates brain anisotropy.
ContributorsAnderies, Barrett James (Author) / Kostelich, Eric (Thesis director) / Kuang, Yang (Committee member) / Stepien, Tracy (Committee member) / Harrington Bioengineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2016-05
136330-Thumbnail Image.png
Description
We model communication among social insects as an interacting particle system in which individuals perform one of two tasks and neighboring sites anti-mimic one another. Parameters of our model are a probability of defection 2 (0; 1) and relative cost ci > 0 to the individual performing task i. We

We model communication among social insects as an interacting particle system in which individuals perform one of two tasks and neighboring sites anti-mimic one another. Parameters of our model are a probability of defection 2 (0; 1) and relative cost ci > 0 to the individual performing task i. We examine this process on complete graphs, bipartite graphs, and the integers, answering questions about the relationship between communication, defection rates and the division of labor. Assuming the division of labor is ideal when exactly half of the colony is performing each task, we nd that on some bipartite graphs and the integers it can eventually be made arbitrarily close to optimal if defection rates are sufficiently small. On complete graphs the fraction of individuals performing each task is also closest to one half when there is no defection, but is bounded by a constant dependent on the relative costs of each task.
ContributorsArcuri, Alesandro Antonio (Author) / Lanchier, Nicolas (Thesis director) / Kang, Yun (Committee member) / Fewell, Jennifer (Committee member) / Barrett, The Honors College (Contributor) / School of International Letters and Cultures (Contributor) / Economics Program in CLAS (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2015-05
133413-Thumbnail Image.png
Description
Catastrophe events occur rather infrequently, but upon their occurrence, can lead to colossal losses for insurance companies. Due to their size and volatility, catastrophe losses are often treated separately from other insurance losses. In fact, many property and casualty insurance companies feature a department or team which focuses solely on

Catastrophe events occur rather infrequently, but upon their occurrence, can lead to colossal losses for insurance companies. Due to their size and volatility, catastrophe losses are often treated separately from other insurance losses. In fact, many property and casualty insurance companies feature a department or team which focuses solely on modeling catastrophes. Setting reserves for catastrophe losses is difficult due to their unpredictable and often long-tailed nature. Determining loss development factors (LDFs) to estimate the ultimate loss amounts for catastrophe events is one method for setting reserves. In an attempt to aid Company XYZ set more accurate reserves, the research conducted focuses on estimating LDFs for catastrophes which have already occurred and have been settled. Furthermore, the research describes the process used to build a linear model in R to estimate LDFs for Company XYZ's closed catastrophe claims from 2001 \u2014 2016. This linear model was used to predict a catastrophe's LDFs based on the age in weeks of the catastrophe during the first year. Back testing was also performed, as was the comparison between the estimated ultimate losses and actual losses. Future research consideration was proposed.
ContributorsSwoverland, Robert Bo (Author) / Milovanovic, Jelena (Thesis director) / Zicarelli, John (Committee member) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05
137504-Thumbnail Image.png
Description
The reconstruction of piecewise smooth functions from non-uniform Fourier data arises in sensing applications such as magnetic resonance imaging (MRI). This thesis presents a new polynomial based resampling method (PRM) for 1-dimensional problems which uses edge information to recover the Fourier transform at its integer coefficients, thereby enabling the use

The reconstruction of piecewise smooth functions from non-uniform Fourier data arises in sensing applications such as magnetic resonance imaging (MRI). This thesis presents a new polynomial based resampling method (PRM) for 1-dimensional problems which uses edge information to recover the Fourier transform at its integer coefficients, thereby enabling the use of the inverse fast Fourier transform algorithm. By minimizing the error of the PRM approximation at the sampled Fourier modes, the PRM can also be used to improve on initial edge location estimates. Numerical examples show that using the PRM to improve on initial edge location estimates and then taking of the PRM approximation of the integer frequency Fourier coefficients is a viable way to reconstruct the underlying function in one dimension. In particular, the PRM is shown to converge more quickly and to be more robust than current resampling techniques used in MRI, and is particularly amenable to highly irregular sampling patterns.
ContributorsGutierrez, Alexander Jay (Author) / Platte, Rodrigo (Thesis director) / Gelb, Anne (Committee member) / Viswanathan, Adityavikram (Committee member) / Barrett, The Honors College (Contributor) / School of International Letters and Cultures (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2013-05
137559-Thumbnail Image.png
Description
Serge Galams voting systems and public debate models are used to model voting behaviors of two competing opinions in democratic societies. Galam assumes that individuals in the population are independently in favor of one opinion with a fixed probability p, making the initial number of that type of opinion a

Serge Galams voting systems and public debate models are used to model voting behaviors of two competing opinions in democratic societies. Galam assumes that individuals in the population are independently in favor of one opinion with a fixed probability p, making the initial number of that type of opinion a binomial random variable. This analysis revisits Galams models from the point of view of the hypergeometric random variable by assuming the initial number of individuals in favor of an opinion is a fixed deterministic number. This assumption is more realistic, especially when analyzing small populations. Evolution of the models is based on majority rules, with a bias introduced when there is a tie. For the hier- archical voting system model, in order to derive the probability that opinion +1 would win, the analysis was done by reversing time and assuming that an individual in favor of opinion +1 wins. Then, working backwards we counted the number of configurations at the next lowest level that could induce each possible configuration at the level above, and continued this process until reaching the bottom level, i.e., the initial population. Using this method, we were able to derive an explicit formula for the probability that an individual in favor of opinion +1 wins given any initial count of that opinion, for any group size greater than or equal to three. For the public debate model, we counted the total number of individuals in favor of opinion +1 at each time step and used this variable to define a random walk. Then, we used first-step analysis to derive an explicit formula for the probability that an individual in favor of opinion +1 wins given any initial count of that opinion for group sizes of three. The spatial public debate model evolves based on the proportional rule. For the spatial model, the most natural graphical representation to construct the process results in a model that is not mathematically tractable. Thus, we defined a different graphical representation that is mathematically equivalent to the first graphical representation, but in this model it is possible to define a dual process that is mathematically tractable. Using this graphical representation we prove clustering in 1D and 2D and coexistence in higher dimensions following the same approach as for the voter model interacting particle system.
ContributorsTaylor, Nicole Robyn (Co-author) / Lanchier, Nicolas (Co-author, Thesis director) / Smith, Hal (Committee member) / Hurlbert, Glenn (Committee member) / Barrett, The Honors College (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2013-05
137354-Thumbnail Image.png
Description
The objective of the research presented here was to validate the use of kinetic models for the analysis of the dynamic behavior of a contrast agent in tumor tissue and evaluate the utility of such models in determining kinetic properties - in particular perfusion and molecular binding uptake associated with

The objective of the research presented here was to validate the use of kinetic models for the analysis of the dynamic behavior of a contrast agent in tumor tissue and evaluate the utility of such models in determining kinetic properties - in particular perfusion and molecular binding uptake associated with tissue hypoxia - of the imaged tissue, from concentration data acquired with dynamic contrast enhanced magnetic resonance imaging (DCE-MRI) procedure. Data from two separate DCE-MRI experiments, performed in the past, using a standard contrast agent and a hypoxia-binding agent respectively, were analyzed. The results of the analysis demonstrated that the models used may provide novel characterization of the tumor tissue properties. Future research will work to further characterize the physical significance of the estimated parameters, particularly to provide quantitative oxygenation data for the imaged tissue.
ContributorsMartin, Jonathan Michael (Author) / Kodibagkar, Vikram (Thesis director) / Rege, Kaushal (Committee member) / Barrett, The Honors College (Contributor) / Chemical Engineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2013-12
137044-Thumbnail Image.png
Description
In applications such as Magnetic Resonance Imaging (MRI), data are acquired as Fourier samples. Since the underlying images are only piecewise smooth, standard recon- struction techniques will yield the Gibbs phenomenon, which can lead to misdiagnosis. Although filtering will reduce the oscillations at jump locations, it can often have the

In applications such as Magnetic Resonance Imaging (MRI), data are acquired as Fourier samples. Since the underlying images are only piecewise smooth, standard recon- struction techniques will yield the Gibbs phenomenon, which can lead to misdiagnosis. Although filtering will reduce the oscillations at jump locations, it can often have the adverse effect of blurring at these critical junctures, which can also lead to misdiagno- sis. Incorporating prior information into reconstruction methods can help reconstruct a sharper solution. For example, compressed sensing (CS) algorithms exploit the expected sparsity of some features of the image. In this thesis, we develop a method to exploit the sparsity in the edges of the underlying image. We design a convex optimization problem that exploits this sparsity to provide an approximation of the underlying image. Our method successfully reduces the Gibbs phenomenon with only minimal "blurring" at the discontinuities. In addition, we see a high rate of convergence in smooth regions.
ContributorsWasserman, Gabriel Kanter (Author) / Gelb, Anne (Thesis director) / Cochran, Doug (Committee member) / Archibald, Rick (Committee member) / Barrett, The Honors College (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2014-05
137147-Thumbnail Image.png
Description
Smart contrast agents allow for noninvasive study of specific events or tissue conditions inside of a patient's body using Magnetic Resonance Imaging (MRI). This research aims to develop and characterize novel smart contrast agents for MRI that respond to temperature changes in tissue microenvironments. Transmission Electron Microscopy, Nuclear Magnetic Resonance,

Smart contrast agents allow for noninvasive study of specific events or tissue conditions inside of a patient's body using Magnetic Resonance Imaging (MRI). This research aims to develop and characterize novel smart contrast agents for MRI that respond to temperature changes in tissue microenvironments. Transmission Electron Microscopy, Nuclear Magnetic Resonance, and cell culture growth assays were used to characterize the physical, magnetic, and cytotoxic properties of candidate nanoprobes. The nanoprobes displayed thermosensitve MR properties with decreasing relaxivity with temperature. Future work will be focused on generating and characterizing photo-active analogues of the nanoprobes that could be used for both treatment of tissues and assessment of therapy.
ContributorsHussain, Khateeb Hyder (Author) / Kodibagkar, Vikram (Thesis director) / Stabenfeldt, Sarah (Committee member) / Barrett, The Honors College (Contributor) / Harrington Bioengineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor)
Created2014-05
136857-Thumbnail Image.png
Description
Glioblastoma Multiforme (GBM) is an aggressive and deadly form of brain cancer with a median survival time of about a year with treatment. Due to the aggressive nature of these tumors and the tendency of gliomas to follow white matter tracks in the brain, each tumor mass has a unique

Glioblastoma Multiforme (GBM) is an aggressive and deadly form of brain cancer with a median survival time of about a year with treatment. Due to the aggressive nature of these tumors and the tendency of gliomas to follow white matter tracks in the brain, each tumor mass has a unique growth pattern. Consequently it is difficult for neurosurgeons to anticipate where the tumor will spread in the brain, making treatment planning difficult. Archival patient data including MRI scans depicting the progress of tumors have been helpful in developing a model to predict Glioblastoma proliferation, but limited scans per patient make the tumor growth rate difficult to determine. Furthermore, patient treatment between scan points can significantly compound the challenge of accurately predicting the tumor growth. A partnership with Barrow Neurological Institute has allowed murine studies to be conducted in order to closely observe tumor growth and potentially improve the current model to more closely resemble intermittent stages of GBM growth without treatment effects.
ContributorsSnyder, Lena Haley (Author) / Kostelich, Eric (Thesis director) / Frakes, David (Committee member) / Barrett, The Honors College (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Harrington Bioengineering Program (Contributor)
Created2014-05
133482-Thumbnail Image.png
Description
Cryptocurrencies have become one of the most fascinating forms of currency and economics due to their fluctuating values and lack of centralization. This project attempts to use machine learning methods to effectively model in-sample data for Bitcoin and Ethereum using rule induction methods. The dataset is cleaned by removing entries

Cryptocurrencies have become one of the most fascinating forms of currency and economics due to their fluctuating values and lack of centralization. This project attempts to use machine learning methods to effectively model in-sample data for Bitcoin and Ethereum using rule induction methods. The dataset is cleaned by removing entries with missing data. The new column is created to measure price difference to create a more accurate analysis on the change in price. Eight relevant variables are selected using cross validation: the total number of bitcoins, the total size of the blockchains, the hash rate, mining difficulty, revenue from mining, transaction fees, the cost of transactions and the estimated transaction volume. The in-sample data is modeled using a simple tree fit, first with one variable and then with eight. Using all eight variables, the in-sample model and data have a correlation of 0.6822657. The in-sample model is improved by first applying bootstrap aggregation (also known as bagging) to fit 400 decision trees to the in-sample data using one variable. Then the random forests technique is applied to the data using all eight variables. This results in a correlation between the model and data of 9.9443413. The random forests technique is then applied to an Ethereum dataset, resulting in a correlation of 9.6904798. Finally, an out-of-sample model is created for Bitcoin and Ethereum using random forests, with a benchmark correlation of 0.03 for financial data. The correlation between the training model and the testing data for Bitcoin was 0.06957639, while for Ethereum the correlation was -0.171125. In conclusion, it is confirmed that cryptocurrencies can have accurate in-sample models by applying the random forests method to a dataset. However, out-of-sample modeling is more difficult, but in some cases better than typical forms of financial data. It should also be noted that cryptocurrency data has similar properties to other related financial datasets, realizing future potential for system modeling for cryptocurrency within the financial world.
ContributorsBrowning, Jacob Christian (Author) / Meuth, Ryan (Thesis director) / Jones, Donald (Committee member) / McCulloch, Robert (Committee member) / Computer Science and Engineering Program (Contributor) / School of Mathematical and Statistical Sciences (Contributor) / Barrett, The Honors College (Contributor)
Created2018-05