Security and Usability in Mobile and IoT Systems

158606-Thumbnail Image.png
Description
Mobile and Internet-of-Things (IoT) systems have been widely used in many aspects

of human’s life. These systems are storing and operating on more and more sensitive

data of users. Attackers may want to obtain the data to peek at users’ privacy or

pollute

Mobile and Internet-of-Things (IoT) systems have been widely used in many aspects

of human’s life. These systems are storing and operating on more and more sensitive

data of users. Attackers may want to obtain the data to peek at users’ privacy or

pollute the data to cause system malfunction. In addition, these systems are not

user-friendly for some people such as children, senior citizens, and visually impaired

users. Therefore, it is of cardinal significance to improve both security and usability

of mobile and IoT systems. This report consists of four parts: one automatic locking

system for mobile devices, one systematic study of security issues in crowdsourced

indoor positioning systems, one usable indoor navigation system, and practical attacks

on home alarm IoT systems.

Chapter 1 overviews the challenges and existing solutions in these areas. Chapater

2 introduces a novel system ilock which can automatically and immediately lock the

mobile devices to prevent data theft. Chapter 3 proposes attacks and countermeasures

for crowdsourced indoor positioning systems. Chapter 4 presents a context-aware indoor

navigation system which is more user-friendly for visual impaired people. Chapter

5 investigates some novel attacks on commercial home alarm systems. Chapter 6

concludes the report and discuss the future work.
Date Created
2020
Agent

Stochastic Analysis of Networked Systems

158599-Thumbnail Image.png
Description
This dissertation presents a novel algorithm for recovering missing values of co-evolving time series with partial embedded network information. The idea is to connect two sources of data through a shared low dimensional latent space. The proposed algorithm, named NetDyna,

This dissertation presents a novel algorithm for recovering missing values of co-evolving time series with partial embedded network information. The idea is to connect two sources of data through a shared low dimensional latent space. The proposed algorithm, named NetDyna, is an Expectation-Maximization algorithm, and uses the Kalman filter and matrix factorization approaches to infer the missing values both in the time series and embedded network. The experimental results on real datasets, including a Motes dataset and a Motion Capture dataset, show that (1) NetDyna outperforms other state-of-the-art algorithms, especially with partially observed network information; (2) its computational complexity scales linearly with the time duration of time series; and (3) the algorithm recovers the embedded network in addition to missing time series values.

This dissertation also studies a load balancing algorithm, the so called power-of-two-choices(Po2), for many-server systems (with N servers) and focuses on the convergence of stationary distribution of Po2 in the both light and heavy traffic regimes to the solution of mean-field system. The framework of Stein’s method and state space collapse (SSC) are used to analyze both regimes.

In both regimes, the thesis first uses the argument of state space collapse to show that the probability of the state being far from the mean-field solution is small enough. By a simple Markov inequality, it is able to show that the probability is indeed very small with a proper choice of parameters.

Then, for the state space close to the solution of mean-field model, the thesis uses Stein’s method to show that the stochastic system is close to a linear mean-field model. By characterizing the generator difference, it is able to characterize the dominant terms in both regimes. Note that for heavy traffic case, the lower and upper bound analysis of a tridiagonal matrix, which arises from the linear mean-field model, is needed. From the dominant term, it allows to calculate the coefficient of the convergence rate.

In the end, comparisons between the theoretical predictions and numerical simulations are presented.
Date Created
2020
Agent

Scheduling in Wireless and Healthcare Networks

158513-Thumbnail Image.png
Description
This dissertation studies the scheduling in two stochastic networks, a co-located wireless network and an outpatient healthcare network, both of which have a cyclic planning horizon and a deadline-related performance metric.

For the co-located wireless network, a time-slotted system is

This dissertation studies the scheduling in two stochastic networks, a co-located wireless network and an outpatient healthcare network, both of which have a cyclic planning horizon and a deadline-related performance metric.

For the co-located wireless network, a time-slotted system is considered. A cycle of planning horizon is called a frame, which consists of a fixed number of time slots. The size of the frame is determined by the upper-layer applications. Packets with deadlines arrive at the beginning of each frame and will be discarded if missing their deadlines, which are in the same frame. Each link of the network is associated with a quality of service constraint and an average transmit power constraint. For this system, a MaxWeight-type problem for which the solutions achieve the throughput optimality is formulated. Since the computational complexity of solving the MaxWeight-type problem with exhaustive search is exponential even for a single-link system, a greedy algorithm with complexity O(nlog(n)) is proposed, which is also throughput optimal.

The outpatient healthcare network is modeled as a discrete-time queueing network, in which patients receive diagnosis and treatment planning that involves collaboration between multiple service stations. For each patient, only the root (first) appointment can be scheduled as the following appointments evolve stochastically. The cyclic planing horizon is a week. The root appointment is optimized to maximize the proportion of patients that can complete their care by a class-dependent deadline. In the optimization algorithm, the sojourn time of patients in the healthcare network is approximated with a doubly-stochastic phase-type distribution. To address the computational intractability, a mean-field model with convergence guarantees is proposed. A linear programming-based policy improvement framework is developed, which can approximately solve the original large-scale stochastic optimization in queueing networks of realistic sizes.
Date Created
2020
Agent

Coordinated Operation of the Electric Power System with Water Distribution Systems: Modeling, Control, Simulation, and Quantification of Resilience

158388-Thumbnail Image.png
Description
The electric power system (EPS) is an extremely complex system that has operational interdependencies with the water delivery and treatment system (WDTS). The term water-energy nexus is commonly used to describe the critical interdependencies that naturally exist between the EPS

The electric power system (EPS) is an extremely complex system that has operational interdependencies with the water delivery and treatment system (WDTS). The term water-energy nexus is commonly used to describe the critical interdependencies that naturally exist between the EPS and water distribution systems (WDS). Presented in this work is a framework for simulating interactions between these two critical infrastructure systems in short-term and long-term time-scales. This includes appropriate mathematical models for system modeling and for optimizing control of power system operation with consideration of conditions in the WDS. Also presented is a complete methodology for quantifying the resilience of the two interdependent systems.

The key interdependencies between the two systems are the requirements of water for the cooling cycle of traditional thermal power plants as well as electricity for pumping and/or treatment in the WDS. While previous work has considered the dependency of thermoelectric generation on cooling water requirements at a high-level, this work considers the impact from limitations of cooling water into network simulations in both a short-term operational framework as well as in the long-term planning domain.

The work completed to set-up simulations in operational length time-scales was the development of a simulator that adequately models both systems. This simulation engine also facilitates the implementation of control schemes in both systems that take advantage of the knowledge of operating conditions in the other system. Initial steps for including the influence of anticipated water availability and water rights attainability within the combined generation and transmission expansion planning problem is also presented. Lastly, the framework for determining the infrastructural-operational resilience (IOR) of the interdependent systems is formulated.

Adequately modeling and studying the two systems and their interactions is becoming critically important. This importance is illustrated by the possibility of unforeseen natural or man-made events or by the likelihood of load increase in the systems, either of which has the risk of putting extreme stress on the systems beyond that experienced in normal operating conditions. Therefore, this work addresses these concerns with novel modeling and control/policy strategies designed to mitigate the severity of extreme conditions in either system.
Date Created
2020
Agent

Quantifying Information Leakage via Adversarial Loss Functions: Theory and Practice

158139-Thumbnail Image.png
Description
Modern digital applications have significantly increased the leakage of private and sensitive personal data. While worst-case measures of leakage such as Differential Privacy (DP) provide the strongest guarantees, when utility matters, average-case information-theoretic measures can be more relevant. However, most

Modern digital applications have significantly increased the leakage of private and sensitive personal data. While worst-case measures of leakage such as Differential Privacy (DP) provide the strongest guarantees, when utility matters, average-case information-theoretic measures can be more relevant. However, most such information-theoretic measures do not have clear operational meanings. This dissertation addresses this challenge.

This work introduces a tunable leakage measure called maximal $\alpha$-leakage which quantifies the maximal gain of an adversary in inferring any function of a data set. The inferential capability of the adversary is modeled by a class of loss functions, namely, $\alpha$-loss. The choice of $\alpha$ determines specific adversarial actions ranging from refining a belief for $\alpha =1$ to guessing the best posterior for $\alpha = \infty$, and for the two specific values maximal $\alpha$-leakage simplifies to mutual information and maximal leakage, respectively. Maximal $\alpha$-leakage is proved to have a composition property and be robust to side information.

There is a fundamental disjoint between theoretical measures of information leakages and their applications in practice. This issue is addressed in the second part of this dissertation by proposing a data-driven framework for learning Censored and Fair Universal Representations (CFUR) of data. This framework is formulated as a constrained minimax optimization of the expected $\alpha$-loss where the constraint ensures a measure of the usefulness of the representation. The performance of the CFUR framework with $\alpha=1$ is evaluated on publicly accessible data sets; it is shown that multiple sensitive features can be effectively censored to achieve group fairness via demographic parity while ensuring accuracy for several \textit{a priori} unknown downstream tasks.

Finally, focusing on worst-case measures, novel information-theoretic tools are used to refine the existing relationship between two such measures, $(\epsilon,\delta)$-DP and R\'enyi-DP. Applying these tools to the moments accountant framework, one can track the privacy guarantee achieved by adding Gaussian noise to Stochastic Gradient Descent (SGD) algorithms. Relative to state-of-the-art, for the same privacy budget, this method allows about 100 more SGD rounds for training deep learning models.
Date Created
2020
Agent

Fundamental Limits of Gaussian Communication Networks in the Presence of Intelligent Jammers

157976-Thumbnail Image.png
Description
The open nature of the wireless communication medium makes it inherently vulnerable to an active attack, wherein a malicious adversary (or jammer) transmits into the medium to disrupt the operation of the legitimate users. Therefore, developing techniques to manage the

The open nature of the wireless communication medium makes it inherently vulnerable to an active attack, wherein a malicious adversary (or jammer) transmits into the medium to disrupt the operation of the legitimate users. Therefore, developing techniques to manage the presence of a jammer and to characterize the effect of an attacker on the fundamental limits of wireless communication networks is important. This dissertation studies various Gaussian communication networks in the presence of such an adversarial jammer.

First of all, a standard Gaussian channel is considered in the presence of a jammer, known as a Gaussian arbitrarily-varying channel, but with list-decoding at the receiver. The receiver decodes a list of messages, instead of only one message, with the goal of the correct message being an element of the list. The capacity is characterized, and it is shown that under some transmitter's power constraints the adversary is able to suspend the communication between the legitimate users and make the capacity zero.

Next, generalized packing lemmas are introduced for Gaussian adversarial channels to achieve the capacity bounds for three Gaussian multi-user channels in the presence of adversarial jammers. Inner and outer bounds on the capacity regions of Gaussian multiple-access channels, Gaussian broadcast channels, and Gaussian interference channels are derived in the presence of malicious jammers. For the Gaussian multiple-access channels with jammer, the capacity bounds coincide. In this dissertation, the adversaries can send any arbitrary signals to the channel while none of the transmitter and the receiver knows the adversarial signals' distribution.

Finally, the capacity of the standard point-to-point Gaussian fading channel in the presence of one jammer is investigated under multiple scenarios of channel state information availability, which is the knowledge of exact fading coefficients. The channel state information is always partially or fully known at the receiver to decode the message while the transmitter or the adversary may or may not have access to this information. Here, the adversary model is the same as the previous cases with no knowledge about the user's transmitted signal except possibly the knowledge of the fading path.
Date Created
2019
Agent

Steady State Analysis of Load Balancing Algorithms in the Heavy Traffic Regime

157816-Thumbnail Image.png
Description
This dissertation studies load balancing algorithms for many-server systems (with N servers) and focuses on the steady-state performance of load balancing algorithms in the heavy traffic regime. The framework of Stein’s method and (iterative) state-space collapse (SSC) are used to

This dissertation studies load balancing algorithms for many-server systems (with N servers) and focuses on the steady-state performance of load balancing algorithms in the heavy traffic regime. The framework of Stein’s method and (iterative) state-space collapse (SSC) are used to analyze three load balancing systems: 1) load balancing in the Sub-Halfin-Whitt regime with exponential service time; 2) load balancing in the Beyond-Halfin-Whitt regime with exponential service time; 3) load balancing in the Sub-Halfin-Whitt regime with Coxian-2 service time.

When in the Sub-Halfin-Whitt regime, the sufficient conditions are established such that any load balancing algorithm that satisfies the conditions have both asymptotic zero waiting time and zero waiting probability. Furthermore, the number of servers with more than one jobs is o(1), in other words, the system collapses to a one-dimensional space. The result is proven using Stein’s method and state space collapse (SSC), which are powerful mathematical tools for steady-state analysis of load balancing algorithms. The second system is in even “heavier” traffic regime, and an iterative refined procedure is proposed to obtain the steady-state metrics. Again, asymptotic zero delay and waiting are established for a set of load balancing algorithms. Different from the first system, the system collapses to a two-dimensional state-space instead of one-dimensional state-space. The third system is more challenging because of “non-monotonicity” with Coxian-2 service time, and an iterative state space collapse is proposed to tackle the “non-monotonicity” challenge. For these three systems, a set of load balancing algorithms is established, respectively, under which the probability that an incoming job is routed to an idle server is one asymptotically at steady-state. The set of load balancing algorithms includes join-the-shortest-queue (JSQ), idle-one-first(I1F), join-the-idle-queue (JIQ), and power-of-d-choices (Pod) with a carefully-chosen d.
Date Created
2019
Agent

Water Supply Infrastructure Modeling and Control under Extreme Drought and/or Limited Power Availability

157075-Thumbnail Image.png
Description
The phrase water-energy nexus is commonly used to describe the inherent and critical interdependencies between the electric power system and the water supply systems (WSS). The key interdependencies between the two systems are the power plant’s requirement of water for

The phrase water-energy nexus is commonly used to describe the inherent and critical interdependencies between the electric power system and the water supply systems (WSS). The key interdependencies between the two systems are the power plant’s requirement of water for the cooling cycle and the water system’s need of electricity for pumping for water supply. While previous work has considered the dependency of WSS on the electrical power, this work incorporates into an optimization-simulation framework, consideration of the impact of short and long-term limited availability of water and/or electrical energy.

This research focuses on the water supply system (WSS) facet of the multi-faceted optimization and control mechanism developed for an integrated water – energy nexus system under U.S. National Science Foundation (NSF) project 029013-0010 CRISP Type 2 – Resilient cyber-enabled electric energy and water infrastructures modeling and control under extreme mega drought scenarios. A water supply system (WSS) conveys water from sources (such as lakes, rivers, dams etc.) to the treatment plants and then to users via the water distribution systems (WDS) and/or water supply canal systems (WSCS). Optimization-simulation methodologies are developed for the real-time operation of water supply systems (WSS) under critical conditions of limited electrical energy and/or water availability due to emergencies such as extreme drought conditions, electric grid failure, and other severe conditions including natural and manmade disasters. The coupling between WSS and the power system was done through alternatively exchanging data between the power system and WSS simulations via a program control overlay developed in python.

A new methodology for WDS infrastructural-operational resilience (IOR) computation was developed as a part of this research to assess the real-time performance of the WDS under emergency conditions. The methodology combines operational resilience and component level infrastructural robustness to provide a comprehensive performance assessment tool.

The optimization-simulation and resilience computation methodologies developed were tested for both hypothetical and real example WDS and WSCS, with results depicting improved resilience for operations of the WSS under normal and emergency conditions.
Date Created
2019
Agent

Towards Developing Computer Vision Algorithms and Architectures for Real-world Applications

156887-Thumbnail Image.png
Description
Computer vision technology automatically extracts high level, meaningful information from visual data such as images or videos, and the object recognition and detection algorithms are essential in most computer vision applications. In this dissertation, we focus on developing algorithms used

Computer vision technology automatically extracts high level, meaningful information from visual data such as images or videos, and the object recognition and detection algorithms are essential in most computer vision applications. In this dissertation, we focus on developing algorithms used for real life computer vision applications, presenting innovative algorithms for object segmentation and feature extraction for objects and actions recognition in video data, and sparse feature selection algorithms for medical image analysis, as well as automated feature extraction using convolutional neural network for blood cancer grading.

To detect and classify objects in video, the objects have to be separated from the background, and then the discriminant features are extracted from the region of interest before feeding to a classifier. Effective object segmentation and feature extraction are often application specific, and posing major challenges for object detection and classification tasks. In this dissertation, we address effective object flow based ROI generation algorithm for segmenting moving objects in video data, which can be applied in surveillance and self driving vehicle areas. Optical flow can also be used as features in human action recognition algorithm, and we present using optical flow feature in pre-trained convolutional neural network to improve performance of human action recognition algorithms. Both algorithms outperform the state-of-the-arts at their time.

Medical images and videos pose unique challenges for image understanding mainly due to the fact that the tissues and cells are often irregularly shaped, colored, and textured, and hand selecting most discriminant features is often difficult, thus an automated feature selection method is desired. Sparse learning is a technique to extract the most discriminant and representative features from raw visual data. However, sparse learning with \textit{L1} regularization only takes the sparsity in feature dimension into consideration; we improve the algorithm so it selects the type of features as well; less important or noisy feature types are entirely removed from the feature set. We demonstrate this algorithm to analyze the endoscopy images to detect unhealthy abnormalities in esophagus and stomach, such as ulcer and cancer. Besides sparsity constraint, other application specific constraints and prior knowledge may also need to be incorporated in the loss function in sparse learning to obtain the desired results. We demonstrate how to incorporate similar-inhibition constraint, gaze and attention prior in sparse dictionary selection for gastroscopic video summarization that enable intelligent key frame extraction from gastroscopic video data. With recent advancement in multi-layer neural networks, the automatic end-to-end feature learning becomes feasible. Convolutional neural network mimics the mammal visual cortex and can extract most discriminant features automatically from training samples. We present using convolutinal neural network with hierarchical classifier to grade the severity of Follicular Lymphoma, a type of blood cancer, and it reaches 91\% accuracy, on par with analysis by expert pathologists.

Developing real world computer vision applications is more than just developing core vision algorithms to extract and understand information from visual data; it is also subject to many practical requirements and constraints, such as hardware and computing infrastructure, cost, robustness to lighting changes and deformation, ease of use and deployment, etc.The general processing pipeline and system architecture for the computer vision based applications share many similar design principles and architecture. We developed common processing components and a generic framework for computer vision application, and a versatile scale adaptive template matching algorithm for object detection. We demonstrate the design principle and best practices by developing and deploying a complete computer vision application in real life, building a multi-channel water level monitoring system, where the techniques and design methodology can be generalized to other real life applications. The general software engineering principles, such as modularity, abstraction, robust to requirement change, generality, etc., are all demonstrated in this research.
Date Created
2018
Agent

Security and Privacy in Mobile Devices: Novel Attacks and Countermeasures

156796-Thumbnail Image.png
Description
Mobile devices have penetrated into every aspect of modern world. For one thing, they are becoming ubiquitous in daily life. For the other thing, they are storing more and more data, including sensitive data. Therefore, security and privacy of mobile

Mobile devices have penetrated into every aspect of modern world. For one thing, they are becoming ubiquitous in daily life. For the other thing, they are storing more and more data, including sensitive data. Therefore, security and privacy of mobile devices are indispensable. This dissertation consists of five parts: two authentication schemes, two attacks, and one countermeasure related to security and privacy of mobile devices.

Specifically, in Chapter 1, I give an overview the challenges and existing solutions in these areas. In Chapter 2, a novel authentication scheme is presented, which is based on a user’s tapping or sliding on the touchscreen of a mobile device. In Chapter 3, I focus on mobile app fingerprinting and propose a method based on analyzing the power profiles of targeted mobile devices. In Chapter 4, I mainly explore a novel liveness detection method for face authentication on mobile devices. In Chapter 5, I investigate a novel keystroke inference attack on mobile devices based on user eye movements. In Chapter 6, a novel authentication scheme is proposed, based on detecting a user’s finger gesture through acoustic sensing. In Chapter 7, I discuss the future work.
Date Created
2018
Agent