Not This Exit: Analyzing the Impact of VPN Exit IPs on Network Alchemy

193476-Thumbnail Image.png
Description
Virtual Private Networks (VPNs) are used in a wide range of applications, rangingfrom commercial applications like accessing resources remotely to security and pri- vacy for targeted users like journalists, Non-governmental organizations (NGOs), etc. However, VPNs were not inherently designed with security in

Virtual Private Networks (VPNs) are used in a wide range of applications, rangingfrom commercial applications like accessing resources remotely to security and pri- vacy for targeted users like journalists, Non-governmental organizations (NGOs), etc. However, VPNs were not inherently designed with security in mind. The interaction between the kernel processes and the connection tracking framework is uncoordi- nated. This leaves VPNs vulnerable to certain attacks due to their implementation. This work explores the extent to which these attacks are possible on certain imple- mentations of VPN servers which have a separate exit IP and entry IP on the VPN server. Further, this work also formally models the VPN connection tracking behavior between servers and clients. The formal models enables a deeper analysis to identify exactly at what point of the VPN process the vulnerabilities are introduced and if the instances of VPN which have separate entry and exit IPs are still vulnerable to the same attacks. Through simulations done in a virtual lab environment and testing on formal models, it is observed that having a separate exit and entry IP leaves may affect the practicality of certain attacks.
Date Created
2024
Agent

Speedcuber Timer: Creating an Open-Source Platform for Smart Rubik’s Cube Applications

187330-Thumbnail Image.png
Description
Since the early 2000s the Rubik’s Cube has seen growing usage at speedsolving competitions and as an effective tool to teach Science, Technology, Engineering, Mathematics (STEM) topics at hundreds of schools and universities across the world. Recently, cube manufacturers have

Since the early 2000s the Rubik’s Cube has seen growing usage at speedsolving competitions and as an effective tool to teach Science, Technology, Engineering, Mathematics (STEM) topics at hundreds of schools and universities across the world. Recently, cube manufacturers have begun embedding sensors to enable digital face tracking. The live feedback from these so called “smartcubes” enables a new wave of immersive solution tutorials and interactive educational games using the cube as a controller. Existing smartcube software has several limitations. Manufacturers’ applications support only a narrow set of puzzle form factors and application platforms, fragmenting the ecosystem. Most apps require an active internet connection for key features, limiting where users can practice with a smartcube. Finally, existing applications focus on a single 3x3x3connection, losing opportunities afforded by new form factors. This research demonstrates an open-source smartcube application which mitigates these limitations. Particular attention is given to creating an Application Programming Interface (API) for smartcube communication and building representative solve analysis tools. These innovations have included successful negotiations to re-license existing open-source Rubik’sCube software projects to support deployment on multiple platforms, particularly iOS. The resulting application supports smartcubes from three manufacturers, runs on two platforms (Android and iOS), functions entirely offline after an initial download of remote assets, demonstrates concurrent connections with up to six smartcubes, and supports all current and anticipated smartcube form factors. These foundational elements can accelerate future efforts to build smartcube applications, including automated performance feedback systems and personalized gamification of learning experiences. Such advances will hopefully enhance the Rubik’s Cube’s value both as a competitive toy and as a pedagogical tool in educational institutions worldwide.
Date Created
2023
Agent

Investigating the Utility of Agile and Lean Software Process Metrics for Open Source Software Communities: An Exploratory Study

171448-Thumbnail Image.png
Description
The adoption of Open Source Software (OSS) by organizations has become a strategic need in a wide variety of software applications and platforms. Open Source has changed the way organizations develop, acquire, use, and commercialize software. Further, OSS projects often

The adoption of Open Source Software (OSS) by organizations has become a strategic need in a wide variety of software applications and platforms. Open Source has changed the way organizations develop, acquire, use, and commercialize software. Further, OSS projects often incorporate similar principles and practices as Agile and Lean software development projects. Contrary to traditional organizations, the environment in which these projects function has an impact on process-related elements like the flow of work and value definition. Process metrics are typically employed during Agile Software Engineering projects as a means of providing meaningful feedback. Investigating these metrics to see if OSS projects and communities can utilize them in a beneficial way thus becomes an interesting research topic. In that context, this exploratory research investigates whether well-established Agile and Lean software engineering metrics provide useful feedback about OSS projects. This knowledge will assist in educating the Open Source community about the applications of Agile Software Engineering and its variations in Open Source projects. Each of the Open Source projects included in this analysis has a substantial development team that maintains a mature, well-established codebase with process flow information. These OSS projects listed on GitHub are investigated by applying process flow metrics. The methodology used to collect these metrics and relevant findings are discussed in this thesis. This study also compares the results to distinctive Open Source project characteristics as part of the analysis. In this exploratory research best-fit versions of published Agile and Lean software process metrics are applied to OSS, and following these explorations, specific questions are further addressed using the data collected. This research's original contribution is to determine whether Agile and Lean process metrics are helpful in OSS, as well as the opportunities and obstacles that may arise when applying Agile and Lean principles to OSS.
Date Created
2022
Agent

Predicting Student Dropout in Self-Paced MOOC Course

161629-Thumbnail Image.png
Description
One persisting problem in Massive Open Online Courses (MOOCs) is the issue of student dropout from these courses. The prediction of student dropout from MOOC courses can identify the factors responsible for such an event and it can further initiate

One persisting problem in Massive Open Online Courses (MOOCs) is the issue of student dropout from these courses. The prediction of student dropout from MOOC courses can identify the factors responsible for such an event and it can further initiate intervention before such an event to increase student success in MOOC. There are different approaches and various features available for the prediction of student’s dropout in MOOC courses.In this research, the data derived from the self-paced math course ‘College Algebra and Problem Solving’ offered on the MOOC platform Open edX offered by Arizona State University (ASU) from 2016 to 2020 was considered. This research aims to predict the dropout of students from a MOOC course given a set of features engineered from the learning of students in a day. Machine Learning (ML) model used is Random Forest (RF) and this model is evaluated using the validation metrics like accuracy, precision, recall, F1-score, Area Under the Curve (AUC), Receiver Operating Characteristic (ROC) curve. The average rate of student learning progress was found to have more impact than other features. The model developed can predict the dropout or continuation of students on any given day in the MOOC course with an accuracy of 87.5%, AUC of 94.5%, precision of 88%, recall of 87.5%, and F1-score of 87.5% respectively. The contributing features and interactions were explained using Shapely values for the prediction of the model. The features engineered in this research are predictive of student dropout and could be used for similar courses to predict student dropout from the course. This model can also help in making interventions at a critical time to help students succeed in this MOOC course.
Date Created
2021
Agent

Compass Portal

Description

Compass portal features tools that help teachers, psychologists, behavioral specialists gain insights on students’ performance through activities they have completed.

Date Created
2021-05
Agent

Compass Portal

Description

COMPASS portal features tools that help teachers, psychologists, behavioral Specialists gain insights on students’ performance through activities they have completed.

Date Created
2021-05
Agent

Adaptive mHealth interventions for improving youth responsiveness and clinical outcomes

157565-Thumbnail Image.png
Description
Mobile health (mHealth) applications (apps) hold tremendous potential for addressing chronic health conditions. Smartphones are now the most popular form of computing, and the ubiquitous “always with us, always on” nature of mobile technology makes them amenable to interventions aimed

Mobile health (mHealth) applications (apps) hold tremendous potential for addressing chronic health conditions. Smartphones are now the most popular form of computing, and the ubiquitous “always with us, always on” nature of mobile technology makes them amenable to interventions aimed and managing chronic disease. Several challenges exist, however, such as the difficulty in determining mHealth effects due to the rapidly changing nature of the technology and the challenges presented to existing methods of evaluation, and the ability to ensure end users consistently use the technology in order to achieve the desired effects. The latter challenge is in adherence, defined as the extent to which a patient conducts the activities defined in a clinical protocol (i.e. an intervention plan). Further, higher levels of adherence should lead to greater effects of the intervention (the greater fidelity to the protocol, the more benefit one should receive from the protocol). mHealth has limitations in these areas; the ability to have patients sustainably adhere to a protocol, and the ability to drive intervention effect sizes. My research considers personalized interventions, a new approach of study in the mHealth community, as a potential remedy to these limitations. Specifically, in the context of a pediatric preventative anxiety protocol, I introduce algorithms to drive greater levels of adherence and greater effect sizes by incorporating per-patient (personalized) information. These algorithms have been implemented within an existing mHealth app for middle school that has been successfully deployed in a school in the Phoenix Arizona metropolitan area. The number of users is small (n=3) so a case-by-case analysis of app usage is presented. In addition simulated user behaviors based on models of adherence and effects sizes over time are presented as a means to demonstrate the potential impact of personalized deployments on a larger scale.
Date Created
2019
Agent

Template-Based Question Answering over Linked Data using Recursive Neural Networks

156879-Thumbnail Image.png
Description
The Semantic Web contains large amounts of related information in the form of knowledge graphs such as DBpedia. These knowledge graphs are typically enormous and are not easily accessible for users as they need specialized knowledge in query languages (such

The Semantic Web contains large amounts of related information in the form of knowledge graphs such as DBpedia. These knowledge graphs are typically enormous and are not easily accessible for users as they need specialized knowledge in query languages (such as SPARQL) as well as deep familiarity of the ontologies used by these knowledge graphs. So, to make these knowledge graphs more accessible (even for non- experts) several question answering (QA) systems have been developed over the last decade. Due to the complexity of the task, several approaches have been undertaken that include techniques from natural language processing (NLP), information retrieval (IR), machine learning (ML) and the Semantic Web (SW). At a higher level, most question answering systems approach the question answering task as a conversion from the natural language question to its corresponding SPARQL query. These systems then utilize the query to retrieve the desired entities or literals. One approach to solve this problem, that is used by most systems today, is to apply deep syntactic and semantic analysis on the input question to derive the SPARQL query. This has resulted in the evolution of natural language processing pipelines that have common characteristics such as answer type detection, segmentation, phrase matching, part-of-speech-tagging, named entity recognition, named entity disambiguation, syntactic or dependency parsing, semantic role labeling, etc.

This has lead to NLP pipeline architectures that integrate components that solve a specific aspect of the problem and pass on the results to subsequent components for further processing eg: DBpedia Spotlight for named entity recognition, RelMatch for relational mapping, etc. A major drawback in this approach is error propagation that is a common problem in NLP. This can occur due to mistakes early on in the pipeline that can adversely affect successive steps further down the pipeline. Another approach is to use query templates either manually generated or extracted from existing benchmark datasets such as Question Answering over Linked Data (QALD) to generate the SPARQL queries that is basically a set of predefined queries with various slots that need to be filled. This approach potentially shifts the question answering problem into a classification task where the system needs to match the input question to the appropriate template (class label).

This thesis proposes a neural network approach to automatically learn and classify natural language questions into its corresponding template using recursive neural networks. An obvious advantage of using neural networks is the elimination for the need of laborious feature engineering that can be cumbersome and error prone. The input question would be encoded into a vector representation. The model will be trained and evaluated on the LC-QuAD Dataset (Large-scale Complex Question Answering Dataset). The dataset was created explicitly for machine learning based QA approaches for learning complex SPARQL queries. The dataset consists of 5000 questions along with their corresponding SPARQL queries over the DBpedia dataset spanning 5042 entities and 615 predicates. These queries were annotated based on 38 unique templates that the model will attempt to classify. The resulting model will be evaluated against both the LC-QuAD dataset and the Question Answering Over Linked Data (QALD-7) dataset.

The recursive neural network achieves template classification accuracy of 0.828 on the LC-QuAD dataset and an accuracy of 0.618 on the QALD-7 dataset. When the top-2 most likely templates were considered the model achieves an accuracy of 0.945 on the LC-QuAD dataset and 0.786 on the QALD-7 dataset.

After slot filling, the overall system achieves a macro F-score 0.419 on the LC- QuAD dataset and a macro F-score of 0.417 on the QALD-7 dataset.
Date Created
2018
Agent

State Press Live

133874-Thumbnail Image.png
Description
As our relationship with technology continues to encourage people to spend more time engaged online, traditional means of journalism must adapt in order to communicate with audiences. While many news organizations default to social media outlets, the goal of this

As our relationship with technology continues to encourage people to spend more time engaged online, traditional means of journalism must adapt in order to communicate with audiences. While many news organizations default to social media outlets, the goal of this project is to allow users a more direct experience with reporters, photographers and editors. It will allow The State Press, the official, student-run news organization covering ASU, to create content within Slack, an internal messaging platform commonly used in newsrooms. Secondly, it will provide a means for viewers to conveniently ingest their news as it unfolds, with updates, media, and analysis appearing in front of them without having to refresh the page.
Date Created
2018-05
Agent

MEASURING AIR QUALITY USING WIRELESS SELF-POWERED DEVICES

136225-Thumbnail Image.png
Description
High concentrations of carbon monoxide and particulate matter can cause respiratory disease, illness, and death in high doses. Air pollution is a concern in many urban areas of emerging markets that rely on outdated technologies for transportation and electricity generation;

High concentrations of carbon monoxide and particulate matter can cause respiratory disease, illness, and death in high doses. Air pollution is a concern in many urban areas of emerging markets that rely on outdated technologies for transportation and electricity generation; rural air quality is also a concern when noting the high prevalence of products of incomplete combustion resulting from open fires for cooking and heating. Monitoring air quality is an essential step to identifying these and other factors that affect air quality, and thereafter informing engineering and policy decisions to improve the quality of air. This study seeks to measure changes in air quality across spatial and temporal domains, with a specific focus on microclimates within an urban area. A prototype, low-cost air quality monitoring device has been developed to measure the concentrations of particulate matter, ozone, and carbon monoxide multiple times per minute. The device communicates data wirelessly via cell towers, and can run off-grid using a solar PV-battery system. The device can be replicated and deployed across urban regions for high-fidelity emissions monitoring to explore the effect of anthropogenic and environmental factors on intra-hour air quality. Hardware and software used in the device is described, and the wireless data communication protocols and capabilities are discussed.
Date Created
2015-05
Agent