Filtering by
- Creators: Arizona State University
- Creators: Department of Information Systems
- Creators: Armfield, Jessica Ann
Choropleth maps are a common form of online cartographic visualization. They reveal patterns in spatial distributions of a variable by associating colors with data values measured at areal units. Although this capability of pattern revelation has popularized the use of choropleth maps, existing methods for their online delivery are limited in supporting dynamic map generation from large areal data. This limitation has become increasingly problematic in online choropleth mapping as access to small area statistics, such as high-resolution census data and real-time aggregates of geospatial data streams, has never been easier due to advances in geospatial web technologies. The current literature shows that the challenge of large areal data can be mitigated through tiled maps where pre-processed map data are hierarchically partitioned into tiny rectangular images or map chunks for efficient data transmission. Various approaches have emerged lately to enable this tile-based choropleth mapping, yet little empirical evidence exists on their ability to handle spatial data with large numbers of areal units, thus complicating technical decision making in the development of online choropleth mapping applications. To fill this knowledge gap, this dissertation study conducts a scalability evaluation of three tile-based methods discussed in the literature: raster, scalable vector graphics (SVG), and HTML5 Canvas. For the evaluation, the study develops two test applications, generates map tiles from five different boundaries of the United States, and measures the response times of the applications under multiple test operations. While specific to the experimental setups of the study, the evaluation results show that the raster method scales better across various types of user interaction than the other methods. Empirical evidence also points to the superior scalability of Canvas to SVG in dynamic rendering of vector tiles, but not necessarily for partial updates of the tiles. These findings indicate that the raster method is better suited for dynamic choropleth rendering from large areal data, while Canvas would be more suitable than SVG when such rendering frequently involves complete updates of vector shapes.
This study examines validity evidence of a state policy-directed teacher evaluation system implemented in Arizona during school year 2012-2013. The purpose was to evaluate the warrant for making high stakes, consequential judgments of teacher competence based on value-added (VAM) estimates of instructional impact and observations of professional practice (PP). The research also explores educator influence (voice) in evaluation design and the role information brokers have in local decision making. Findings are situated in an evidentiary and policy context at both the LEA and state policy levels.
The study employs a single-phase, concurrent, mixed-methods research design triangulating multiple sources of qualitative and quantitative evidence onto a single (unified) validation construct: Teacher Instructional Quality. It focuses on assessing the characteristics of metrics used to construct quantitative ratings of instructional competence and the alignment of stakeholder perspectives to facets implicit in the evaluation framework. Validity examinations include assembly of criterion, content, reliability, consequential and construct articulation evidences. Perceptual perspectives were obtained from teachers, principals, district leadership, and state policy decision makers. Data for this study came from a large suburban public school district in metropolitan Phoenix, Arizona.
Study findings suggest that the evaluation framework is insufficient for supporting high stakes, consequential inferences of teacher instructional quality. This is based, in part on the following: (1) Weak associations between VAM and PP metrics; (2) Unstable VAM measures across time and between tested content areas; (3) Less than adequate scale reliabilities; (4) Lack of coherence between theorized and empirical PP factor structures; (5) Omission/underrepresentation of important instructional attributes/effects; (6) Stakeholder concerns over rater consistency, bias, and the inability of test scores to adequately represent instructional competence; (7) Negative sentiments regarding the system's ability to improve instructional competence and/or student learning; (8) Concerns regarding unintended consequences including increased stress, lower morale, harm to professional identity, and restricted learning opportunities; and (9) The general lack of empowerment and educator exclusion from the decision making process. Study findings also highlight the value of information brokers in policy decision making and the importance of having access to unbiased empirical information during the design and implementation phases of important change initiatives.
are constantly changing, and adapting to these changes in an academic curriculum
can be challenging. Given a specific aspect of a domain, there can be various levels of
proficiency that can be achieved by the students. Considering the wide array of needs,
diverse groups need customized course curriculum. The need for having an archetype
to design a course focusing on the outcomes paved the way for Outcome-based
Education (OBE). OBE focuses on the outcomes as opposed to the traditional way of
following a process [23]. According to D. Clark, the major reason for the creation of
Bloom’s taxonomy was not only to stimulate and inspire a higher quality of thinking
in academia – incorporating not just the basic fact-learning and application, but also
to evaluate and analyze on the facts and its applications [7]. Instructional Module
Development System (IMODS) is the culmination of both these models – Bloom’s
Taxonomy and OBE. It is an open-source web-based software that has been
developed on the principles of OBE and Bloom’s Taxonomy. It guides an instructor,
step-by-step, through an outcomes-based process as they define the learning
objectives, the content to be covered and develop an instruction and assessment plan.
The tool also provides the user with a repository of techniques based on the choices
made by them regarding the level of learning while defining the objectives. This helps
in maintaining alignment among all the components of the course design. The tool
also generates documentation to support the course design and provide feedback
when the course is lacking in certain aspects.
It is not just enough to come up with a model that theoretically facilitates
effective result-oriented course design. There should be facts, experiments and proof
that any model succeeds in achieving what it aims to achieve. And thus, there are two
research objectives of this thesis: (i) design a feature for course design feedback and
evaluate its effectiveness; (ii) evaluate the usefulness of a tool like IMODS on various
aspects – (a) the effectiveness of the tool in educating instructors on OBE; (b) the
effectiveness of the tool in providing appropriate and efficient pedagogy and
assessment techniques; (c) the effectiveness of the tool in building the learning
objectives; (d) effectiveness of the tool in document generation; (e) Usability of the
tool; (f) the effectiveness of OBE on course design and expected student outcomes.
The thesis presents a detailed algorithm for course design feedback, its pseudocode, a
description and proof of the correctness of the feature, methods used for evaluation
of the tool, experiments for evaluation and analysis of the obtained results.
To boost students’ learning experience, adaptive selection was built on the generated questions. Bayesian Knowledge Tracing was used as embedded assessment of the student’s current competence so that a suitable question could be selected based on the student’s previous performance. A between-subjects experiment with 42 participants was performed, where half of the participants studied with adaptive selected questions and the rest studied with mal-adaptive order of questions. Both groups significantly improved their test scores, and the participants in adaptive group registered larger learning gains than participants in the control group.
To explore the possibility of generating rich instructional feedback for machine-generated questions, a question-paragraph mapping task was identified. Given a set of questions and a list of paragraphs for a textbook, the goal of the task was to map the related paragraphs to each question. An algorithm was developed whose performance was comparable to human annotators.
A multiple-choice question with high quality distractors (incorrect answers) can be pedagogically valuable as well as being much easier to grade than open-response questions. Thus, an algorithm was developed to generate good distractors for multiple-choice questions. The machine-generated multiple-choice questions were compared to human-generated questions in terms of three measures: question difficulty, question discrimination and distractor usefulness. By recruiting 200 participants from Amazon Mechanical Turk, it turned out that the two types of questions performed very closely on all the three measures.