Computing and Information Systems - Theses

Permanent URI for this collection

Search Results

Now showing 1 - 7 of 7
  • Item
    Thumbnail Image
    Dauphin A Programming Language for Statistical Signal Processing - from principles to practice
    Kyprianou, Ross ( 2018)
    This dissertation describes the design and implementation of a new programming language called Dauphin for the signal processing domain. Dauphin's focus is on the primitive concepts and algorithmic structures of signal processing. In this language, random variables and probability distributions are as fundamental and easy to use as the numeric types of other languages. The basic algorithms of signal processing --- estimation, detection, classification and so on --- become the standard function calls. Too much time is expended by researchers in re-writing these basic algorithms for each application. Dauphin allows you to code these algorithms directly, so they can be coded once and put into libraries for future use. Ultimately, Dauphin aims to extend the power of the researcher by allowing them to focus on the real problems and simplify the process of implementing their ideas. The first half of this dissertation describes Dauphin and the design issues of existing languages used for signal processing that motivated its development. It includes a general investigation into programming language design and the identification of specific design criteria that impact signal processing programming. These criteria directed the features in Dauphin that support writing signal processing algorithms. Of equal importance, the criteria also provide a means to compare, with some objectivity, the suitability of different languages for signal processing. Following the discussion on language design, Dauphin's features are described in detail, then details related to Dauphin's implementation are presented, including a description of Dauphin's semantics and type system. The second half of the dissertation presents practical applications of the Dauphin language, focussing on three broad areas associated with signal processing: classification, estimation and Monte Carlo methods. These non-trivial applications, combined with examples throughout the dissertation, demonstrate that Dauphin is simple and natural to use, easy to learn and has sufficient expressiveness for general programming in the signal processing domain.
  • Item
    Thumbnail Image
    The use of clinical decision support systems for the development of medical students’ diagnostic reasoning skills
    Khumrin, Piyapong ( 2018)
    Computer-aided learning systems (e-learning systems) can help medical students gain more experience with diagnostic reasoning and decision-making. Within this context, providing feedback that matches student needs (i.e. personalised feedback) is both critical and challenging. Prior research showed that using Clinical Decision Support System (CDSS) to assist doctors improves the effectiveness and efficiency of diagnostic and treatment processes. However, the application of CDSS to the developmental process of clinical reasoning in a clinical teaching environment is still limited. In this research, we developed a new diagnostic decision support system embedded in a learning tool, called DrKnow. Students interact with twenty virtual patients to investigate though the learning steps similar to bedside teaching to arrive at a proper final diagnosis. DrKnow with CDSS-based design monitors students’ activities and provide personalised feedback to support students’ diagnostic decisions. We developed expert knowledge within DrKnow based on the machine learning models trained on 208 realworld clinical cases presenting with abdominal pain, to predict 5 diagnoses (appendicitis, gastroenteritis, urinary tract infection, ectopic pregnancy, and pelvic inflammatory disease). We assessed which of these models are likely to be most effective by predictive accuracy and clinical appropriateness when the model prediction was transformed to feedback. These models were leveraged to generate different kinds of feedback provided during along the process of decision making (interim feedback) and at the end of scenario (final feedback) based on the specific information requested by students from the virtual patients and their active diagnostic hypotheses. Students used this tool to explore one or more common clinical presentations, assessing patient histories, selecting and evaluating appropriate investigations and integrating these findings to select the most appropriate diagnosis. Based on the clinical information they request and prioritise, DrKnow presents key decision points and suggest three provisional diagnoses as they work through the virtual cases. Once students make a final diagnosis, DrKnow presents students with information about their overall diagnostic performance as well as recommendations for diagnosing similar cases. The analysis of the decisions of students as compared to those of DrKnow shows that DrKnow provided appropriate feedback on supporting students to select appropriate differential diagnoses and effective assessment of students diagnostic performance. Although DrKnow still has some limitations, we argue that the implementation of CDSS-based learning support for the development of diagnostic reasoning skills represented by DrKnow provides an effective learning process enabling positive student learning outcomes, while simultaneously overcoming the resource challenges of expert clinician supported bedside teaching.
  • Item
    Thumbnail Image
    Machine learning with adversarial perturbations and noisy labels
    Ma, Xingjun ( 2018)
    Machine learning models such as traditional random forests (RFs) and modern deep neural networks (DNNs) have been successfully used to solve complex learning problems in many applications such as speech recognition, image classification, face recognition, gaming agents and self-driving cars. For example, DNNs have demonstrated near or even surpassing human-level performance in image classification tasks. Despite their current success, these models are still vulnerable to noisy real-world situations where illegitimate or noisy data may exist to corrupt learning. Studies have shown that by adding small, human imperceptible (in the case of images) adversarial perturbations, normal samples can be perturbed into "adversarial examples'', and DNNs can be made to misclassify adversarial examples with a high level of confidence. This arouses security concerns when employing DNNs in security-sensitive applications such as fingerprint recognition, face verification and autonomous cars. Studies have also found that DNNs can overfit to noisy (incorrect) labels and as a result, generalize poorly. This has been one of the key challenges when applying DNNs in noisy real-world scenarios where even high-quality datasets tend to contain noisy labels. Another open question in machine learning is whether actionable knowledge (or "feedback") can be generated from prediction models to support decision making towards some long-term learning goals (for example, mastering a certain type of skills in a simulation-based learning (SBL) environment). We view the feedback generation problem from a new perspective of adversarial perturbation, and explore the possibility of using adversarial techniques to generate feedback. In this thesis, we investigate machine learning models including DNNs and RFs, and their learning behavior through the lens of adversarial perturbations and noisy labels, with the aim of achieving more secure and robust machine learning. We also explore the possibility of using adversarial techniques in a real-world application: to support skill acquisition in SBL environments through the provision of performance feedback. The first part of our work is on the investigation of DNNs and their vulnerability to adversarial perturbations, in the context of image classification. In contrast to existing work, we develop new understandings of adversarial perturbations by exploring DNN representation space with the Local Intrinsic Dimensionality (LID) measure. In particular, we characterize adversarial subspaces in the vicinity of adversarial examples using LID, and find that adversarial subspaces are of higher intrinsic dimensionality than normal data subspaces. We not only provide a theoretical explanation of the high dimensionality of adversarial subspaces, but also empirically demonstrate that such properties can be used to effectively discriminate adversarial examples generated using state-of-the-art attacking methods. The second part of our work is to explore the possibility of using adversarial techniques in a beneficial way to generate interactive feedback for intelligent tutoring in SBL environments. Feedback is actions (in the form of feature changes) generated from a pre-trained prediction model that can be delivered to a leaner in an SBL environment to correct mistakes or improve skills. We demonstrate that such feedback can be generated accurately and efficiently using properly constrained adversarial techniques with DNNs. In addition to DNNs, we also explore, in the third part of our work, adversarial feedback generation from RF models. Adversarial perturbations can be easily generated from DNNs using gradient descent and backpropagation, however, it is still an open question whether such perturbations can be generated from models such as RFs that do not work with gradients. This part of our work confirms that adversarial perturbations can also be crafted from RFs for the provision of feedback in SBL. In particular, we propose a perturbation method that can find the optimal space transition from one undesired class (e.g. 'novice') to the desired class (e.g. 'expert'), based on a geometric view of the RF decision space as overlapping high dimensional rectangles. We demonstrate empirically that our proposed method has high effectiveness as well as high efficiency when compared to existing methods, making it possible to be used for real-time feedback generation in SBL. The fourth part of our work focuses on DNNs and noisy label learning: training accurate DNNs on data with noisy labels. In this work, we investigate the learning behaviours of DNNs, and show that DNNs exhibit two distinct learning styles when trained on clean versus noisy labels. A LID-based characterization of the intrinsic dimensionality of DNN subspace (inspired by the first part of our work) allows us to identify the two stages of learning from dimensionality compression to dimensionality expansion on datasets with noisy labels. Based on the observation that dimensionality expansion is associated with overfitting to noisy labels, we further propose a heuristic learning strategy to avoid the later stage of dimensionality expansion, so as to robustly train DNNs in the presence of noisy labels. In summary, this work has contributed significantly to existing knowledge through: novel dimensional characterization of DNNs, effective discrimination of adversarial attacks, robust deep learning strategies against noisy labels, and novel approaches to feedback generation. All work is supported by theoretical analysis, empirical results and publications.
  • Item
    Thumbnail Image
    Towards highly accurate publication information extraction from academic homepages
    Zhang, Yiqing ( 2018)
    More and more researchers list their research profiles in academic homepages online. Publications from a researcher's academic homepage contain rich information, such as the researcher's fields of expertise, research interests, and collaboration network. Extracting publication information from academic homepages is an essential step in automatic profile analysis, which enables many applications such as academic search, bibliometrics and citation analysis. The publications extracted from academic homepages can also be a supplementary source for bibliographic databases. We investigate two publication extraction problems in this thesis: (i) Given an academic homepage, how can we precisely extract all the individual publication strings from the homepage? Here, a publication string is a text string that describes a publication record. We call this problem publication string extraction. (ii) Given a publication string, how can we extract different fields, such as publication authors, publication title, and publication venue, from the publication string? We call this problem publication field extraction. There are two types of traditional approaches to these two problems, rule-based approaches and machine learning based approaches. Rule-based approaches cannot accommodate the large variety of styles in the homepages, and they require significant efforts in rule designing. Machine learning based approaches rely on a large amount of high-quality training data as well as suitable model structures. To tackle these challenges, we first collect two datasets and annotate them manually. We propose a training data enhancement method to generate large sets of semi-real data for training our models. For the publication string extraction problem, we propose a PubSE model that can model the structure of a publication list in both line-level and webpage-level. For the publication field extraction problem, we propose an Adaptive Bi-LSTM-CRF model that can utilize the generated and the manually labeled training data to the full extent. Extensive experiment results show that the proposed methods outperform the state-of-the-art methods in the publication extraction problems studied.
  • Item
    Thumbnail Image
    Indoor localization supported by landmark graph and locomotion activity recognition
    Gu, Fuqiang ( 2018)
    Indoor localization is important for a variety of applications such as location-based services, mobile social networks, and emergency response. Although a number of indoor localization systems have been proposed in recent years, they have different limitations in terms of accuracy, cost, coverage, complexity, and applicability. In order to achieve a higher accuracy with relatively low cost, hybrid methods combining multiple positioning techniques have been used. However, hybrid methods usually require an infrastructure of beacons or transmitters, which may not be available in many environments or it may be available at a high cost. Spatial knowledge is available in many scenarios, and can be used to assist localization without additional cost. Landmarks are one of the spatial constraints useful for indoor localization. Indoor localization systems that use landmarks have been proposed in the literature, but they are usually applied for tracking robots by using laser scanners or/and cameras. The systems using these devices are economically or/and computationally expensive, and hence are not suitable for indoor pedestrian localization. Although landmarks based on the built-in smartphone sensors are also used in some indoor localization systems, the performance of these systems relies highly on the completeness of landmarks. A mismatch of landmarks may cause a large localization error and even lead to the failure of localization. The advent of sensor-equipped smart devices has enabled a variety of activity recognition and inference, including locomotion (e.g., walking, running, standing). The sensors built in the smart devices can capture the intensity and duration of activity, and even are able to sense the activity context. Such information can be used to enhance the localization accuracy or reduce the energy consumption and deployment cost while maintaining the accuracy. For example, the knowledge of locomotion activities can be used to optimize the step length estimation of people, which will contribute to the improvement of localization accuracy. However, it is challenging to precisely recognize activities related to indoor localization with smartphones. The hypothesis of this research is that accurate and reliable indoor localization can be achieved by fusing smartphone sensor data with locomotion activities and a landmark graph. This hypothesis is tested using the novel algorithms proposed and developed in this research. The proposed framework consists of four main phases, namely recognizing locomotion activities related to indoor localization from sensor data, improving the accuracy of step counting and step length estimation for pedestrian dead reckoning method, developing a landmark graph-based indoor localization method, and implementing quick WiFi fingerprint collection. The main contributions of this research are as follows. First, a novel method is proposed for locomotion activity recognition by automatically learning useful features from sensor data using a deep learning model. Second, robust and accurate algorithms are proposed for step counting and step length estimation to improve the performance of pedestrian dead reckoning, which will be fused with spatial information. Third, the concept of sensory landmarks and the landmark graph is proposed, and a landmark graph-based method is developed for indoor localization. Fourth, a practical, fast, and reliable fingerprint collection method is designed, which uses the landmark graph-based localization method for automatically estimating the location of reference points used to associate the collected fingerprints.
  • Item
    Thumbnail Image
    Analysing the interplay of location, language and links utilising geotagged Twitter content
    Afshin, Rahimi ( 2018)
    Language use and interactions on social media are geographically biased. In this work we utilise this bias in predictive models of user geolocation and lexical dialectology. User geolocation is an important component of applications such as personalised search and recommendation systems. We propose text-based and network-based geolocation models, and compare them over benchmark datasets yielding state-of-the- art performance. We also propose hybrid and joint text and network geolocation models that improve upon text or network only models and show that the joint models are able to achieve reasonable performance in minimal supervision scenarios, as often happens in real world datasets. Finally, we also propose the use of continuous representations of location, which enables regression modelling of geolocation and lexical dialectology. We show that our proposed data-driven lexical dialectology model provides qualitative insights in studying geographical lexical variation.
  • Item
    Thumbnail Image
    Scaling learning algorithms using locality sensitive hashing
    Aye, Zay Maung Maung ( 2018)
    With increasing volumes of data, it is imperative that data analysis can appropriately scale. However, many common machine learning algorithms, e.g., metric learning, manifold landmark learning, and processing trajectories, suffer poor computational complexity in the size of training data. In this thesis, we propose generic methods for scaling up learning algorithms by utilizing locality sensitive hashing. First, finding representative samples utilizing locality sensitive hashing is proposed. The usefulness of these samples is demonstrated on large-scale supervised metric learning. Our methods achieve quadratic speed up with only minimal decrease in accuracy. Second, representative samples are leveraged for adaptive minibatch selection for fitting Gaussian processes for landmarking manifolds. Our methods exploit the compatibility of locality sensitive hashing and the manifold assumption in high-dimensional data, thereby limiting expensive optimization to relevant regions of the data. Training the state-of-the-art learner with our compressed dataset achieves superior accuracy compared to training with randomly selected samples. We also demonstrate that our methods can be used to find manifold landmarks without learning Gaussian processes at all, which leads to orders-of-magnitude speed up with only minimal decrease in accuracy. And finally, we propose locality sensitive hashing based feature hashing methods which map variable length trajectories to constant length trajectories for efficient similarity computation in Euclidean space. Our methods can accelerate trajectory clustering while achieving competitive accuracy in comparison to clustering using more complicated distance function, such as Dynamic Time Warping.