Computing and Information Systems - Theses

Permanent URI for this collection

Search Results

Now showing 1 - 10 of 41
  • Item
    Thumbnail Image
    Improvised coordination in agent organisations
    Keogh, Kathleen Nora ( 2018)
    This thesis investigates coordination between intelligent software agents operating within agent organisations. Motivated by the prospect of agents working with humans in real world complex domains, the thesis focuses on flexible behaviour and improvisation in agent organisations. Methods used to design organisations of software agents are explored with particular consideration given to problem situations that cannot be defined with a detailed pre-scripted solution for coordinated action. A conceptual model that describes the components that are needed in an agent based model in a multi-agent system is referred to in this thesis as a meta-model. A number of agent organisation-based meta-models and frameworks for coordination of agents have been proposed such as OperA, OMACS and SharedPlans. There is however, no specific meta-model or approach that addresses agent improvisation and unscripted coordination. The reality of complex coordination in people's behaviour is analysed and used to develop requirements for agents' behaviour. A meta-model is proposed to include components to address these requirements. A process outlining how to design and implement such organisations is presented. The meta-model draws on features in existing models in the literature and describes components to guide agents to behave with flexibility at run time. The thesis argues that coordinated agents benefit from an explicit representation of an organisational model and policies to guide agents' run time behaviour. Policies are proposed to maintain consistent knowledge and mutual plans between team members. Coordination is explicit and some flexibility is given to agents to improvise beyond the solution anticipated at design-time. Agents can mutually adjust individual plans to fit in with others so the multi-agent organisation is able to dynamically adapt to a changing environment. The meta-model and design approach is successfully demonstrated and validated using an implementation of a simulation system. In this demonstration system, agents in multiple organisations collaborate and coordinate to resolve a problem within an artificial simulation world.
  • Item
    Thumbnail Image
    Dauphin A Programming Language for Statistical Signal Processing - from principles to practice
    Kyprianou, Ross ( 2018)
    This dissertation describes the design and implementation of a new programming language called Dauphin for the signal processing domain. Dauphin's focus is on the primitive concepts and algorithmic structures of signal processing. In this language, random variables and probability distributions are as fundamental and easy to use as the numeric types of other languages. The basic algorithms of signal processing --- estimation, detection, classification and so on --- become the standard function calls. Too much time is expended by researchers in re-writing these basic algorithms for each application. Dauphin allows you to code these algorithms directly, so they can be coded once and put into libraries for future use. Ultimately, Dauphin aims to extend the power of the researcher by allowing them to focus on the real problems and simplify the process of implementing their ideas. The first half of this dissertation describes Dauphin and the design issues of existing languages used for signal processing that motivated its development. It includes a general investigation into programming language design and the identification of specific design criteria that impact signal processing programming. These criteria directed the features in Dauphin that support writing signal processing algorithms. Of equal importance, the criteria also provide a means to compare, with some objectivity, the suitability of different languages for signal processing. Following the discussion on language design, Dauphin's features are described in detail, then details related to Dauphin's implementation are presented, including a description of Dauphin's semantics and type system. The second half of the dissertation presents practical applications of the Dauphin language, focussing on three broad areas associated with signal processing: classification, estimation and Monte Carlo methods. These non-trivial applications, combined with examples throughout the dissertation, demonstrate that Dauphin is simple and natural to use, easy to learn and has sufficient expressiveness for general programming in the signal processing domain.
  • Item
    Thumbnail Image
    Understanding information security strategy in organisations
    Horne, Craig Andrew ( 2018)
    The research topic under investigation in this thesis is information security strategy in organisations and I propose a substantive theory for understanding this phenomenon under varying environmental and internal conditions. My original contribution to knowledge includes a definition for information security strategy, criteria for organisational environment and information assessment, a conceptual model of information security strategy, a substantive theory on information security strategy, and a descriptive set of benefits that can be adopted after strategy selection and approval. Organisations are progressively undertaking digital transformation of their products and services to reduce costs, improve customer relationships, and consolidate operations. Information is the “lifeblood” of any organisation and is increasingly being used to support this digital transformation across the entire organisation. Yet, the boundaries of information, its value, and importance in supporting organisational goals are frequently overlooked, creating security exposures and vulnerabilities. One reason for this is a lack of attention paid to cataloguing and controlling valuable information being used as a business resource. Others are that usage of emerging disruptive technology such as cloud-based applications can create porous network borders, that security controls used to protect information can be expensive and complex, and that organisational leaders may resist the implementation of security controls due to a perception that they impede productivity. This then leads to increased risk to information, affecting organisational leaders in the governing body, who currently have no consistent guidance available to help them in selecting a strategy or setting a strategic direction for information security. To address this problem, I examine a range of concepts when adopting an approach to securing information, by interviewing security leaders in larger organisations. In a qualitative study, I interviewed twenty-five participants and took a phenomenological approach to understanding their lived experiences with developing and using an information security strategy. I used grounded theory methodology and techniques to analyse the interview transcripts and their organisation’s information security strategy documents when permitted, to understand significant information security concepts and their relationships in an organisational context. The results show that organisational leaders choose from four main strategies when making decisions to secure their organisation’s information, which are Fortification, Devaluation, Outsourcing and Minimisation. Their selection depends on consideration of organisational factors including constraints on outsourcing decisions and the value of information held within the organisation. This facilitated the development of a conceptual model of information security strategy and a substantive theory on information security strategy. The implications of this are that organisations can continue business operations towards the achievement of strategic goals using information as a resource, and that the selection of an information security strategy can lead to a more complete understanding of the comprehensive strategic plans required to implement operational security controls throughout an organisation, making them more applicable and cost effective.
  • Item
    Thumbnail Image
    The use of clinical decision support systems for the development of medical students’ diagnostic reasoning skills
    Khumrin, Piyapong ( 2018)
    Computer-aided learning systems (e-learning systems) can help medical students gain more experience with diagnostic reasoning and decision-making. Within this context, providing feedback that matches student needs (i.e. personalised feedback) is both critical and challenging. Prior research showed that using Clinical Decision Support System (CDSS) to assist doctors improves the effectiveness and efficiency of diagnostic and treatment processes. However, the application of CDSS to the developmental process of clinical reasoning in a clinical teaching environment is still limited. In this research, we developed a new diagnostic decision support system embedded in a learning tool, called DrKnow. Students interact with twenty virtual patients to investigate though the learning steps similar to bedside teaching to arrive at a proper final diagnosis. DrKnow with CDSS-based design monitors students’ activities and provide personalised feedback to support students’ diagnostic decisions. We developed expert knowledge within DrKnow based on the machine learning models trained on 208 realworld clinical cases presenting with abdominal pain, to predict 5 diagnoses (appendicitis, gastroenteritis, urinary tract infection, ectopic pregnancy, and pelvic inflammatory disease). We assessed which of these models are likely to be most effective by predictive accuracy and clinical appropriateness when the model prediction was transformed to feedback. These models were leveraged to generate different kinds of feedback provided during along the process of decision making (interim feedback) and at the end of scenario (final feedback) based on the specific information requested by students from the virtual patients and their active diagnostic hypotheses. Students used this tool to explore one or more common clinical presentations, assessing patient histories, selecting and evaluating appropriate investigations and integrating these findings to select the most appropriate diagnosis. Based on the clinical information they request and prioritise, DrKnow presents key decision points and suggest three provisional diagnoses as they work through the virtual cases. Once students make a final diagnosis, DrKnow presents students with information about their overall diagnostic performance as well as recommendations for diagnosing similar cases. The analysis of the decisions of students as compared to those of DrKnow shows that DrKnow provided appropriate feedback on supporting students to select appropriate differential diagnoses and effective assessment of students diagnostic performance. Although DrKnow still has some limitations, we argue that the implementation of CDSS-based learning support for the development of diagnostic reasoning skills represented by DrKnow provides an effective learning process enabling positive student learning outcomes, while simultaneously overcoming the resource challenges of expert clinician supported bedside teaching.
  • Item
    Thumbnail Image
    Safety critical multi-agent systems
    Liu, Ching Louis ( 2018)
    Artificial intelligence (AI) has come of age, especially with the rapid growth in the variety and quantity of data, together with automation in industrial and private applications, such as 3D printing and automated manufacturing, all of which require computing for control and interaction with humans at some point. Computer-controlled machines that operate in industrial or domestic settings must also operate to ensure the safety of any humans in the vicinity or who interact with the machines. Further, if we are to realise the dream of autonomous agents interacting freely with humans, then the issue of safety becomes even more critical. In this regard, the aim of this thesis is to propose methods that complement existing Agent-Oriented Software Engineering methods and provide a means of safety engineering at the agent design stage. Our proposed methods can also build accident knowledge in such a way that a safety analysis on one type of multi-agent system can be transferred to another provided that the other multi-agent system shares enough similarities with the first. The current situation is that it is difficult to apply agent-oriented methods in situations where safety is critical because of the lack of available agent-specific safety analysis methods. Traditional safety engineering methods are not tailored to the analysis of the full capability of agents and although a number of attempts has been made to automate traditional safety engineering methods, a gap between the dynamic behaviour of multi-agent systems and safety analysis remains. Further, there is no single accepted definition of a multi-agent system, but there is a list of concepts common to most definitions and widely accepted among different methodologies: the concept of an open system, dynamic behaviour and adaption to name a few. Current safety analysis methods do not fully handle concepts with much success. Of the existing methods, which we review in chapter 3, all require domain knowledge as well as expertise in the application of the methods themselves and are limited by the size of the component that can be analysed. This thesis contributes to safety analysis in agent-oriented software engineering by providing safety analysis methods that generate tangible safety goals based on previous accident data and system behaviour. Another contribution of this thesis is that our method enables agents to dynamically calculate accident likelihood and then, through a specific systems level ontology, to translate the safety analysis from one multi-agent system to another with similar agent characteristics. An example of where this latter case can be applied is to provide estimations on the design of a new multi-agent system that does not yet have any accident data. We first look at ways of modelling system behaviour and, importantly, the interactions between different agents. Then, we present a way to convert the interaction model to a Bayesian network that combines data from multiple previous accidents and a method for identifying which system component to change to improve the safety of the overall multi-agent system. When we apply this method to real-life situations, we find that the current limitation is the lack of data at the right level of detail. However, exploring the interactions in the system and the relationships between agents, we can overcome the limitations in data to some extent. Our approach can be used to estimate the accident rate by combining accident data from different existing physical systems. Doing this provides a quick way to estimate the accident rate and provide design feedback to the multi-agent system designer. Our thesis will advance the application of multi-agent systems by improving their safety aspects. Moreover, the ability provided by our Bayesian networks to dynamically calculate the likelihood of accidents provides agents with the means to improve safety as they encounter new incidents. Our method of translating the analysis from one type of multi-agent system to another on the basis of ontology provides an interesting approach for sharing accident knowledge between related systems when they are in the field.
  • Item
    Thumbnail Image
    Compositional morphology through deep learning
    Vylomova, Ekaterina ( 2018)
    Most human languages have sophisticated morphological systems. In order to build successful models of language processing, we need to focus on morphology, the internal structure of words. In this thesis, we study two morphological processes: inflection (word change rules, e.g. run -- runs) and derivation (word formation rules, e.g. run -- runner). We first evaluate the ability of contemporary models that are trained using the distributional hypothesis, which states that a word's meaning can be expressed by the context in which it appears, to capture these types of morphology. Our study reveals that inflections are predicted at high accuracy whereas derivations are more challenging due to irregularity of meaning change. We then demonstrate that supplying the model with character-level information improves predictions and makes usage of language resources more efficient, especially in morphologically rich languages. We then address the question of to what extent and which information about word properties (such as gender, case, number) can be predicted entirely from a word's sentential content. To this end, we introduce a novel task of contextual inflection prediction. Our experiments on prediction of morphological features and a corresponding word form from sentential context show that the task is challenging, and as morphological complexity increases, performance significantly drops. We found that some morphological categories (e.g., verbal tense) are inherent and typically cannot be predicted from context while others (e.g., adjective number and gender) are contextual and inferred from agreement. Compared to morphological inflection tasks, where morphological features are explicitly provided, and the system has to predict only the form, accuracy on this task is much lower. Finally, we turn to word formation, derivation. Experiments with derivations show that they are less regular and systematic. We study how much a sentential context is indicative of a meaning change type. Our results suggest that even though inflections are more productive and regular than derivations, the latter also present cases of high regularity of meaning and form change, but often require extra information such as etymology, word frequency, and more fine-grained annotation in order to be predicted at high accuracy.
  • Item
    Thumbnail Image
    Machine learning methods for biomedical relation extraction
    Panyam Chandrasekarasastry, Nagesh ( 2018)
    Automated text mining has emerged as an important method for effective comprehension of the fast growing body of biomedical publications. Within this topic, relation extraction refers to the goal of automated extraction of relations between well known entities, from unstructured text. Biomedical relation extraction tasks such as the Protein-Protein Interaction and Chemical-induced-Disease (CID) relation extraction are motivated by critical applications in toxicology studies and drug discovery. In this thesis, our research focus is to develop machine learning techniques to improve Biomedical relation extraction. In this context, we address the following research questions: *How to produce a high performance biomedical relation extraction system without task specific optimizations? *How to incorporate similarity measures developed for biomedical relation extraction with syntactic parse structures, into state of the art machine learning algorithms? *How to produce a high performance supervised model for biomedical relation extraction system with minimal annotation costs? We demonstrate state of the art performance by using syntactic-parse-kernels and SVM-classifiers, without the use of task specific feature design or custom heuristics. In our work, we use tree kernels to boost the performance of chemical-disease relation task from 57.9% to 61.7%. In contrast, prior work [139] relied on task specific rules to boost their performance from 56.0% to 61.3%. To leverage the classical Approximate Subgraph Matching (ASM) similarity measure defined for dependency graphs, we derived a Mercer’s kernel from the classical ASM for use with SVMs—a powerful machine learning classifier. We show that the ASM graph kernel is effective in exploiting dependency graphs to achieve high performance biomedical relation extraction.Over two biomedical relation extraction tasks, namely Seedev-binary and chemical-disease relation extraction from sentences, the classical ASM distance measure achieves an F-score of 14.5% and 49.0% respectively. In comparison, ASM kernel achieves a substantially better performance of 26.1% and 58.1% respectively for these tasks. ASM kernel also outperforms tree kernels for the Seedev-binary and chemical-disease relation extraction tasks. We show that the ASM kernel attains a comparable performance to the state of the art all-path-graph kernel on several datasets such as the Chemical-Disease relation extraction from sentences and protein-protein interaction extraction from the BioInfer dataset. To produce a high performance supervised model for relation extraction with minimal annotation costs, we propose Bayesian hierarchical active sampling, a novel clustering based method for multi-instance active learning. In coarsely labeled datasets such as with document level annotations used in biomedical relation extraction—only bags, that is, a set of instances are annotated and instance level annotations are not provided. In this setting, the goal of multi-instance active learning is to supplement the coarsely labeled dataset with a minimal set of instance labels to boost the overall performance of the supervised model. We demonstrate that hierarchical active sampling is effective for multi-instance active learning. For example, on the Protein-Protein Interaction extraction task, with 50 instance labelings, our method achieves an Area Under Curve (AUC) measure of 76% as compared to 73% achieved by the baseline methods, namely multi-instance uncertainty and expected-gradient length.
  • Item
    Thumbnail Image
    Accurate and efficient human activity recognition
    Cheng, Weihao ( 2018)
    Human Activity Recognition (HAR) is a promising technology which enables artificial intelligence systems to identify user's physical activities such as walking, running, and cycling. Recently, the demand for HAR is continuously increasing in pace with the rapid development of ubiquitous computing techniques. Major applications of HAR including fitness tracking, safety monitoring, and contextual recommendation have been widely applied in people's daily lives. For example, a music App on smartphones can use HAR to detect the current activity of the user and recommend activity-related songs. State-of-the-art HAR methods are based on the machine learning technique, where a classification model is trained on a dataset to infer a number of predefined activities. The data for HAR is usually in the form of time series, which can be collected by sensors such as accelerometers, microphones, and cameras. In this thesis, we mainly focus on HAR using the data from inertial sensors, such as accelerations from accelerometers. A large number of existing studies on HAR aim to obtain high recognition accuracy. However, efficiency is also an important aspect of HAR. In this thesis, we attempt to improve HAR methods for both accuracy and efficiency. Toward this goal, we first devise accurate HAR methods, and then improve the efficiency of HAR while maintaining the accuracy. More specifically, we tackle three problems. The first problem is to accurately recognize the current activity during activity transitions. Existing HAR methods train classification models based on tailored time series containing single activity. However, in practical scenarios, a piece of time series data could capture multiple interleaving activities causing activity transitions. Thus, recognition of the current activity, i.e., the most recent one, is a critical problem to investigate. The second problem is to accurately predict complex activities from ongoing observations. Many time-critical applications, such as safety monitoring, require early recognition of complex activities which are performed over a long period of time. However, without being fully observed, complex activities are hard to be recognized due to their complicated patterns. Therefore, predicting complex activities from ongoing observations is an important task to study. The third problem is to improve energy-efficiency of HAR on mobile devices while maintaining high accuracy. Many applications of HAR are based on mobile devices. However, due to the limited battery capacity, real-time HAR requires minimization of energy cost to extend the operating spans of the devices. Generally, the cost can be cut down by reducing algorithmic computations and sensing frequencies. Yet it is worth to find a maximal cost reduction while preserving a high recognition accuracy. In this thesis, we present a set of algorithms to address the proposed problems. The key contributions of the thesis can be summarized as follows: 1. We propose a method to accurately recognize the current activity in the presence of multiple activities with transitions. The method partitions a time series matching the occurring activities, where the maximum classification error of the activities is minimized. 2. We propose a method to accurately predict complex activities over time from ongoing multivariate time series. The method utilizes an action sequence model and a complex activity model, which make predictions alternately based on each other as the observed data increases. 3. We propose a method to minimize the computational cost of HAR while maintaining high recognition accuracy. The method uses a Markov Decision Process (MDP) to select an optimal subset of feature representations for ensemble classification that minimizes redundant computations. 4. We propose a method to minimize a combined measurement of sensing cost and classification error of HAR. The method uses MDP to select appropriate sensing rate to sample the incoming data points, where the sparsity of the outcome time series is ensured to preserve the recognition accuracy.
  • Item
    Thumbnail Image
    The psychological contract and its impact on interpersonal relationship quality in IT outsourcing
    Jeong, Jongkil ( 2018)
    Relationship Quality (ReQ) levels established between project members is now considered to be a key indicator of IT Outsourcing (ITO) success. To enhance the current understanding of how ReQ is established between individual project members engaged in ITO, a construct known as the Psychological Contract (PsyCon) has been investigated in recent studies. This is because the PsyCon enables investigation of ITO relationships from mutual, individual, and psychological perspectives which offers a fresh approach towards better understanding of ITO relationships. However, despite empirical evidence that suggests that the PsyCon plays a critical role in how individuals perceive ITO relationships, studies to date have been few and far between, often with conflicting results. Hence, the main objective of this thesis is to improve our understanding on how the fulfilment (breach) of certain Psychological Contract (PsyCon) obligations impact the level of Relationship Quality (ReQ) shared between two partners engaged in an IT Outsourcing (ITO) agreement. To address the main research objective, I first developed and proposed a research model based on a synthesis of the existing literature, along with three theoretical propositions that represented the causal links between the PsyCon and ReQ in ITO. These propositions were tested with a deductive, qualitative multiple-case study consisting of 12 dyadic relationships spread across 15 individual participants; 7 organisations and 4 different ITO agreements. Results from this study indicate that the fulfilment (and/or breach) of the PsyCon obligations identified in this study did in fact have a direct impact over the ReQ levels established within interpersonal ITO relationships. The findings also suggest that pre-existing levels of ReQ may amplify or dampen the effect that the fulfilment (and/or breach) of the PsyCon obligation has on future ReQ levels as well. These findings not only address the inconsistency over the definition, boundaries and application of the PsyCon and ReQ in the context of ITO, but also provide a new perspective on how practitioners and researchers alike can deal with the ever-increasing complexities surrounding the ITO agreement through the establishment and management of stronger interpersonal dyadic relationships.
  • Item
    Thumbnail Image
    Understanding role of provenance in bioinformatics workflows and enabling interoperable computational analysis sharing
    Khan, Farah Zaib ( 2018)
    The automation of computational analyses in data-intensive domains such as genomics through scientific workflows is a widely adopted practice in many fields of research nowadays. Computationally driven data-intensive experiments using workflows enable Automation, Scaling, Adaption and Provenance support (ASAP). Provenance data collection is an essential factor for any computational workflow-centric research to achieve reproducibility, transparency and support trust in the published results. At present capture of provenance information across the plethora of workflow management systems and custom software platforms in the bioinformatics domain is not well supported and as such, there exist numerous challenges associated with the effective sharing, publication, understandability, reproducibility and repeatability of scientific workflows. This thesis focuses on providing a unified, interoperable and systematised view of provenance with specific focus on workflow environments in the bioinformatics domain. We identify and overcome the current disconnect between various workflows systems and their existing provenance representations. Through empirical analysis of complex genomic data analysis workflows using three exemplar workflow systems, we identify implicit assumptions that arise. These assumptions produce an incomplete view of provenance resulting in insufficient details that impact on workflow enactment requirements and ultimately on the reproducibility of the given analysis. We propose a set of recommendations to mitigate against such assumptions and enable workflow systems to document and capture complete provenance information that can subsequently be used for re-enacting workflows in other contexts and potentially using other workflow platforms. Based on this empirical case study and pragmatic analysis of related literature, we define a hierarchical provenance framework offering `Levels of Provenance and Resource Sharing''. Each level of this framework addresses specific provenance recommendations and supports the capture of rich provenance information, with the topmost layer enabling the sharing of comprehensive and executable workflows utilising retrospective provenance. To realise this framework, we leverage community-driven, domain-neutral, platform-independent and open-source standards to implement ``CWLProv'' - a format for the methodical representation of provenance supporting workflow enactment aggregating resources specific to the given enactment and associated workflow configuration settings. We realise CWLProv through the Common Workflow Language (CWL) for workflow definition and utilise Research Objects (ROs) for resource aggregation and PROV-Data Model (PROV-DM) to support the capture of retrospective provenance information as required for subsequent workflow enactments. To demonstrate the applicability of CWLProv, we extend an existing workflow executor (cwltool) to provide a reference implementation that generates metadata and provenance-rich interoperable workflow-centric ROs. This approach aggregates and preserves data and methods needed to support the coherent sharing of computational analyses and experiments. Evaluation of CWLProv using real-life bioinformatics pipelines is demonstrated to highlight the utility of the approach demonstrating the interoperability of workflow analyses and the benefits to research reproducibility more generally.