Jim Warren’s research while affiliated with University of Auckland and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (10)


Figure 1: An example network produced by NEAT -Blue: Inputs; Green: Outputs; White: Hidden nodes; Red: Unreachable nodes. This graph exemplifies the major challenges of the naive approach to backpropagation (see 3.3). First, there are unreachable nodes X and Y which either do not connect to the inputs or do not connect to the output through their directed graph. X cannot be reached from the inputs and Y does not connect to the output. Second, there are skip-layer connections such as from 2 to the output. This has two paths 2-3-O and 2-O resulting in a skip-layer effect. If implemented naively this requires 13 operations (one for each connection).
Figure 2: This shows the resulting solution produced by PropNEAT. The graph traversals identify the unreachable nodes and these are removed. The nodes of the same depth from the input are grouped into layers, in this case [1,2] and [3,4,5] as depth 1 and 2 respectively. Where there are skip layers (e.g., 2-O), the outputs of the shallower layer are concatenated to the outputs of subsequent layer and otherwise treated as normal. The subsequent weights layer is then applied across all of these inputs. This provides a consistent layer-based structure that can be mapped to the tensor algebra operations. After the graph-traversal operations, and excluding concatenation as trivial, this requires 3 tensor operations (one for each layer connection).
Figure 3: Scatter, correlation and histogram plots for PropNEAT Model complexity over all iterations for all datasets. "True" models are the highest-performing on validation data, used for final analysis. "False" models are other candidates with lower validation performance. Significance is shown with * indicating p<0.05, ** indicating p<0.01, *** indicating p<0.001
PropNEAT -- Efficient GPU-Compatible Backpropagation over NeuroEvolutionary Augmenting Topology Networks
  • Preprint
  • File available

November 2024

·

21 Reads

Michael Merry

·

Patricia Riddle

·

Jim Warren

We introduce PropNEAT, a fast backpropagation implementation of NEAT that uses a bidirectional mapping of the genome graph to a layer-based architecture that preserves the NEAT genomes whilst enabling efficient GPU backpropagation. We test PropNEAT on 58 binary classification datasets from the Penn Machine Learning Benchmarks database, comparing the performance against logistic regression, dense neural networks and random forests, as well as a densely retrained variant of the final PropNEAT model. PropNEAT had the second best overall performance, behind Random Forest, though the difference between the models was not statistically significant apart from between Random Forest in comparison with logistic regression and the PropNEAT retrain models. PropNEAT was substantially faster than a naive backpropagation method, and both were substantially faster and had better performance than the original NEAT implementation. We demonstrate that the per-epoch training time for PropNEAT scales linearly with network depth, and is efficient on GPU implementations for backpropagation. This implementation could be extended to support reinforcement learning or convolutional networks, and is able to find sparser and smaller networks with potential for applications in low-power contexts.

Download

Multivariate Sequential Analytics for Cardiovascular Disease Event Prediction

December 2022

·

46 Reads

·

3 Citations

Methods of Information in Medicine

Background Automated clinical decision support for risk assessment is a powerful tool in combating cardiovascular disease (CVD), enabling targeted early intervention that could avoid issues of overtreatment or undertreatment. However, current CVD risk prediction models use observations at baseline without explicitly representing patient history as a time series. Objective The aim of this study is to examine whether by explicitly modelling the temporal dimension of patient history event prediction may be improved. Methods This study investigates methods for multivariate sequential modelling with a particular emphasis on long short-term memory (LSTM) recurrent neural networks. Data from a CVD decision support tool is linked to routinely collected national datasets including pharmaceutical dispensing, hospitalization, laboratory test results, and deaths. The study uses a 2-year observation and a 5-year prediction window. Selected methods are applied to the linked dataset. The experiments performed focus on CVD event prediction. CVD death or hospitalization in a 5-year interval was predicted for patients with history of lipid-lowering therapy. Results The results of the experiments showed temporal models are valuable for CVD event prediction over a 5-year interval. This is especially the case for LSTM, which produced the best predictive performance among all models compared achieving AUROC of 0.801 and average precision of 0.425. The non-temporal model comparator ridge classifier (RC) trained using all quarterly data or by aggregating quarterly data (averaging time-varying features) was highly competitive achieving AUROC of 0.799 and average precision of 0.420 and AUROC of 0.800 and average precision of 0.421, respectively. Conclusion This study provides evidence that the use of deep temporal models particularly LSTM in clinical decision support for chronic disease would be advantageous with LSTM significantly improving on commonly used regression models such as logistic regression and Cox proportional hazards on the task of CVD event prediction.


Human Versus Machine: How Do We Know Who Is Winning? ROC Analysis for Comparing Human and Machine Performance under Varying Cost-Prevalence Assumptions

December 2021

·

29 Reads

·

1 Citation

Methods of Information in Medicine

Background Receiver operating characteristic (ROC) analysis is commonly used for comparing models and humans; however, the exact analytical techniques vary and some are flawed. Objectives The aim of the study is to identify common flaws in ROC analysis for human versus model performance, and address them. Methods We review current use and identify common errors. We also review the ROC analysis literature for more appropriate techniques. Results We identify concerns in three techniques: (1) using mean human sensitivity and specificity; (2) assuming humans can be approximated by ROCs; and (3) matching sensitivity and specificity. We identify a technique from Provost et al using dominance tables and cost-prevalence gradients that can be adapted to address these concerns. Conclusion Dominance tables and cost-prevalence gradients provide far greater detail when comparing performances of models and humans, and address common failings in other approaches. This should be the standard method for such analyses moving forward.


Venn diagrams representing differences in overlapping mental models. A Three mental models overlap with a subset being globally understood. B Although each pair intersects, there is no globally shared mental model. Inspired by and adapted from [34]
Example of charts referencing cardiovascular disease risk prediction models for primary-care use in New Zealand from Fig. 1 of [52]
Card sorting algorithm in progress from Fig. 1 of [41]
A mental models approach for defining explainable artificial intelligence

December 2021

·

456 Reads

·

27 Citations

BMC Medical Informatics and Decision Making

Background Wide-ranging concerns exist regarding the use of black-box modelling methods in sensitive contexts such as healthcare. Despite performance gains and hype, uptake of artificial intelligence (AI) is hindered by these concerns. Explainable AI is thought to help alleviate these concerns. However, existing definitions for explainable are not forming a solid foundation for this work. Methods We critique recent reviews on the literature regarding: the agency of an AI within a team; mental models, especially as they apply to healthcare, and the practical aspects of their elicitation; and existing and current definitions of explainability, especially from the perspective of AI researchers. On the basis of this literature, we create a new definition of explainable, and supporting terms, providing definitions that can be objectively evaluated. Finally, we apply the new definition of explainable to three existing models, demonstrating how it can apply to previous research, and providing guidance for future research on the basis of this definition. Results Existing definitions of explanation are premised on global applicability and don’t address the question ‘understandable by whom?’. Eliciting mental models can be likened to creating explainable AI if one considers the AI as a member of a team. On this basis, we define explainability in terms of the context of the model, comprising the purpose, audience, and language of the model and explanation. As examples, this definition is applied to regression models, neural nets, and human mental models in operating-room teams. Conclusions Existing definitions of explanation have limitations for ensuring that the concerns for practical applications are resolved. Defining explainability in terms of the context of their application forces evaluations to be aligned with the practical goals of the model. Further, it will allow researchers to explicitly distinguish between explanations for technical and lay audiences, allowing different evaluations to be applied to each.


Figure 1. Diagram of anticipated workflow. MCR (medical case reviewer); CM (case manager); Veterans' Affairs (New Zealand Veteran Affairs).
Figure 2. VeCHAT main navigation screen.
Figure 3. VeCHAT report dashboard example.
Positive screens in the 34 respondents
Acceptability of VeCHAT to veteran participants
VeCHAT: a proof-of-concept study on screening and managing veterans

March 2021

·

119 Reads

·

4 Citations

Journal of Primary Health Care

INTRODUCTIONNew Zealand veterans may have complex mental and physical complaints related to multiple exposures to war environments. They are entitled to, but often do not, access a range of physical, mental health and social services funded through Veterans’ Affairs New Zealand. eCHAT (electronic Case-finding and Help Assessment Tool) is a self-completed electronic holistic screen for substance misuse, problem gambling, anger control, physical inactivity, depression, anxiety, exposure to abuse; and assesses whether help is wanted for identified issues. AIMA proof-of-concept study was conducted to develop a modified version of eCHAT (VeCHAT) with remote functionality for clinical assessment of mental health and lifestyle issues of contemporary veterans, and assesses acceptability by veterans and Veterans’ Affairs staff, and feasibility of implementation. METHODS We used a co-design approach to develop VeCHAT. Veterans’ Affairs and service organisations invited veterans to remotely complete VeCHAT and a subsequent short online acceptability survey. Veterans’ Affairs medical and case manager staff underwent semi-structured interviews on feasibility and acceptability of VeCHAT use. RESULTSThirty-four veterans completed VeCHAT. The tool proved acceptable to veterans and Veterans’ Affairs staff. Key emergent themes related to tool functionality, design, ways and barriers to use, and suggested improvements. Veterans’ Affairs staff considered VeCHAT use to be feasible with much potential. DISCUSSIONCapacity of Veterans’ Affairs to respond if their engagement with veterans increases and employment of VeCHAT is scaled up, is unknown. Work is needed to assess how introducing VeCHAT as a standard procedure might influence Veterans’ Affairs case management processes.


Clinically Comprehensible Clustering of Change in Time Series: The C4TS Algorithm

January 2019

·

34 Reads

Multivariate time series provide many challenges for analysis, especially if there are additional requirements of being able to translate analytical results easily to the clinical context, and scale analyses to large numbers of records. The Clinically Comprehensible Clustering of Change in Time Series (C4TS) algorithm is a novel approach to reduce anaesthetic records from a multivariate physiological time series to a set of features, directly relatable to the clinical context, scalable to large data sets, and compatible with a wide range of statistical analyses. These features are windows of the records, located at points of change detected via a drift detector and clustered into cohesive groups. A proof of concept version of C4TS was implemented and evaluated on a retrospective set of 9,246 cardiac, thoracic and otorhinological electronic anaesthetic records from one hospital in New Zealand. C4TS captured distinct features that were interpretable by clinicians, and were amenable to statistical analysis relating them to 30-day mortality yielding plausible results that would be worth exploring with a more robust implementation of the algorithm.


Health Consumer Usage Patterns in Management of CVD using Data Mining Techniques

January 2019

·

27 Reads

·

1 Citation

The Healthcare system is exposed to the increasing impact of chronic diseases including cardiovascular diseases; it is of much importance to analyze and understand the health trajectories for efficient planning and fair allotment of resources. This work proposes an approach based on mining clinical data to support the exploration of health trajectories related to cardiovascular diseases. As the health data are highly confidential, we aimed to conduct our experiments using a large, synthetic, longitudinal dataset, constituted to represent the CVD risk factors distribution and temporal sequence of events related to heart failure hospitalization and readmission. This research work analyses and represents the temporal events or states of the patient's trajectory with the aim of understanding the patient's journey in the management of the chronic condition and its complications by using data mining techniques. This study focuses on developing an efficient algorithm to find cohesive clusters for handling the temporal events. Clustering health trajectories have been carried out by proposing an improved version of the Ant-based clustering algorithm. Insights from this study can potentially result in evidence that these approaches are useful in understanding and analyzing patient's health trajectories for better management of the chronic condition and its progression.


Multivariate Sequential Analytics for Treatment Trajectory Forecasting

January 2019

·

39 Reads

Chronic conditions, especially cardiovascular disease account for a large burden on modern healthcare systems. These conditions are by their nature ones that unfold over a long period of time, typically involving many healthcare events, treatments and changes of patient status. The gold standard in public health informatics for risk assessment is regression-based. While these techniques are effective in identifying factors contributing to risk, they produce reductive scores (e.g. probability of a specific class of event, like a heart attack) or binary prediction results, and moreover, they are sequence agnostic. In the area of long-term chronic disease management, multivariate sequential modeling offers an opportunity to forecast disease progression and treatment trajectory in a fine-grained manner in order to aid clinical decision making. This paper investigates the suitability of Long short-term memory, a type of recurrent neural network, in conducting multivariate sequential modeling in the healthcare domain, specifically in the task of forecasting. The eventual goal is to apply this technique to linked New Zealand health data through the Vascular Informatics using Epidemiology and the Web (VIEW) research project. This paper presents initial experiments and results for modeling patients' treatment trajectories during hospitalization using the Medical Information Mart for Intensive Care (MIMIC-III) data set.


Towards a Youth Mental Health Screening Analytics Tool

January 2019

·

54 Reads

Risky health behaviours and mental health problems in adolescents constitute a global concern, and are relatively common in New Zealand. A self-administered screening tool, YouthCHAT, has been developed and validated for general practices and school clinics. As YouthCHAT is more widely deployed, an opportunity emerges to analyse aggregated amassed data to extend epidemiological knowledge of psychosocial issues and facilitate monitoring health of the youth population. The present research designs and prototypes a data analytics portal to facilitate such analysis. The design process is supported by analysis of a dataset generated by a YouthCHAT trial, a scanning study of existing tools that aggregate and analyse health survey data, stakeholder interviews, development of personas and scenarios of use, and creation of a simulated dataset on the scale of an expected future YouthCHAT database. The result is a prototype providing descriptive and inferential analyses and visualisations meeting expected requirements of future users.


A Population-Level Data Analytics Portal for Self-Administered Lifestyle and Mental Health Screening

January 2016

·

20 Reads

·

2 Citations

Studies in Health Technology and Informatics

This paper describes development of a prototype data analytics portal for analysis of accumulated screening results from eCHAT (electronic Case-finding and Help Assessment Tool). eCHAT allows individuals to conduct a self-administered lifestyle and mental health screening assessment, with usage to date chiefly in the context of primary care waiting rooms. The intention is for wide roll-out to primary care clinics, including secondary school based clinics, resulting in the accumulation of population-level data. Data from a field trial of eCHAT with sexual health questions tailored to youth were used to support design of a data analytics portal for population-level data. The design process included user personas and scenarios, screen prototyping and a simulator for generating large-scale data sets. The prototype demonstrates the promise of wide-scale self-administered screening data to support a range of users including practice managers, clinical directors and health policy analysts.

Citations (4)


... Hsu et al. [11] proposed a methodology for cardiovascular disease-related event detection using recurrent neural networks. The research work was carried out using a 2-year observation and 5-year prediction window. ...

Reference:

Fuzzy rule-based intelligent cardiovascular disease prediction using complex event processing
Multivariate Sequential Analytics for Cardiovascular Disease Event Prediction

Methods of Information in Medicine

... The majority of the published papers that focus on XAI definition either review existing literature [30,39,40] or seek definitions inspired by existing literature and their own expertise [16,17,32,33,[41][42][43][44][45][46][47]. Existing definitions of XAI often fail to provide specific guidance. ...

A mental models approach for defining explainable artificial intelligence

BMC Medical Informatics and Decision Making

... This paper makes a contribution by demonstrating the use of co-design to develop an approach to veteran well-being during the transition, with programme features and system characteristics, which participants have suggested would reduce excessive alcohol consumption. In the veteran domain, participatory research approaches are gaining traction [39]; however, co-design has been more commonly used to develop service screening methods [58] or information technology solutions [59,60], than programmes or service ecosystems. Furthermore, some co-design attempts have involved a majority of experts and only some veteran input [61] rather than adopting a deeper participatory approach that places the end-user in the driver's seat during the design of programmes intended for them. ...

VeCHAT: a proof-of-concept study on screening and managing veterans

Journal of Primary Health Care

... This can assist in appropriate provision of mental health and addiction services to align service provision with population need, and improve services through benchmarking. A data analytics portal has been designed and prototyped to support a range of users including practice managers, clinical directors and health policy analysts based on initial YouthCHAT field trial data [100]. ...

A Population-Level Data Analytics Portal for Self-Administered Lifestyle and Mental Health Screening
  • Citing Article
  • January 2016

Studies in Health Technology and Informatics