Myroslava O. Dzikovska’s research while affiliated with University of Edinburgh and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (68)


Fig. 1 Summary of available data in the IDEAL household energy dataset, and IDEAL app features received, by type of participant. home.install_type and home.study_class refer to fields in the home table in the dataset, and indicate how to identify and select each group of participants by type of installation or study group. Those interested in NILM research should utilise the enhanced group of homes. For details of the different study groups, see the Research design section. For a summary of the data available, see Tables 3 and 4; for details, see the Data acquisition section. For a summary of the IDEAL app features received by each group, see the project website, http://www.energyoracle.org/energy-feedback.html.
Fig. 2 Heat map of propagation rates of homes' IDEAL sensor systems, by date. Plot shows hour-by-hour propagation rates from homes as a proportion of expected readings, based on data from IDEAL sensors. One horizontal line per home, ordered by participation start date. Grey indicates periods outside a home's participation in the study.
Fig. 4 Schematic diagram of the IDEAL system. Arrows indicate direction of data flow between participating homes, the University servers and the project participants.
Sociodemographic and building information about the homes in the IDEAL household energy dataset. Note varying y-axis scales. Figures are based on data collected during each household’s installation visit. Blue indicates homes with standard installations; green indicates homes with enhanced installations.
The IDEAL household energy dataset, electricity, gas, contextual sensor data and survey data for 255 UK homes
  • Article
  • Full-text available

May 2021

·

568 Reads

·

57 Citations

Scientific Data

·

Jonathan Kilgour

·

·

[...]

·

The IDEAL household energy dataset described here comprises electricity, gas and contextual data from 255 UK homes over a 23-month period ending in June 2018, with a mean participation duration of 286 days. Sensors gathered 1-second electricity data, pulse-level gas data, 12-second temperature, humidity and light data for each room, and 12-second temperature data from boiler pipes for central heating and hot water. 39 homes also included plug-level monitoring of selected electrical appliances, real-power measurement of mains electricity and key sub-circuits, and more detailed temperature monitoring of gas- and heat-using equipment, including radiators and taps. Survey data included occupant demographics, values, attitudes and self-reported energy awareness, household income, energy tariffs, and building, room and appliance characteristics. Linked secondary data comprises weather and level of urbanisation. The data is provided in comma-separated format with a custom-built API to facilitate usage, and has been cleaned and documented. The data has a wide range of applications, including investigating energy demand patterns and drivers, modelling building performance, and undertaking Non-Intrusive Load Monitoring research.

Download

Beetle-Grow: An Effective Intelligent Tutoring System for Data Collection

April 2016

·

59 Reads

·

2 Citations

We present the Beetle-Grow intelligent tutoring system, which combines active experimentation, self-explanation, and formative feedback using natural language interaction. It runs in a standard web browser and has a fresh, engaging design. The underlying back-end system has previously been shown to be highly effective in teaching basic electricity and electronics concepts. Beetle-Grow has been designed to capture student interaction and indicators of learning in a form suitable for data mining, and to support future work on building tools for interactive tutoring that improve after experiencing interaction with students, as human tutors do.


The joint student response analysis and recognizing textual entailment challenge: making sense of student responses in educational applications

August 2015

·

82 Reads

·

32 Citations

Language Resources and Evaluation

We present the results of the joint student response analysis (SRA) and 8th recognizing textual entailment challenge. The goal of this challenge was to bring together researchers from the educational natural language processing and computational semantics communities. The goal of the SRA task is to assess student responses to questions in the science domain, focusing on correctness and completeness of the response content. Nine teams took part in the challenge, submitting a total of 18 runs using methods and features adapted from previous research on automated short answer grading, recognizing textual entailment and semantic textual similarity. We provide an extended analysis of the results focusing on the impact of evaluation metrics, application scenarios and the methods and features used by the participants. We conclude that additional research is required to be able to leverage syntactic dependency features and external semantic resources for this task, possibly due to limited coverage of scientific domains in existing resources. However, each of three approaches to using features and models adjusted to application scenarios achieved better system performance, meriting further investigation by the research community.


The CADENCE Corpus

April 2015

·

43 Reads

·

4 Citations

Papers on voice interfaces for people with cognitive impairment or demenita only provide small snapshots of actual interactions, if at all. This is a major obstacle to the development of better interfaces. Transcripts of interactions between users and systems contain rich evidence of typical language patterns, indicate how users conceptualise their computer interlocutor, and highlight key design issues. In this paper, we introduce the CADENCE corpus and outline how it can be used to stimulate replicable research on inclusive voice interfaces. The CADENCE corpus is first data set of its kind to include rich data from people with cognitive impairment and free for research use. The corpus consists of transcribed spoken interactions between older people with and without cognitive impairment and a simulated Intelligent Cognitive Assistant and includes comprehensive data on users' cognitive abilities.


BEETLE II: Deep Natural Language Understanding and Automatic Feedback Generation for Intelligent Tutoring in Basic Electricity and Electronics

September 2014

·

903 Reads

·

133 Citations

International Journal of Artificial Intelligence in Education

Within STEM domains, physics is considered to be one of the most difficult topics to master, in part because many of the underlying principles are counter-intuitive. Effective teaching methods rely on engaging the student in active experimentation and encouraging deep reasoning, often through the use of self-explanation. Supporting such instructional approaches poses a challenge for developers of Intelligent Tutoring Systems. We describe a system that addresses this challenge by teaching conceptual knowledge about basic electronics and electricity through guided experimentation with a circuit simulator and reflective dialogue to encourage effective self-explanation. The Basic Electricity and Electronics Tutorial Learning Environment (BEETLE II) advances the state of the art in dynamic adaptive feedback generation and natural language processing (NLP) by extending symbolic NLP techniques to support unrestricted student natural language input in the context of a dynamically changing simulation environment in a moderately complex domain. This allows contextually-appropriate feedback to be generated “on the fly” without requiring curriculum designers to anticipate possible student answers and manually author multiple feedback messages. We present the results of a system evaluation. Our curriculum is highly effective, achieving effect sizes of 1.72 when comparing pre- to post-test learning gains from our system to those of a no-training control group. However, we are unable to demonstrate that dynamically generated feedback is superior to a non-NLP feedback condition. Evaluation of interpretation quality demonstrates its link with instructional effectiveness, and provides directions for future research and development.


System Comparisons: Is There Life after Null?

July 2013

·

16 Reads

Lecture Notes in Computer Science

It is common practice to compare gain scores in order to determine the effectiveness of adding features to a training system. Here we argue that relying on one measure of overall system effectiveness may result in overlooking valuable lessons available from a comparison of different versions of a system. To illustrate our point, we present the results of comparing a Natural Language Processing (NLP) based adaptive feedback system to a system that does not utilize NLP capabilities. We show that, while there were no learning gain differences between the two systems, the correlates to gain were different. In the non-NLP system, only student performance during the training was correlated to learning gain. In the adaptive system, more variables correlated with learning, including measures of system capability and student satisfaction. This level of analysis suggests that the two systems are not equivalent and points us towards modifications that may improve effectiveness.


Combining Semantic Interpretation and Statistical Classification for Improved Explanation Processing in a Tutorial Dialogue System

July 2013

·

19 Reads

·

9 Citations

Lecture Notes in Computer Science

We present an approach for combining symbolic interpretation and statistical classification in the natural language processing (NLP) component of a tutorial dialogue system. Symbolic NLP approaches support dynamic generation of context-adaptive natural language feedback, but lack robustness. In contrast, statistical classification approaches are robust to ill-formed input but provide less detail for context-specific feedback generation. We describe a system design that combines symbolic interpretation with statistical classification to support context-adaptive, dynamically generated natural language feedback, and show that the combined system significantly improves interpretation quality while retaining the adaptivity benefits of a symbolic interpreter.





Citations (59)


... Деякі останні експериментальні системи можуть також адаптувати свої стратегії до настрою користувача, як-от до розчарування чи нудьги, про що може свідчити поведінка, тон голосу і навіть вираз обличчя або жести, проаналізовані за допомогою комп'ютерного зору. Інші прототипні системи також прагнуть до об'ємнішого моделювання психіки користувача, яке базуватиметься, зокрема, на покращеному розумінні його мови (наприклад, Callaway et al., 2007). ...

Reference:

Прикладні аспекти комп'ютерної лінгвістики. Навчально-методичний посібник [Applications of computational linguistics. A handbook]
The Beetle and BeeDiff tutoring systems
  • Citing Conference Paper
  • October 2007

... The (publicly available) datasets considered in our system are UKDALE, REFIT, and IDEAL [11]- [13] (though, users could upload other datasets, as well). Each dataset comprises several houses monitored by sensors that record the total main and appliance-level power for a period of time (used only during evaluation). ...

The IDEAL household energy dataset, electricity, gas, contextual sensor data and survey data for 255 UK homes

Scientific Data

... Responses where the grading results deviate from human-labeled scores are selected for the inner iteration. In the inner iteration (lines [11][12][13][14][15][16][17][18][19], the Reflector identifies erroneous grading responses and suggests rubric improvements. These suggestions are then refined by the Refiner, generating an updated rubric for the next inner iteration. ...

Semeval-2013 task 7: The joint student response analysis and 8th recognizing textual entailment challenge
  • Citing Article
  • January 2013

... Our demonstration will showcase the Beetle-Grow intelligent tutoring system [3]. Student interaction data ( Figure 1) and other indicators of learning are logged in a format suitable for data mining and as training data for machine learning. ...

Beetle-Grow: An Effective Intelligent Tutoring System for Data Collection
  • Citing Conference Paper
  • April 2016

... All three, mean, standard deviation, and range, can be found for less than half (42.7%) of the datasets (e.g., "The mean age of the subjects was 54.9 ± 13.4 (SD) yr (range 36-70 yr)" [64]). Meanwhile, some documentation noted only the minimum (e.g., "participants aged 50 or older" [180]) or the age requirement for participation (e.g., "18 or older" [13]). ...

The CADENCE Corpus
  • Citing Conference Paper
  • April 2015

... In Beetle II, students' utterances are analyzed by means of a process that includes two stages: in the first, the TRIPS dialog analyzer [85] generates a semantic representation that is domain independent; and in the second, the contextual interpreter applies a reference resolution approach and a set of rules to obtain a representation in terms of the Beetle II domain. Later, in [86], the group of researchers responsible for Beetle II presented a study of how to improve the robustness of the semantic interpreter. Basically, the improvement consisted of implementing a classifier based on lexical similarity within the symbolic approach. ...

Combining Semantic Interpretation and Statistical Classification for Improved Explanation Processing in a Tutorial Dialogue System
  • Citing Conference Paper
  • July 2013

Lecture Notes in Computer Science

... The TRIPS spoken dialog architecture [127], has been used to develop a number of dialog systems over almost a decade on tasks such as emergency response and evacuation planning [128]. The initial implementation handled turn-taking in the standard rigid way, where as the later version featured incremental interpretation and generation and some other features [129]. ...

Towards a generic dialogue shell
  • Citing Article
  • January 2000

Natural Language Engineering

... Most research on automated grading has focused on ASAG, developed and evaluated on datasets with an average response length of fewer than 20 words, such as Beetle [20] and Texas 2011 [21]. One notable exception is [8], which considers longer answer grading on the RiceChem dataset (average answer length: 120 words) by framing grading as a textual entailment task, achieving approximately 65% accuracy on four questions. ...

The joint student response analysis and recognizing textual entailment challenge: making sense of student responses in educational applications
  • Citing Article
  • August 2015

Language Resources and Evaluation

... Another line of relevant work is on student response analysis in tutoring systems, where the goal is to classify student answers as correct, incorrect, or incomplete. (Dzikovska et al., 2013) explored this in the context of the SemEval-2013 Task 7. In our case, the roles are reversed-we classify the tutor's replies. Nonetheless, techniques such as using pre-trained language models and handling class imbalance are common challenges across these domains. ...

SemEval-2013 Task 7: The Joint Student Response Analysis and 8th Recognizing Textual Entailment Challenge