Ralph R. Martin’s research while affiliated with Cardiff University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (19)


Framework for developing models of shape complexity
Von Mises distribution of turning angles for a polygon
Repetitive visual stimulation, following Attneave [2]
Three L shapes, following Psarra and Grajewski [42]
The training data set with different levels of complexity selected for this study

+6

Visual complexity of shapes: a hierarchical perceptual learning model
  • Article
  • Full-text available

January 2021

·

568 Reads

·

17 Citations

The Visual Computer

Lingchen Dai

·

·

·

[...]

·

Jinhui Yu

Understanding how people perceive the visual complexity of shapes has important theoretical as well as practical implications. One school of thought, driven by information theory, focuses on studying the local features that contribute to the perception of visual complexity. Another school, in contrast, emphasizes the impact of global characteristics of shapes on perceived complexity. Inspired by recent discoveries in neuroscience, our model considers both local features of shapes: edge lengths and vertex angles, and global features: concaveness, and is in 92% agreement with human subjective ratings of shape complexity. The model is also consistent with the hierarchical perceptual learning theory, which explains how different layers of neurons in the visual system act together to yield a perception of visual shape complexity.

Download

Accurate Dynamic SLAM using CRF-based Long-term Consistency

October 2020

·

42 Reads

·

64 Citations

IEEE Transactions on Visualization and Computer Graphics

Accurate camera pose estimation is essential and challenging for real world dynamic 3D reconstruction and augmented reality applications. In this paper, we present a novel RGB-D SLAM approach for accurate camera pose tracking in dynamic environments. Previous methods detect dynamic components only across a short time-span of consecutive frames. Instead, we provide a more accurate dynamic 3D landmark detection method, followed by the use of long-term consistency via conditional random fields, which leverages long-term observations from multiple frames. Specifically, we first introduce an efficient initial camera pose estimation method based on distinguishing dynamic from static points using graph-cut RANSAC. These static/dynamic labels are used as priors for the unary potential in the conditional random fields, which further improves the accuracy of dynamic 3D landmark detection. Evaluation using the TUM \zjc{and Bonn RGB-D dynamic datasets shows that our approach significantly outperforms state-of-the-art methods, providing much more accurate camera trajectory estimation in a variety of highly dynamic environments. We also show that dynamic 3D reconstruction can benefit from the camera poses estimated by our RGB-D SLAM approach.


Livestock detection in aerial images using a fully convolutional network

March 2019

·

3,510 Reads

·

51 Citations

In order to accurately count the number of animals grazing on grassland, we present a livestock detection algorithm using modified versions of U-net and Google Inception-v4 net. This method works well to detect dense and touching instances. We also introduce a dataset for livestock detection in aerial images, consisting of 89 aerial images collected by quadcopter. Each image has resolution of about 3000×4000 pixels, and contains livestock with varying shapes, scales, and orientations. We evaluate our method by comparison against Faster RCNN and Yolo-v3 algorithms using our aerial livestock dataset. The average precision of our method is better than Yolo-v3 and is comparable to Faster RCNN.


Learning guidelines for automatic indoor scene design

February 2019

·

88 Reads

·

3 Citations

In this work, we address a novel and practical problem of automatically generating a room design from given room function and basic geometry, which can be described as picking appropriate objects from a given database, and placing the objects with a group of pre-defined criteria. We formulate both object selection and placement problems as probabilistic models. The object selection is first formulated as a supervised generative model, to take room function into consideration. Object placement problem is then formulated as a Bayesian model, where parameters are inferred with Maximizing a Posteriori (MAP) objective. We solve the placement problem efficiently by introducing a solver based on Markov Chain Monte Carlo with a specific proposal function designed for the problem.



Semantic 3D indoor scene enhancement using guide words

June 2017

·

115 Reads

·

9 Citations

The Visual Computer

We propose a novel framework for semantically enhancing a 3D indoor scene in agreement with a user-provided guide word. To do so, we make changes to furniture colors and place small objects in the scene. The relevance of specific furniture colors and small objects to each guide word is learned from a database of annotated images, taking into account both their frequency and specificity to that guide word. Enhancement suggestions are generated by optimizing a scoring function, which combines the relevance of both enhancement factors, i.e., furniture colors and small objects. During optimization, a submodular set function is adopted to ensure that a diverse set of enhancement suggestions is produced. Our experiments show that this framework can generate enhancement suggestions that are both compatible with the input guide word, and comparable to ones designed by humans.


Automatic Data-Driven Room Design Generation

June 2017

·

132 Reads

·

20 Citations

Lecture Notes in Computer Science

In this work, we address a novel and practical problem of automatically generating a room design from given room function and basic geometry, which can be described as picking appropriate objects from a given database, and placing the objects with a group of pre-defined criteria. We formulate both object selection and placement problems as probabilistic models. The object selection is first formulated as a supervised generative model, to take room function into consideration. Object placement problem is then formulated as a Bayesian model, where parameters are inferred with Maximizing a Posteriori (MAP) objective. By introducing a solver based on Markov Chain Monte Carlo (MCMC), the placement problem is solved efficiently.


Static Scene Illumination Estimation from Videos with Applications

April 2017

·

18 Reads

·

22 Citations

Journal of Computer Science and Technology

We present a system that automatically recovers scene geometry and illumination from a video, providing a basis for various applications. Previous image based illumination estimation methods require either user interaction or external information in the form of a database. We adopt structure-from-motion and multi-view stereo for initial scene reconstruction, and then estimate an environment map represented by spherical harmonics (as these perform better than other bases). We also demonstrate several video editing applications that exploit the recovered geometry and illumination, including object insertion (e.g., for augmented reality), shadow detection, and video relighting.


Fig. 6. Comparison of performance gain (i.e., ∆CC) between saliency-augmented IQMs using fixed and adaptive use of saliency for each saliency models. 
A Saliency Dispersion Measure for Improving Saliency-Based Image Quality Metrics

January 2017

·

49 Reads

·

38 Citations

IEEE Transactions on Circuits and Systems for Video Technology

Objective image quality metrics (IQMs) potentially benefit from the addition of visual saliency. However, challenges to optimising the performance of saliency-based IQMs remain. A previous eye-tracking study has shown that gaze is concentrated in fewer places in images with highly salient features than in images lacking salient features. From this, it can be inferred that the former are more likely to benefit from adding a saliency term to an IQM. To understand whether these ideas still hold when using computational saliency instead of eyetracking data, we first conducted a statistical evaluation using 15 state of the art saliency models and 10 well-known IQMs. We then used the results to devise an algorithm which adaptively incorporates saliency in IQMs for natural scenes, based on saliency dispersion. Experimental results demonstrate this can give significant improvements.


Figure 1: Capture device. Left: camera rig. Center: rig mounted on a car. Right: rig mounted on a tripod.  
Fig. 2. A typical stitched panorama. Region 0 is captured by the upwardspointing camera. Regions 1-5 are captured by the other cameras. Purple lines indicate the boundary seams between overlapping adjacent video streams.
Fig. 3. One dimensional schematic illustration of results produced by different blending algorithms.
A Comparative Study of Blending Algorithms for Realtime Panoramic Video Stitching

May 2016

·

2,450 Reads

·

36 Citations

IEEE Transactions on Image Processing

Panoramic video stitching consists of two major steps: remapping each candidate video stream to its final position and compositing them to generate seamless results. Given videos captured with cameras in fixed relative positions, the remapping step can be done directly using a precomputed look-up table. Greater challenges lie in the more time-consuming composition step. Real world applications typically use blending to perform composition; the performance of the whole system largely depends on the efficiency of the blending algorithm. In this paper, we provide in-depth analysis of the application of several state-of-the-art image blending techniques to realtime panoramic video stitching, as realtime panoramic video stitching enables near-immediate broadcast. Test videos were captured under various conditions, and stitched using various blending methods. Both computational efficiency and quality of composition results were evaluated. Source code and test videos are all publicly available.


Citations (18)


... The ways people process shapes are complicated and still not fully understood [3,36,41]. A range of potential visual features may influence shape perception [10,34], and the role of these features in categorical encoding specifically is not well understood. In this paper, we aim to understand how shapes impact categorical perception in multiclass scatterplots. ...

Reference:

Shape It Up: An Empirically Grounded Approach for Designing Shape Palettes
Visual complexity of shapes: a hierarchical perceptual learning model

The Visual Computer

... DynaSLAM II [23] tightly couples the scene structure, camera poses, and dynamic object trajectories within the same optimization window. LC-CRF SLAM [24] utilizes dynamic feature partitioning methods based on graph cuts and Conditional Random Fields to segment dynamic pixels in the scene. DRG-SLAM [25] proposes a dynamic feature extraction method that combines semantic segmentation and epipolar constraints, and it also improves the robustness of the SLAM system in weakly textured and dynamic environments using point-line features. ...

Accurate Dynamic SLAM using CRF-based Long-term Consistency
  • Citing Article
  • October 2020

IEEE Transactions on Visualization and Computer Graphics

... There is unanimity among researchers on the scarcity of publicly available, high-quality CV datasets of livestock (Bahlo and Dahlhaus, 2021;Han et al., 2019;Ocholla et al., 2024;Ong et al., 2023;Psota et al., 2019;Vayssade et al., 2023). While the volume of livestock data in public sources continues to grow with additions from individuals and research institutions, it remains far below the scale of generic CV datasets like ImageNet (Russakovsky et al., 2015) and MS COCO (Lin et al., 2014). ...

Livestock detection in aerial images using a fully convolutional network

... For multimedia applications such as virtual reality and computer games, room design issues arise. Room functions and basic geometry automatically generate room designs, select appropriate objects from a library, and formulate both object selection and placement problems as probabilistic models [5]. With the development of 3D tool software, designers can quickly design rooms with different details. ...

Learning guidelines for automatic indoor scene design

... The An important question then arises regarding the language in which these rules should be represented. Earlier work, such as Merrel et al. [10] and Yu et al. [51], proposed rules as geometric formulas, based on shapes, angles, and distance, and many subsequent approaches since have followed suit ( [47]- [50], [52]). Such representations are not easy to explain to occupants who tend to think in terms of furnishings, and, for the same reason, they are also difficult for interior design experts to express. ...

Automatic Data-Driven Room Design Generation
  • Citing Conference Paper
  • June 2017

Lecture Notes in Computer Science

... Güneş et al. examined living room wall colors and concluded that red represents disgust and happiness, green represents happiness and blue represents neutrality [47]. Kaya et al1 stated that among residential building types, purple is the least preferred color, and blue and red are more associated with residential building types [48]. It can be seen that the color planning of interior design is really important. ...

Semantic 3D indoor scene enhancement using guide words

The Visual Computer

... The existing shadow generation methods can be divided into two categories: rendering-based and image-to-image translation methods. Some renderingbased shadow generation methods [54][55][56] require strong user interactions to obtain explicit illumination condition, reflectance, and scene geometry to generate plausible shadows. Although these methods [57][58][59] recover illumination information and scene geometry from a single image, inaccurate estimation typically produces unsatisfactory results. ...

Static Scene Illumination Estimation from Videos with Applications
  • Citing Article
  • April 2017

Journal of Computer Science and Technology

... Measuring the perceptual contrast or how 'contrasty' or 'punchy' an image appears is a challenging task. Inspired by multi-level approaches in entropy computation [46] and structural similarity measures [42], we present our own approximation of a multi-level contrast measure. Multi-level contrast follows a multi-grid approach where at each level n, the full resolution image is divided into a grid of n×n patches and patch-specific variance of pixel intensity is computed. ...

A Saliency Dispersion Measure for Improving Saliency-Based Image Quality Metrics

IEEE Transactions on Circuits and Systems for Video Technology

... All algorithms, except for UDIS that does not include an explicit blending step, were evaluated with a multi-band blending post-processing, since that has been established as the optimal blending algorithm in terms of quality/speed trade-off [36]. Noticing that the best results among the mesh-based models were given by REW [25], only the REW results are shown in this section. ...

A Comparative Study of Blending Algorithms for Realtime Panoramic Video Stitching

IEEE Transactions on Image Processing