Akio Kosaka’s research while affiliated with Purdue University West Lafayette and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (42)


Estimation of subjective age based on facial images of others: Comparative studies of the Americans and the Japanese
  • Conference Paper
  • Full-text available

February 2011

·

23 Reads

Yasuhiro Azuma

·

Masato Konishi

·

Naoyuki Miyamoto

·

[...]

·

Akio Kosaka

We designed a method for estimating the subjective age of a person; using this method, one evaluates one's own age by estimating whether a person shown in a facial image looks older or younger than oneself. Thus far, experiments have shown that Japanese people tend to underestimate their subjective age. In this study, we conducted an international comparative study to focus on the socio-psychological effects that may influence the underestimation of subjective age. Experiments were performed in which American and Japanese participants viewed American facial images; in addition, the American participants viewed Japanese facial images. Through these experiments, it was confirmed that the underestimation of age occurs despite differences in Japanese and American societies and cultures; however, the tendency of underestimation of age is not related to facial images but to social and cultural factors that influence the participants. Moreover, we carried out experiments of conventional look age estimates and compared the results to those of the subjective age estimates.

Download

Figure 1. Shift values of subjective age (Participant: Japanese, Facial images: Japanese) 
Figure 2. Examples of facial images (Class: 25–29 years) (We have obtained individual permissions to publish these photographs.) C. Rating experiment For both male and female participants belonging to different age and gender groups, we selected facial images from their own age class as well as the next younger and older classes. For example, for a 32-year-old male participant, 5 male and 5 female images were selected from the same age class (30–34) and from the next younger (25–29) and older (35–39) classes. Therefore, the total number of facial images used was 30 (5 images/class × 3 classes × 2 genders). Next, we experimented with a rating scale for the facial images used as stimuli. The participants were shown random facial images on the computer monitor, and they evaluated whether the person shown in the image looked older or younger than themselves. The evaluation had 5 ranks: “Definitely older than myself (2),” “Probably older than myself (1),” “Not able to estimate (0),” “Probably younger than myself (–1),” and “Definitely younger than myself (–2).” A range of responses was adopted in order to determine the relative position to other ages as opposed to estimating the chronological ages of the people whose facial images were shown. 
Figure 3. Example of non-linear regression 
A Comparative Assessment of One's Own Age from Facial Images of Others: Two Case Studies for the Americans and the Japanese

November 2009

·

115 Reads

·

6 Citations

We have proposed a method for estimating the subjective age of a person; that is, a method that yields a person's age on the basis of estimations made by that person about the age of other people by observing their facial images. Thus far, experiments have shown that Japanese people tend to underestimate their subjective age. In this study, we focus on the socio-psychological effects that may influence the underestimation of subjective age. We conducted an international comparative study. In this study, experiments were performed in which American and Japanese participants viewed American facial images; in addition, the American participants also viewed Japanese facial images. Through these experiments, it was confirmed that the subjective age generally tended toward the negative direction despite differences in the nationalities and cultures of the Japanese and American participants and their facial images. Moreover, it was found that nationality and culture may have some effects on the estimation; for example, American males did not exhibit the tendency to underestimate the age, unlike Japanese male. When estimating facial images of different nationalities, the variance generally tended to be larger, although the average was similar. This study suggests that the underestimation of age occurs despite differences in Japanese and American societies and cultures; however, the tendency of underestimation of age is not related to facial images but to social and cultural factors that influence the participants.


Motion detection for subcellular structure trafficking

September 2009

·

13 Reads

·

1 Citation

Conference proceedings: ... Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Conference

A detailed understanding of subcellular structure motility is critical to understanding how cells regulate the delivery of specific proteins from the site of synthesis to the site of action. Diverse modalities have been observed for subcellular dynamics, such as directional movement, random movement, tethered movement, object appearing, and disappearing. Motion modality detection is important for efficient subcelluar structure tracking and population study. In this paper, we present a new technique, called "divergence filter," for detecting subcellular structure motion modalities. The plausibility of the efficient technique was applied to caveolar membrane trafficking sequences obtained using confocal microscopy.


Hand–Eye Calibration Applied to Viewpoint Selection for Robotic Vision

November 2008

·

509 Reads

·

69 Citations

IEEE Transactions on Industrial Electronics

Viewpoint calibration is a method to manipulate hand-eye for generating calibration parameters for active viewpoint control and object grasping. In robot vision applications, accurate vision sensor calibration and robust vision-based robot control are essential for developing an intelligent and autonomous robotic system. This paper presents a new approach to hand-eye robotic calibration for vision-based object modeling and grasping. Our method provides a 1.0-pixel level of image registration accuracy when a standard Puma/Kawasaki robot generates an arbitrary viewpoint. To attain this accuracy, our new formalism of hand-eye calibration deals with a lens distortion model of a vision sensor. Our most distinguished approach of optimizing intrinsic parameters is to utilize a new parameter estimation algorithm using an extended Kalman filter. Most previous approaches did not even consider the optimal estimates of the intrinsic and extrinsic camera parameters, or chose one of the estimates obtained from multiple solutions, which caused a large amount of estimation error in hand-eye calibration. We demonstrate the power of this new method for: (1) generating 3-D object models using an interactive 3-D modeling editor; (2) recognizing 3-D objects using stereovision systems; and (3) grasping 3-D objects using a manipulator. Experimental results using Puma and Kawasaki robots are shown.


A New Kalman-Filter-Based Framework for Fast and Accurate Visual Tracking of Rigid Objects

November 2008

·

81 Reads

·

76 Citations

IEEE Transactions on Robotics

The best of Kalman-filter-based frameworks reported in the literature for rigid object tracking work well only if the object motions are smooth (which allows for tight uncertainty bounds to be used for where to look for the object features to be tracked). In this contribution, we present a new Kalman-filter-based framework that carries out fast and accurate rigid object tracking even when the object motions are large and jerky. The new framework has several novel features, the most significant of which is as follows: the traditional backtracking consists of undoing one-at-a-time the model-to-scene matchings as the pose-acceptance criterion is violated. In our new framework, once a violation of the pose-acceptance criterion is detected, we seek the best largest subset of the candidate scene features that fulfill the criterion, and then continue the search until all the model features have been paired up with their scene correspondents (while, of course, allowing for nil-mapping for some of the model features). With the new backtracking framework, our Kalman filter is able to update on a real-time basis the pose of a typical industrial 3-D object moving at the rate of approximately 5 cm/s (typical for automobile assembly lines) using off-the-shelf PC hardware. Pose updating occurs at the rate of 7 frames per second and is immune to large jerks introduced manually as the object is in motion. The objects are tracked with an average translational accuracy of 4.8 mm and the average rotational accuracy of 0.27deg.


A novel evidence accumulation framework for Robust multi-camera person detection

October 2008

·

242 Reads

·

12 Citations

We propose a novel evidence accumulation framework that accurately estimates the positions of humans in a 3D environment. The framework consists of a network of distributed agents having different functionalities. The modular structure of the network allows scalability to large surveillance areas and robust operation. The framework does not assume reliable measurements in single cameras (referred to as dasiasensing agentspsila in our framework) or reliable communication between different agents. There is a position uncertainty associated with single camera measurements and it is reduced through an uncertainty reducing transform that performs evidence accumulation using multiple camera measurements. Our framework has the advantage that single camera measurements do not need to be temporally synchronized to perform evidence accumulation. The system has been tested for detecting single and multiple humans in the environment. We conducted experiments to evaluate the localization accuracy of the position estimates obtained from the system by comparing them with the ground truth. Also, two different configurations of the agents were tested to compare their detection performance.


Stereo Vision for Industrial Applications

November 2007

·

32 Reads

·

10 Citations

Introduction to Stereo Vision Mathematical Preliminaries Basic Strategy for Model-Based Stereo Vision Stereo Vision for Dump Truck Recognition and Localization Model-Based Stereo Vision for Recognition and Localization of Alternator Covers Conclusions References


A multi-Kalman filtering approach for video tracking of human-delineated objects in cluttered environments (DOI:10.1016/j.cviu.2006.02.003)

June 2006

·

30 Reads

·

35 Citations

Computer Vision and Image Understanding

In this paper, we propose a new approach that uses a motion–estimation based framework for video tracking of objects in cluttered environments. Our approach is semi-automatic, in the sense that a human is called upon to delineate the boundary of the object to be tracked in the first frame of the image sequence. The approach presented requires no camera calibration; therefore it is not necessary that the camera be stationary. The heart of the approach lies in extracting features and estimating motion through multiple applications of Kalman filtering. The estimated motion is used to place constraints on where to seek feature correspondences; successful correspondences are subsequently used for Kalman-based recursive updating of the motion parameters. Associated with each feature is the frame number in which the feature makes its first appearance in an image sequence. All features that make first-time appearances in the same frame are grouped together for Kalman-based updating of motion parameters. Finally, in order to make the tracked object look visually familiar to the human observer, the system also makes its best attempt at extracting the boundary contour of the object—a difficult problem in its own right since self-occlusion created by any rotational motion of the tracked object would cause large sections of the boundary contour in the previous frame to disappear in the current frame. Boundary contour is estimated by projecting the previous-frame contour into the current frame for the purpose of creating neighborhoods in which to search for the true boundary in the current frame. Our approach has been tested on a wide variety of video sequences, some of which are shown in this paper.


Erratum to A multi-Kalman filtering approach for video tracking of human-delineated objects in cluttered environments [Comput. Vision Image Understanding 99 (2005) 1-57]

June 2006

·

7 Reads

·

3 Citations

Computer Vision and Image Understanding

This paper proposes a novel method, using constant inter-frame motion, for self-calibration from an image sequence of an object rotating around a single axis with varying camera internal parameters. Our approach makes use of the facts that in many commercial ...


A Deformable Model for Automatic CT Liver Extraction1

October 2005

·

18 Reads

·

23 Citations

Academic Radiology

This study was performed to design an automatic liver region extraction system to facilitate clinical liver size estimation and further serve as a prestage for liver reconstruction and volume estimation. We present a modification of the well-known snakes algorithm for extracting liver regions in noisy CT images. Our modification addresses the issues of selection of the control points on an estimate of the contour and the determination of the weighting coefficients. The weighting coefficients are determined dynamically on the basis of the distance between the control points and the local curvature of the contour. The proposed method was used in extracting liver regions from 98 cross-sectional abdominal images. The overall performance was estimated by comparisons with original liver regions. The deformable model method enables an efficient and effective automatic liver region extraction in noisy environments. This approach eliminates human-in-the loop, which is the common practice for the majority of current methods.


Citations (31)


... The technology of 3D vision has been widely applied to really life in a number of fields. 3D vision is commonly utilized to realize the tasks including human activity detection, 3D terrain mapping, navigation, obstacle detection and bio-inspired autonomous guidance [19]. Nishida and Kitamura raised an application detecting human activity that can quickly detect the human daily activity events in the real world [20]. ...

Reference:

A Survey of Learning Approaches and Application for 3D Vision
Stereo Vision for Industrial Applications
  • Citing Chapter
  • November 2007

... The grid-based framework (examples are FINALE 1 and Moravec 12 ) requires a geometrical representation of space and relies on the camera image to pinpoint landmark features in the ''vicinity'' of the robot for determining its position. Another camera-based strategy is a navigation system, which uses a topological representation of space and an ensemble of neural networks to guide a robot through interior space (examples are NEURO-NAV 13,14 and FUZZY-NAV 15 ). To give an example of the capabilities of these camera-based navigation techniques, FINALE can (equipped with a high-performance camera under optimal conditions) navigate a robot in motion at an average speed of 17 m/min using an ordinary Personal-computer-(PC)-based self-positioning algorithm that processes images based on camera data. ...

FUZZY-NAV: A Vision-Based Robot Navigation Architecture using Fuzzy Inference for Uncertainty-Reasoning
  • Citing Article
  • January 1995

... Gao et al. [38] used extended Kalman filter for tracking an object in a cluttered environment. The method is semi-automated whereby the user helps in the initial segmentation by selecting the region of interest. ...

Erratum to A multi-Kalman filtering approach for video tracking of human-delineated objects in cluttered environments [Comput. Vision Image Understanding 99 (2005) 1-57]
  • Citing Article
  • June 2006

Computer Vision and Image Understanding

... Debido a la experiencia que se tenía en aplicaciones marinas del sonar, este fue uno de los primeros sensores en ser utilizado en la robótica móvil. Actualmente, nuevos y más sofisticados sistemas han sido desarrollados, como son cámaras de video y rayos laser que compiten con el sonar [5,6]. Pero el sonar aún continúa mostrando ventajas técnicas y económicas con respecto a los otros sistemas mencionados. ...

Vision-guided mobile robot navigation using retroactive updating of position uncertainty
  • Citing Conference Paper
  • June 1993

... Dengan demikian, pelacakan menjadi masalah optimasi sederhana yang diselesaikan dengan cara menghitung ulang bobot kuadrat terkecil secara iteratif. Yoon et al. [15] menyajikan pelacakan objek berbasis model untuk menghitung pose kamera 3D. Algoritma yang digunakan adalah Extended Kalman Filter (EKF) untuk memberikan skema pembaruan pose tambahan dalam kerangka prediksi-verifikasi. ...

A New Approach to the Use of Edge Extremities for Model-based Object Tracking

... Based on our VFC architecture, we processed sensing cooperation for autonomous driving. We consider two kinds of sensing cooperation tasks which are accumulative sensing tasks such as [32] and best-quality sensing tasks such as [33]. We first adopt greedy algorithm to optimal cooperative sensing strategy to enhance sensing coverage. ...

A novel evidence accumulation framework for Robust multi-camera person detection

... The annoying thing is that discrepancies usually arise between the estimated positions of objects and the truth, primarily due to environmental factors, [14], [15], [16], [17]. Since discrepancies are not white, they can be treated as colored measurement noise (CMN), making accurate detection difficult, [18], [19], [20], [21], [22]. ...

A New Kalman-Filter-Based Framework for Fast and Accurate Visual Tracking of Rigid Objects
  • Citing Article
  • November 2008

IEEE Transactions on Robotics

... To accomplish this, precise conversion of object locations from the image coordinate system to the robot coordinate system is required. The general procedure to achieve it involves constructing a precise mathematical model for accurate conversion [1][2][3][4][5][6] and optimizing this model through calibration [7][8][9][10][11][12][13][14][15], known as hand-eye calibration. This can be accomplished by taking images of a board with a specific pattern, such as a checkerboard pattern, from various positions and then using the image group taken to optimize a mathematical model. ...

Hand–Eye Calibration Applied to Viewpoint Selection for Robotic Vision

IEEE Transactions on Industrial Electronics

... Moreover, different ethnicities and genders might present different types of skin texture. A review showed that the variance in human estimation of age based on facial images of different nationalities generally tended to be larger [8]. Therefore, the impact of skin texture difference on age estimation is a potential research area. ...

A Comparative Assessment of One's Own Age from Facial Images of Others: Two Case Studies for the Americans and the Japanese