Science topic

Object Recognition - Science topic

Explore the latest questions and answers in Object Recognition, and find Object Recognition experts.
Questions related to Object Recognition
  • asked a question related to Object Recognition
Question
2 answers
Deep Learning for Computer Vision
Explore how deep learning is revolutionizing the way machines see and understand the world around us!
🔍 In this lecture, we delve into:
✅ The power of Convolutional Neural Networks (CNNs)
✅ Image and Object Recognition
✅ Semantic Segmentation and Localization
✅ Advanced Object Detection techniques like RCNN, Fast-RCNN, and Faster-RCNN
🎥 Watch the full lecture here: https://www.youtube.com/watch?v=Ql0sApkfXpk
Join us as we uncover cutting-edge techniques and their applications in this exciting domain of AI. Let's shape the future together!
#DeepLearning #ComputerVision #AI #MachineLearning #Education #Innovation
Relevant answer
Answer
Dear Rahul Jain ,
Computer vision (CV) is the scientific field which defines how machines interpret the meaning of images and videos. Computer vision algorithms analyze certain criteria in images and videos, and then apply interpretations to predictive or decision making tasks.
Today, deep learning techniques are most commonly used for computer vision. This article explores different ways you can use deep learning for computer vision. In particular, you will learn about the advantages of using convolutional neural networks (CNNs), which provide a multi-layered architecture that allows neural networks to focus on the most relevant features in the image.
Regards,
Shafagat
  • asked a question related to Object Recognition
Question
3 answers
Seeking insights on leveraging deep learning techniques to improve the accuracy and efficiency of object recognition in machine vision systems.
Relevant answer
Answer
Identification. Image classification using deep learning categorizes images or image regions to distinguish between similarly looking objects including those with subtle imperfections. Image classification can, for example, determine if the lips of glass bottles are safe or not.
Regards,
Shafagat
  • asked a question related to Object Recognition
Question
3 answers
Hello Everybody,
As the title says, I am searching for a public 3D Object Recognition and Pose Estimation Dataset.
I've utilized Google for three days, so I thought I might as well ask here.
The dataset should contain model and scene pointcloud data, ideally stored in pcd or ply format. (I am working with the point cloud library)
There is no need for thoundand files of training data, 10 files f.e. would be fine as I just want to evaluate an object recognition pipeline fast.
So basically the algorithms tries to fit 3D models clouds into a point cloud of a scene photographed by an ASUS Xtion camera.
Relevant answer
Answer
In the field of autonomous driving there is a well-known dataset KITTI ( ) which contains point clouds, if this domain works for your research.
  • asked a question related to Object Recognition
Question
11 answers
I am currently searching for a topic for my research which is about using machine vision and object recognition to control a robot (serial or parallel it does not matter). Unfortunately I can not find a problem to be solved. can any one recommend some new points of research ?
Relevant answer
Answer
Robotics are used in many materials handling applications because they are more efficient in certain tasks and take people out of potentially unsafe situations. To work effectively, robotics rely on sensors to interact and perceive their environment. This white paper from ifm illustrates the importance of 3D image sensors in robotics and the systems used to operate and manage these.
The interaction of the human eye with the visual centre in the brain creates a three-dimensional image of the environment. In robotics, such a three-dimensional image is important to enable robots to act independently and without causing any danger outside of safety barriers...
  • asked a question related to Object Recognition
Question
3 answers
I would like suggestions that mainly be able to analyze the New Object Recognition Test (NORT). Free software!
Relevant answer
  • asked a question related to Object Recognition
Question
4 answers
To answer this question, you should define what is meant by object recognition. You can then include research evidence showing how object recognition is performed.
Relevant answer
Answer
Pattern/object recognition is concerned with the processes involved in the identification of images and objects. This essentially involves taking information that enters the visual system and comparing this with information stored in memory, and finding a match. There are three approaches within pattern recognition; template and prototype theories, feature comparison theories and structural theories. The focus of this essay is feature comparison theories, their advantages and disadvantages and their overall success in pattern/object recognition.
Regards,
Shafagat
  • asked a question related to Object Recognition
Question
6 answers
Hi Everyone,
I'm currently practising an object detection model which should detect a car, person, truck, etc. in both day and night time. Now, I have started gathering data for both day and night time. I'm not sure whether to train a separate model for daylight and another model for the night-light or to combine together and train it?
can anyone suggest to me the data distribution for each class at day and night light? I presume it should be a uniform distribution. Please correct me if I'm wrong.
Eg: for person: 700 images at daylight and another 700 images for nightlight
Any suggestion would be helpful.
Thanks in Advance.
Relevant answer
Answer
Ben Harper, thanks very much for your recommendation.
  • asked a question related to Object Recognition
Question
5 answers
Hi,
Are there datasets for objects recognition for industrial applications ? as tools
  • asked a question related to Object Recognition
Question
8 answers
I want to Identify darkest object in uploaded image. I have tried Imagej. In IMAGEJ, for each image I have to do different threshold and analyzing. Here some are getting excluded for different images with same value of threshold. I want to learn if counting is possible very accurate automatically with some image processing technique. Is it possible to identify with OpenCV?
Relevant answer
Answer
Since, your problem is to identify darkest object. you can get the histogram of the image. Then apply adapive thresholding to find the threshold automatically (threshold will be different for different images but that will be automated)
Moreover, it seems that the objects shape is hexagonal that you want to detect. I think you can train a HOG+SVM classifier with bounding boxes.
  • asked a question related to Object Recognition
Question
3 answers
Dear colleagues:
I had an interesting discussion with my labmates about what is considered exploration in object location or object recognition tasks. In my experience/opinion, object exploration performed by roedents is defined just when the animal snif directly the object, with a clear "intention" to explore (figure1). According to my labmates and published pappers, exploration is defined when the animal brings its head closer to the object. Automated softwares quantify any entry in a determined circle around the object.
My concerns are related to:
1.- when the the animal sometimes hides behind an object, in this scenario is not exploring, just being there, sometimes even immobile (figure2).
2.- sometimes rats use the object to rear and snif pointing its nose to the upper part of the arena, not in the object direction (figure3).
Can you share your opinions about this concern????
Relevant answer
Answer
It is by opposing intent that we prove its existence. (That comes obliquely here from security operations which have a duty through unambiguous warnings or non-lethal opposition to prove hostile intent in challengers before the use of force).
So, if you dangle the object above the ground but in the limits of reach, then your subjects prove their exploratory intent by both stretching up and sniffing. All sniffers are then more obviously explorers, and hiders and recliners are ruled out.
More generally, if the subjects had to overcome discomfort or fear in order to venture ever closer to an object, then it's much more likely that their intent is to explore. The greater or more numerous the obstacles, the greater the intent.
  • asked a question related to Object Recognition
Question
2 answers
Dear community,
I'm looking into ways how to do an a-priori power analysis for an fMRI experiment where the main analysis will be a representational similarity analysis (RSA).
The experiment will present the same stimuli in two successive fMRI-sessions (with a behavioral training in between). For each fMRI session, I plan to do a model-based RSA on brain responses elicited by the stimuli. Voxel restriction will be done with a searchlight procedure. The most interesting outcome will be the difference in these results between the two training sessions, as estimated with a GLM contrast.
I think this is not uncommon as i found other experiments adopting a similar analysis procedure. I found no clue however on how to estimate the necessary sample size to achieve a certain statistical power (say 80%).
Since this is a bit of a frankenstein made from other common statistical approaches, I'm not sure if the general logic of fMRI-power analysis applies here.
Has anybody experience in this area or can point me to literature that contemplates this issue?
Thanks,
Oliver
Relevant answer
Answer
Topic is bit dusty, but I wonder if you got any answers?
  • asked a question related to Object Recognition
Question
3 answers
I want to start a project in which I want to use machine learning tools (neural networks) to recognise objects. For this purpose I am looking for the right hardware component with respect to camera and enlightning.
The following requirements to the hardware are given:
The resolution shall be at least 4000x3000 pixels or 4K.
The device shall be designed for continuous operation. It must not shut down after a certain time.
The depth of focus shall be good enough. The camera is about 1.5 meters away from the objects and it may occur that some objects have large packing heights.
In general exposure time, aperture and focus shall be able to be fixed manually.
Also the products shall not be too expensive in comparison with similar products.
Relevant answer
Answer
Depends upon many conditions, e.g. The algorithm you want to use, any integrated cameras or sensors, and your budget.
  • asked a question related to Object Recognition
Question
14 answers
I am using UAV data for mapping geomorphological processes in different environments, from coastal and estuarine subtropical areas to subpolar and polar glacial landscapes, and I want to profit from the huge amount of information in such high-resolution datasets. So, I was wondering if there are good free options for object-oriented image classification, alternative to eCognition for example?
  • asked a question related to Object Recognition
Question
16 answers
Hello. I am an undergraduate student and I need to collaborate with my professor (who specialises in Computer Vision, Image Processing and 3-D medical imaging). I am looking for research ideas mainly in the topics of object detection, visual object tracking, object recognition, semantic segmentation, localization using u-net or medical imaging. Can anybody help me jot down growing research fields in these areas? Thank you!
Relevant answer
Answer
in computer vision Deep Learning is recently being used very widely by many organization including Microsoft, Google. Deep Learning uses deep net to simulated many layers of non linear processing . feature extraction and transformation and finally cascade them into a more accurate approximation of object identification. also you can check the following ideas:
  1. object detection
  2. efficient video understanding
  3. new applications using CV and related techniques (e.g. NLP)
  • asked a question related to Object Recognition
Question
7 answers
I am looking for visual stimuli that produce a similar effect than the well know “Dalmatian dog illusion” (see Figure attached).
If you look briefly  at the Dalmatian dog illusion  for the first time, it looks like a pattern of meaningless black and white stains (left panel). However, once a priming cue is briefly presented (the red contour in the right panel) the representation of a Dalmatian dog in a field becomes apparent. Once “seen” this representation will remain apparent even after the priming cue is removed and can't be unseen (look at the left panel again without the red contour).
Do you know other types of visual stimuli containing a hidden object shape that never pops-out before and always pops-out after the transient presentation of a priming stimulus? 
Thank you!
Relevant answer
Answer
Here is a similar one, which you will never be able to not see it again once you've found it:
  • asked a question related to Object Recognition
Question
3 answers
Which algorithms makes use of least squares in object Recognition? Is the least squares approximation use in calculating Principal Component Analysis (PCA) and Singular Value Decomposition (SVD) in machine learning?
  • asked a question related to Object Recognition
Question
3 answers
How many hidden layers are there in Faster Region Convolutional neural network used for object recognition?
Relevant answer
Answer
Hi Aravinda Kasukurthi,
The Faster R-CNN has the same number of hidden layers as the Fast R-CNN, the RPN has no hidden layers and is only used as a feature extractor. The Fast R-CNN has three fully connected layers.
  • asked a question related to Object Recognition
Question
12 answers
In image processing, an image is "processed", that is, transformations are applied to an input image and an output image is returned. The transformations can e.g. be "smoothing", "sharpening", "contrasting" and "stretching". The transformation used depends on the context and issue to be solved.
In computer vision, an image or a video is taken as input, and the goal is to understand (including being able to infer something about it) the image and its contents. Computer vision uses image processing algorithms to solve some of its tasks.
The main difference between these two approaches are the goals (not the methods used). For example, if the goal is to enhance an image for later use, then this may be called image processing. If the goal is to emulate human vision, like object recognition, defect detection or automatic driving, then it may be called computer vision.
Relevant answer
Answer
Dear Rohit Yadav,
In image processing, an image is "processed", that is, transformations are applied to an input image and an output image is returned. ... Computer vision uses image processing algorithms to solve some of its tasks. The main difference betweenthese two approaches are the goals (not the methods used).
Regards,
Shafagat
  • asked a question related to Object Recognition
Question
8 answers
Hello everyone,
I have a group of fish (n=28 per group). I have measured the amount of time that each fish spent exploring two different objects (A and B) in a squared shape tank. Then I have calculated an exploration index as: (time spent with A) / (time spent with B + time spent with A).
Now I have some proportional data and I am going to test whether a fish spent time with A more than chance level (0.5) to see if they have a preference for object A. To see if there is a difference between exploration ratios and chance level, I am gonna run one sample t-test. However, as I read some papers, I think I should do arcsin transformation (arcsine square root transformation) for my data. However, when I transform my data and run one-sample t-test I get odd results. For example, my raw exploration ratios show that they are below chance level (mean: 0.322 and SD:0.12) while after transformation of data I see that the ratio are now higher than chance level. I am really confused and I do not now how I should treat with my data. I would really appreciate it if you can advise me
Relevant answer
Answer
You need to transform the data because you are applying a parametric test that requires the fulfillment of two main assumptions: normality and homocedasticity.
The angular transformation (or arcsine transformation) is suitable for proportional data, like yours, because they are restricted in the range 0-1. In this case, one of the recommended transformations is the Arcsine transformation.
I think your confusion is regarding the value to which to compare your proportions after the transformation.
Angular transformation: W = [ASIN (SQRT (p))]
Being: W = transformed value; p = the original proportion.
For p = 0.5 (the value you will use for your test)
W = [ASIN (SQRT (0.5))] = 0.785. That value W = 0.785 will be the one used in your
one-sample t-test.
I hope that is your doubt.
Regards !
  • asked a question related to Object Recognition
Question
12 answers
In the very first frame of the video, I define a ROI by drawing a close line on the image. The goals is to recognize that ROI in a later frame, but that ROI is not a salient object. It is just a part of an object, and it can deform, rotate, translate and even not be fully in the frame.
Essentially, this algorithm should be used to reinitialize trackers once they are lost.
I have used a histogram based algorithm which works somewhat well, but it doesn't "catch" the ROI entirely.
The object is a soft and deformable object, soft tissue in a way, meaning you can expect deformations and also visual changes due to lightning.
  • asked a question related to Object Recognition
Question
4 answers
I have a data for image recognition using neural networks. The images are in pgm format.how to pre-process that data to get into a suitable matrix in cpp.
  • asked a question related to Object Recognition
Question
8 answers
I have a tif image of size around ~10 Gb. I need to perform object classification or pixel classification in this image. The dimension of image data has zyx form. My voxel size in x=0.6, y=0.6 and z=1.2.
Z is the depth of the object.
If I do classification of pixels in each Z plane separately and then merge to get the final shape and volume of object. Would I loose any information and my final shape or volume of object will be wrong?
Relevant answer
Answer
If your data have a very high slice spacing in 3rd dimension and you are getting 2d slices results, you need to integrate neighbouring slices, with a fixed spacing of 1 mm using a linear interpolation to deal with this problem.
  • asked a question related to Object Recognition
Question
8 answers
like a MWM, EPM and object recognition test free software also help how to use and download it
thanks
Relevant answer
Answer
I've used Ctrax successful, but heard that ToxTrac is quicker at extracting data from video files
  • asked a question related to Object Recognition
Question
7 answers
Dear colleagues,
I am looking for the URLs file of the VALIDATION set of ImageNet Large Scale Visual Recognition Competition (ILSVRC) 2012.
I can easily find that of the training set. However, I have troubles reaching out the validation set's file.
BTW, I have the original image set. I just need the source URLs.
Thanks for help.
Relevant answer
Answer
Dear Alaa,
you may find all relevant information on the ILSVRC page [1]. If you already have the original image set, you are fine. The devkit contains the ground truth annotations for the validation images, which are named like "ILSVRC2012_val_00000001.JPEG" etc.
  • asked a question related to Object Recognition
Question
4 answers
I want to code for object recognition using deep learning, where I do not have any database for supervised approach.
I want to perform the same using unsupervised deep learning approach.
Can you please guide me to focus those possible methods to go through for object recognition.
Relevant answer
Answer
Dear Md Junayed Hasan,
Thank you for your time and answer..
I will go through it...
  • asked a question related to Object Recognition
Question
5 answers
I have a project in matlab. I need to recognize the color of that car. 
On other words , How can I recognize any object color using matlab and wrote the result as text for example if car has a red color, then the result will be "red" as text .
  • asked a question related to Object Recognition
Question
8 answers
I am looking for the state of the art methods which are being used for object recognition on moving platforms.
Relevant answer
Answer
This review paper enlist recent techniques.
  • asked a question related to Object Recognition
Question
18 answers
Please do introduce some paper or site.
Relevant answer
Answer
1. Haar Cascade Classifier
2. Template matching
3. Deep CNN
4. Contour finding after background subtraction in case of video
  • asked a question related to Object Recognition
Question
11 answers
Formally, a computer program should be able to scan such an image, perform image processing or any specific treatment on it and produce the followings.
1) The info about multiple geometrical shapes stacked together in front/side and top views in the image.
2) The correlation between the two views as some shapes (or a part of a shape) are hidden in one view but their projections are seen in the other view.
3) The relationship among shapes such as the orientation of a shape w.r.t. to each other.
4) The info about dimensions normally written in text besides arrows.
5) The info about arrows, single-headed, double-headed, straight, slanted, etc.
Relevant answer
Answer
Fabio, is " Convex Polygon Fitting in Robot-Based Neurorehabilitation" the work you referenced which has all the examples and the code?? I have sent you a request for the article. Could you please send it across? Thanks
  • asked a question related to Object Recognition
Question
10 answers
I have a small set of chest CT scans. I am interested to use a deep neural network for denoising of these images. However due to small size of data, I can not train the network, hence I am looking for pretrained networks. I am aware of pretrained CNNs for object recognition or feature extraction (VGG, Resnet, etc.) but not any for denoising). I appreciate any suggestions.
Thanks,
Nastaran Emaminejad
Relevant answer
Answer
Xie, J., Xu, L., & Chen, E. (2012). Image denoising and inpainting with deep neural networks. In Advances in neural information processing systems (pp. 341-349).
Dong, W., Wang, P., Yin, W., Shi, G., Wu, F., & Lu, X. (2018). Denoising Prior Driven Deep Neural Network for Image Restoration. arXiv preprint arXiv:1801.06756.
Lucas, A., Iliadis, M., Molina, R., & Katsaggelos, A. K. (2018). Using Deep Neural Networks for Inverse Problems in Imaging. IEEE SIgnal ProcESSIng MagazInE, 1053(5888/18).
  • asked a question related to Object Recognition
Question
5 answers
Can anyone help me about the best method to classify facial expression database? I have tried using FFT and SVM, but still it's just based on the whole features in image. It doesn't necessarily focus to mouth or eye expression. Thank you
Relevant answer
Answer
Traditionally, Local Binary Patterns (LBP) has been successfully used as features for facial expression recognition. It is very important to extract LBP from regions in a grid over the face image and concatenate them in a single vector.
  • asked a question related to Object Recognition
Question
9 answers
With a list of models (CNN, FFNN, RNN, etc) performances? A kind of MNIST for VOR?
A want to compare performances to well-known models in computer vision.
Relevant answer
Answer
Dear Nils Schaetti,
Look the link, maybe useful.
Regards, Shafagat
  • asked a question related to Object Recognition
Question
6 answers
The research content includes a proposed algorithm for image/object matching and two proposed algorithms for multiple object detection.
The algorithms for image/object matching and multiple object detection are not related.
My question is how to organize them to form a Phd thesis? How to unify them into a big problem to present? What title is appropriate?
Relevant answer
Answer
You should probably try to find a problem you can solve combining/pipelining both types of algorithms, i.e. pretend you had an ultimate goal when you worked on them both. I don't know exactly which algorithms you are talking about, but lets say you are detecting people approaching a government building and you want to identify a felon among them or something like this. Depends on your specific work, really..
  • asked a question related to Object Recognition
Question
9 answers
Hi,
I will be running an experiment that requires participants to distinguish between several novel objects. Ideally, each novel object will be a configuration of 3D geometric shapes (e.g., pyramids, pentagonal-prisms, spirals, discs, cuboids) but objects *cannot* be distinguished from one another based on one particular local feature: the only defining aspect of an object should be its overall configuration.
For example, if we have object A (a configuration of a cuboid, cylinder, and a pyramid) for each of its features there will be at least one other novel object that contains the identical feature (e.g., object B might have the identical cuboid, object C might have the identical pyramid, and so on…) - and thus the objects cannot be differentiated based on local features, and must be differentiated by overall configuration instead. So I’m looking for a stimuli set where features have been manipulated systematically such that objects can be distinguished only by their configuration of features (something corresponding to the linked table would be ideal):
Has such a stimuli set has been used in the past, and if so has it been made available? Any suggestions welcome.
Ryan
Relevant answer
Answer
Hi Ryan
Take a look at Gauthier T, Tarr MJ Becoming a Greeble Expert etc..., Vision Research 1997, 37, pp 1673-1681.
It may help.
Regards
Michael
  • asked a question related to Object Recognition
Question
10 answers
I want to know is any object in a part of picture or not. I do not need to know what that object is.
Relevant answer
Answer
tnx dear @Dibya jyoti Bora
  • asked a question related to Object Recognition
Question
10 answers
Could any one of you please suggest some technical papers/articles in the field of thermal image processing to start with?
Relevant answer
Answer
 This is an important question with a number of possible answer.   In addition to the very helpful answers already given, there is a bit more to add.
Thermal imaging and thermographic measurements are the focus of the following paper:
See, for example, Fig. 1 and the ensuing discussion for an overview of the approach needed to do edge detection on thermal images.
More to the point, see the use of histograms as a means of extracting information from thermal images in
  • asked a question related to Object Recognition
Question
7 answers
I need to apply a machine learning technique to categorize a data set in to a large number of classes (around 60). What would be the best machine learning technique to use? I just want to get an idea. 
Relevant answer
Answer
Hi Chamari,
Basically, there is no best machine learning algorithm, however, I recommend trying linear algorithms first, but if you found them not sufficient, you could give kernel methods and neural networks a try. 
HTH.
Samer
  • asked a question related to Object Recognition
Question
5 answers
Hi,
I am new to active shape models (ASM) and I want to use it in my research to do image segmentation.
For using ASM, there should be a training set to generate the shape statistical model: x = xbar+Pb, where xbaris the mean shape and P is the eigenvectors, x is the shapes obtained by changing the shape parameters b.
In my particular case, there is no training set. However, the mean shape is known, so is the shape constraints. Is there a way to use the statistical model? Use simulated shapes for training? But how to mimic the gray-level profile?
Thanks in advance.
Relevant answer
Answer
Just attach the image files with the "paper clip" below the message...
  • asked a question related to Object Recognition
Question
14 answers
I ran a novel object recognition test on mice that underwent traumatic brain injury (or sham) and were treated with a drug (or vehicle).  My sham controls (both treated and untreated) performed well and showed clear preference for the novel object (>60%), the TBI untreated group exhibited no preference for the novel object.  Curiously, the drug treated TBI group exhibited a pretty strong avoidance of the novel object, only actively investigating it about 25% of the time.  Does anyone with experience in this task have some insight into how to interpret this result?  Is it neophobia? Anxiety?
Relevant answer
Answer
Dear Kate, if you want to take into account the initial exploration of the objects as a baseline, you can utilize the recognition index (RI), as described in:
If the RI confirms the result obtained with the discrimination index (DI), then you might really have found a neophobic phenotype. You said you have counterbalaced for both the objects and the sides, but the best way to check for possible biases is to compare the RI and the DI. If you still find a significant aversion for the new object, then it would be interesting to investigate this phenotype with a different test. As Katherine suggested, you could use the novelty-suppressed feeding test, by which you can evaluate the hyponeophagia phenomenon (https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3197427/).
  • asked a question related to Object Recognition
Question
5 answers
I would like to know the process of recognition of shape and pattern of the object using a digi-cam based on image processing
Relevant answer
Answer
Go through this paper you can detect most of the object shapes using this approach
  • asked a question related to Object Recognition
Question
7 answers
I would like to know a good starting point to carry out my research in the above mentioned topic.
Relevant answer
Answer
Hello Aditya,
What are the Objects you want to recognise ? Are they same object? and more in Number?
Example : 1st Case : You want to Detect Multiple objects of same class .
You want to detect Multiple Oranges in a Picture.
2nd Case : Multi Class Object detection
You want to detect Oranges, Apples, Banana's . 
If you are looking for the first case, it should not be that hard to implement, Apply the object detector multiple times on the images, instead returning when you find the first object.
If its the second case, try building feature detectors for every class and apply every detector onto the image to find the different objects.
In general there are many approaches to recognise the object. Simple features like Color, Shape, or more complex features using HaarWavelets.
Start with one object and move to multiple object recognition
  • asked a question related to Object Recognition
Question
2 answers
hi, i'm working on The biologically inspired hierarchical model for object recognition, Hierarchical Model and X (HMAX), and i want to know that how many images i should use in training stage to extract patchs?
only thing i see in related works is that they just mentioned to number of patches and they didn't say any thing about number of images.
Relevant answer
Answer
It may  5-10
  • asked a question related to Object Recognition
Question
7 answers
Say I have a set of objects to teach CNN but when they are appear in different angles the network doesnt recognise them. I can teach each CNN per angle but it looks like a weak solution. Is there any existing experience to solve recognition issue for 360 degrees of same object?
Relevant answer
Answer
This is an interesting problem.
  • asked a question related to Object Recognition
Question
1 answer
Esp. for Face recognition and face expression recognition
Relevant answer
Answer
The paper "Curvelet Based Feature Extraction" might give some insight.
  • asked a question related to Object Recognition
Question
4 answers
I have an arm robot. An object coordinates will captured by the camera and need to be mapped to the robot to implement the IK (inverse kinematic) algorithm and then robot has to move to a location defined in the camera image displayed by the supervising computer.
I'm wondering, what kind of a vision system should I apply to capture object coordinates? .and to measure surface defects to characterize surface roughness in polishing task?
Relevant answer
Answer
You can extract texture features and then run some supervised classification experiments to determine roughness. You can do this with CVIPtools and run experiments with CVIP-FEPC. CVIPtools software is available here: http://cviptools.ece.siue.edu/
  • asked a question related to Object Recognition
Question
14 answers
Hi all,
I am using slidning windows technique for object detection. But this techniques is very slow. So i am searching for an alternative method for object detection.
Is there any alternative method for object detection ?
Relevant answer
Answer
Hi, 
Can you please elaborate on your context a little? What is the scene you're looking at? One person or multiple persons, static or dynamic background? Identify a particular person or detect people in general?
Simplest case of a static background and detecting a moving person in a video- 
The first possibility is a static frame looking at a particular scene with change only being a person moving. Difference of frames (absolute and thresholded) shall give you the location of a particular blob (where the moving person. You can then use OpenCV contours detecting for a maximum value (use more than one for multiple moving objects) which shall localise the moving person for you. Draw a bounding box and extract the object. For identifying, you can apply your feature based techniques to do that. For example, face-detection and so on. 
More clarity on context can help me come up with a better approach maybe. I hope I have understood the question correctly. 
  • asked a question related to Object Recognition
Question
7 answers
i want detecting, counting and measuring of plant stomata in microscopic image by a computer software. Can I do it by image processing at all? How can I do it? Is it possible programmed by vb.net library or need to use other programming languages or even matlab toolboxes? I attached an example image. So thanks
Relevant answer
Answer
I think the best way is to train a classifier to detect the stomata.
Crop from several images the example of stomata and areas with out stomata.
Use some image descriptor like Histogram of Gradients (Hog) to describe each one the patches.
Train a classifier, SVM for example, to distinguish between stomata and non stomata.
Use a sliding window and the classifier to detect the areas with the stomata.
For now if you don't have enough images  you can use the following code as a staring point. It uses one of the stomata in the image as an example in order to do correlation base detection.
Pay attention that the right low stomata is not detected because it is not a full stomata. You can change the patch to be just the upper side and also detect the right low stomata but then you would have to change some basic staff in the final part of the code.
Hope it is a good base for you good luck.
  • asked a question related to Object Recognition
Question
9 answers
hi all
I have some images. I should convert my images to binary. For all images the histogram has a distribution like the attachment.
I think that if we have a bimodal histogram then choosing a thereshold is easy (there are some methods for this. For example: I know that otsu's method is good for bimodal histogram).
Now with 3 peaks in all histograms, how can I convert my images to binary? (consider to the peak that Corresponding to  zero)
Relevant answer
Answer
Dear Arefeh...
The proper threshold should be the valley point in the histogram... You can calulate it as the point whose value is less than the average of nearest n histogram values from the left side, and also less than the average of of n histogram values lay at the right side... You can set n=5 or 6.... Regards
  • asked a question related to Object Recognition
Question
5 answers
My main idea is to detect an object from a cluttered scene. At first, I capture the image of the object alone. Next, I capture the image of a cluttered scene in which the object is present. Th object must be detected from the cluttered scene. I am taking the pictures using flash only.
Relevant answer
Answer
your work is more aligned towards image recolonization rather then detection. you can use many features of the images containing object there is not objection on the use of feature(mean , entropy, standard deviation, correlation etc are some of the commonly   used features   ) more the number of features more will be the accuracy you can , do the add on like use of wavelets curvelets along with the features to make it more robust.
  • asked a question related to Object Recognition
Question
6 answers
I want to measure major and minor axes on each hole in image that i attached. I've done pre-processing methods on image,finally i have this image that i want to measure axis,how can i do that?
Relevant answer
Answer
This is a very interesting question.
You might want to consider the location of the center of mass of a hole as a means of measuring the major and minor axes of a hole.    Another approach is suggested in the following: find the major and minor axes of the best fit elipse for a hole:
  • asked a question related to Object Recognition
Question
4 answers
I am doing project for 3D model search using shape matching.  For this I have generated 4 texture-less views from 3D model (Front, Top, Side and Isometric) with hidden lines. I have used SIFT algorithm to match these diagram with the one that is provided by an user. But SIFT is mainly for textured object detection so it is not generating up to the mark result. Also there is AKAZE but it also uses textured object detection.
Can anyone suggest me any shape matching algorithm with scale and rotation invariant.
Relevant answer
Answer
This problem is a little bit harder than invariant-based shape matching since you will have occlusions from the perspective projection (if I understand the problem correctly, you are trying to match views of a 3d object to a template). If the observed view does not overlap with the view used to construct the template then the problem is impossible. However, if the view has a high amount of overlap then what you need is a perspective invariant, which would fall within the realm of affine invariant.
You could use our method (attached) if the views are provided as point-sets. You could sample edges or use some other feature point selector to produce the point-sets. It is a little bit sensitive to outliers.
If you try our method and find it has trouble with outliers produce by your feature point selection routine then try Shape Contexts along with VFC. I believe that Ji Zhao has code somewhere. There are literally dozens of point-matching algorithms you can try once you get the data into this format.
If you don't want to do feature-points but can reduce the views to silhouettes then you could try the dijkstra-based method on the cross product graph of the two outlines (ok you have to reduce the silhouettes to curves this time). You would want to use an affine invariant signature in this setup.
Good luck, hope this helps 
  • asked a question related to Object Recognition
Question
2 answers
Hi
I'm trying to find an algorithm for detecting fire in a video.
which method for dynamic texture detection is the best for this purpose?
thanks in advance...
Relevant answer
Answer
  • asked a question related to Object Recognition
Question
27 answers
I'm working on hand-recognition project using Matlab, I'm trying to find convexity defects to define the fingers roots, actually, I have got Convex-hull points ( Convex-contour ) as shown in figure below (blue line), but I don't know how to find convexity defects, convexity defects shown as yellow points in the next figure.
Relevant answer
Answer
You could calculate the angle between each successive line-segment pair. Depending on how the line-segments run (clockwise or anti-clockwise), angles greater than (or less than)180 degrees are concavities. However, you then get all concavities...
In this case, following heuristic might also work: evaluate all points that are between two successive convex-hull points. The point that has the longest perpendicular distance to the line connecting the two convex-hull points, probably is the point you are looking for. 
  • asked a question related to Object Recognition
Question
4 answers
Hi all. I am a postgraduate student in electrical engineering and I would like to hear your opinion on automatic recognition of fasteners. We have heard about plant identification and face recognition but fastener recognition is rarely discussed. From my research, there are thousands of unique fasteners and to identify one fastener with another requires knowledge about the pitch diameter, head type, etc. There is also gauges invented to identify fastener but how about a system which require only a camera and a computer?
Would you like to have an automatic fastener identification system using a camera?
Do you think this system is important in your daily life or in the manufacturing/maintenance industry?
And lastly do you have problem in identifying fastener?
Relevant answer
Answer
salam
have a good day. can you take a look to this code program https://github.com/rishisij/workhouse
  • asked a question related to Object Recognition
Question
3 answers
I am looking Video dataset to make studies on the field of Video processing?
Relevant answer
  • asked a question related to Object Recognition
Question
8 answers
I] Part-I (Orientation Assignment to Keypoint)
In this process
1. First I have selected window of size 16 x 16 around keypoint and calculated magnitude and orientation for each point in window of 16 x 16.
3. Then created a 36 bin histogram of orientation.
4. then I have assigned the mean value of highest bin.(i.e. if 1st bin(0-10) has highest bin of 36 then '5' is assigned as orientation to keypoint.(Is it Correct?))
5. Then I have calculated Gaussian window of size 16 x 16 with sigma value equal to 1.5 times of scale.
6. Then I have multiplied magnitude matrix of size 16 x 16 with Gaussian window
(What is the use of this multiplication?)
Is it require to multiply this multiplication result(Magnitude x Gaussian) with orientation before assigning orientation to keypoint ? (as i found some histogram bins with highest value has less magnitude value.)
As per my logic we should assign the orientation mean to keypoint as orientation of the bin whose value is highest with its magnitude value.
7. Then I have transformed(rotated) coordinates of key point i.e. x,y position of key point with respect to assigned orientation by using 2D transformation. (Is it Correct?)
8. then I have transformed orientation of all sample points included in window of 16 x 16 according to orientation of keypoint.(e.g. if keypoint orientation =5 and if sample point orientation =270 the it will become 275.(Is it Correct ?))
Relevant answer
Answer
Hi,
I am trying to shed some light here:
1. what you get from SIFT is an 128-dim. descriptor for each detected keypoint position, orientation and scale.
2. In the matching step, only the 128-dim descriptor is involved. Please note that the descriptor is already normalized to orientation and scale. The detected position gives you the center point of the image patch that you describe with the 128 dim. descriptor. The scale tells you how large this patch is and the orientation tells how much the patch has to be rotated before descriptor computation.
3. This depends on the application... if you just want to match points between images you compare the 128.dim descriptors and just look at the keypoint position, when corresponding descriptors have been detected.
4. see 2. and 3.
5. Yes, if you rotate an image by 90 degree you should get the same number of keypoints.
6. This is optional... if you implement the standard matching described in Lowe's paper then you have a one-to-many matching.
  • asked a question related to Object Recognition
Question
3 answers
Boundary of the image is generate. from the boundary image i want to generate the datum points between the (index-middle) finger and (ring-little) finger.  
  • asked a question related to Object Recognition
Question
3 answers
Do the Flying saucers contain giant wheels as a century fuse (e.g. to create artificial gravity – by spinning at high-speed)?
Is the shape of a saucer (i.e. shape of magnifying or convex lens) is ideal shape for deflecting space debris (e.g. to minimize damage)?
If mankind wish to travel to nearby planets such as Mars, don’t we need to study the reasons or possible advantages for saucer shape?
I am not saying, aliens travelled to Earth. But we all know that the most popular shape for the UFO is Flying saucers.
I like to know pros and cons and thoughts who have done more investigation. I am just a curious bystander. I saw a small bit on returning of US astronaut after nearly spending one year in the space. Also the news mentioned that, it would take about 1 year just to reach Mars.
This is the weekend, so wish to explore something fun and interesting. If UFO contains a gain-wheel/centrifuge. How many hours a day should we need to run the gain-wheel/centrifuge to maintain healthy bone mass density?
Of course, it is possible to run the gain-wheel/centrifuge at different speeds in order to exert different weight (e.g. ranging from 0.5G or 1.5G). Of course, such power consumption can be meat by a mini nuclear power plant. I am sure, such advanced civilizations could have developed such mini nuclear power plant.
Best Regards,
Raju Chiluvuri
Relevant answer
Answer
oups! this is far away from competencies, sorry.
  • asked a question related to Object Recognition
Question
6 answers
I am doing the work for finger recognition. For each image containing hand object, I have the human labeled ground truth and calculated result by algorithms of each fingertip coordinate. Right now I would like to calculate error between the result by algorithm and ground truth.
Before the error calculating, I believe I need to match fingertips to certain pairs. My intuitive method is to generate fingertip blocks firstly, and for each recognized fingertip block by the algorithm, use SSIM to find the nearest block in labeled data.
Could you give me more suggestions for the corresponding fingertips matching procedure?
Thank you so much for your great help! 
  • asked a question related to Object Recognition
Question
7 answers
I have mixed some detection and tracking algorithms to do multiple pedestrian tracking. At this stage I have my results and I can visually see the tracked pedestrians. However, I don't know how to evaluate my results to show how good my method works. Do you have any suggestions?
Thank you.
Relevant answer
Answer
Hi,
Graham pointed out the fact you must analyze what kind of evaluation is required.
In some MTT methods, if track labels or theoretical positions of pedestrians are known, you can use the OSPA-T distance. It can both combine the tracking accuracy and the cardinality estimation. Full details of this metric can be found in Ristic's paper :
  • asked a question related to Object Recognition
Question
4 answers
I am trying to detect 3d model from live video stream.Model should be detected by any face.how can  do that?
Relevant answer
Answer
Hi Patil,
If you have a CAD model of your target you may extract the edges and find those edge in image.
Not so common, you may also use texture if your CAD model contains texture information.
I used a commercial software to find the pose of a CAD model in images, but if you pretend to get more data and fully control the process is may be better to implement the methods by yourself
  • asked a question related to Object Recognition
Question
11 answers
I'm a student with electrical/mechanical background, in my project I'm searching for a solution for a company who wants to start with 3D cameras for robotics.
At the moment I'm working with Matlab and it works great, the possibility to create your own GUI is a big plus.
But I read Matlab is more for developing purpose and is slower (overhead).
A second software package that I try to use is Halcon, at the moment I've no overview of the possibilities.
But it looks to me that you can program in Halcon's own language hdevelop or using their libraries in your own code (like C++).
Programming in hdevelop with it's GUI seems to be easier/faster than low-level programming (e.g. C++), but I don't know the limitations.
A disadvantage is that there is no community for support, you need to use their documentation.
A third option I read a lot about is OpenCV, but with no low-level programming background this seems too ambitious for me.
I'm not searching the best solution for me, but for a company (although I know the company hasn't a lot of computer engineers).
I was hoping to find software with a good GUI to reduce low-level programming, Halcon seems to be the closest match.
Thanks for your help.
Relevant answer
Answer
Hi Mat,
I use Halcon, it's a very powerful tool mainly for industrial purposes. For investigation it may be used in processes that aren't your focus because some functions are like a black box (its their knowledge and marketing advantage).
There is a group on LinkedIn about Halcon with experienced users that gives faster answers than Halcon support.
And, YES you develop all your code and export it to other languages or use hdevengine in which any modification require just to replace an Halcon file and not to compile again all app
  • asked a question related to Object Recognition
Question
4 answers
We run a small scoring shop for university exams. We use a optical recognition scanner and scan sheets through the scanner to score instructor designed exams.
We have been asked to begin scoring  multiple answer exams. Our current optical recognition software is good at scoring items with only one correct answer. However we are now being asked to score tests where students should indicate all items which are true, up to three correct options for one item. 
Do any of you have a good system for tabulating the correct answer in this type of assessment? Thanks for your help.
Relevant answer
Answer
Hi Laura
You may treat this problem as an ordinal variable. If in your example, the student marks a and b, he gets the maximum score (eg. 3) . If he marks c he obtain the minimum score (eg. 0), if he marks only a or b he obtains a score in the midle (eg. 2). If he marks b and c, or a and c, he gets a regular score (eg. 1). Then you may use an IRT method for an ordinal response or for a mix of binary and ordinal responses. Also it may be that to mark only a was better than to mark only b, then you may give different ordinal value to this selections. For example you may put a value +2 for a, +1 for b, -2 for c and your ordinal variable may be the sum of this values. You should also take account of the lack of response.
  • asked a question related to Object Recognition
Question
6 answers
Is there any difference in performing the novel object recognition test in open arena or in Y-Maze or in multiple chambers? Do they have different purpose? 
Relevant answer
Answer
My preference is to perform Novel Object Recognition in an open area, rather than Y-maze, separate compartments etc. This will mean the objects are continually visible to the animal and hence when they tend to the object it is an active choice of one over the other, rather than simply exploring the object in closest proximity. The other risk with more compartmentalised arenas, or more distance between the objects, is that a spatial aspect to the task is included, rather than purely visual recognition memory.
  • asked a question related to Object Recognition
Question
13 answers
What are we supposed to do during recognition process in order to sparse representing? Actually sparse representation should be done in which part of the process? skin detection, feature extraction or classification of gestures? Is there any projects that have been done with similar subject, to give me a view about how I could go through it?
Relevant answer
Answer
Dear Giovanni
Do you have any related practical example or project, that has used sparse representation?
  • asked a question related to Object Recognition
Question
4 answers
Actually, I have thought to implement HOG feature with temporal context for video data. Dalal and Triggs HOG is for 2 Dimension image. I want to implement it for video sequences as a feature for human action recognition. Where, you have to find gradient in three directions (x,y and t) and follow the procedure of traditional HOG with some nominal changes. So, is anyone has already used this technique? Is this concept is worth as efficient feature?
Relevant answer
Answer
Thank you Nada. I appreciate your help.
  • asked a question related to Object Recognition
Question
6 answers
As in case of object recognition, different work is done in this field with different test objects, how can I compare the performance of my work with any existing method?
Relevant answer
Answer
Dear Zubair Please find the following article which may help you.
  • asked a question related to Object Recognition
Question
3 answers
Ask for advice: Comparing the delay or retrieval activity in object color-based working memory with corresponding activity in object location-based WM.
I would be grateful if someone could give me some advices and paper.
Relevant answer
Answer
Maybe the attached publication by
Ma et al. (2014). Changing concepts of working memory. nature neuroscience, 17(3): 347-356
could give you some orientation?
Regards,
Klaus Blischke
  • asked a question related to Object Recognition
Question
3 answers
I have a skeletal model given by Kinect and now I want to label body parts using it. I know that for finding joint coordinates Kinect algorithm does this in its intermediate steps http://research.microsoft.com/pubs/145347/BodyPartRecognition.pdf
but is there a way to access that information or can you suggest me some other method/code to label body parts.
Relevant answer
Answer
Man it has been a while, I had completely forgotten about the paper that you linked to in your question.
Okay, here is my understanding of what the Kinect device is doing based upon the patent literature, and I'd love to hear from someone else if they have info. that contradicts this.
First, there is a huge difference between what people do during the R&D phase and what they do during the production phase. The reason is that during R&D you want to know as much as possible while during production you want to meet your clients demands as well as possible.  For most of the "Kinect" clients, all they want to know is where is the skeleton, and they want to know this at real time speeds.
I say this because from the patent literature it sounds as if this point-to-bone assignment is only transient inside the Kinect device.  Rather, the Kinect device focuses on what this point-to-bone information tells it about where the joints should be, and only reports that information.
Similarly, from what I have seen with respect to the Kinect SDK, it appears that the only information that is streamed from the Kinect device is {color, depth, skeletal pose}, not {color, depth, pixel assignment, skeletal pose}.
If you think about the size of the pixel assignment information there is a good reason for not reporting it.  That is, the skeletal pose information is basically non-existent with respect to the size of the color and depth information.  Since the pixel assignment information would require at least as much time/memory as the color and depth information this means including the pixel assignment information in the Kinect camera data stream would increase the amount of time required for Kinect to report a single frame by rough 50%.  Since Kinect is already operating at the abysmal frame rate ~30 fps, the final frame rate with pixel assignment would put them somewhere around 15 fps.  I have tracking libraries that I have written that can do as much as Kinect can with any depth data stream that will get you the pixel assignment in about 30 fps without optimization through parallelization. If Kinect was forced to operate at 15 fps no one would buy the device.  So, I don't think that you are going to get the per-pixel information from Kinect.
All that said, unless you want to drop Kinect and write your own tracking algorithm, I would just take the skeletal pose and depth data returned by Kinect and use Linear blend skinning to assign pixel data to skeletal pose.  If you don't want to mess with skinning and you don't want to have to write your own skeletal tracker then you could just license my companies libraries for cheap, but you would still have to set up all of real time image buffers and concurrency infrastructure yourself in order to get real time processing.  So, I would just use LBS.
  • asked a question related to Object Recognition
Question
6 answers
I have performed behavioral tests of open field, elevated plus maze and novel bject recognition in my test mice in CD1 background. Interestingly, these mice didn't show any change in Elevated plus maze test. However in open field they traversed less in the central area of the open field. This was significant. The same difference was there in C57 background also. In novel object the test mice stayed more in proximity of the older object as compared to the novel object. I interpret it as mild anxiety which may be driven by novelty associated fear. However there still remains the question of the mice not showing any changes as compared to control mice in Elevated plus maze. Can someone help me alternate interpretation?An alternate explanation would be immensely helpful. I haven't performed EPM on c57 background. Only open field has been carried out. 
Relevant answer
Answer
The open field test (OFT) is not quite a good test to evaluate anxiety, I would be careful with this kind of interpretation. OFT is more useful for locomotor measurements than it is to anxiety. The novel object recognition is better fitted for memory evaluation, so the interpretation about anxiety in this test should be also done with good care, maybe the novelty suppressed feeding would be better for anxiety measurements. Elevated plus-maze is a considerably good test and widely used for anxiety behavior, so if you do not see any difference at this test it indicates better that you drug or treatment/manipulation does not have a strong effect on anxiety behavior control. 
Therefore, you can consider run another behavior test which is better to evaluate anxiety-like behavior such as novelty suppressed feeding or the radial maze to have a better profile of your animals behavior. As your results do not show an effect with the EPM it might be better to run more tests to confirm the effects you saw with OFT and NOR. 
  • asked a question related to Object Recognition
Question
7 answers
Hello forum,
I have been reading about cascaded classifiers using haar-features for face detection and I have a few simple questions I have to ask/clarify. This is more towards implementation as I am a little confused as to how they work.
1) I understand that during the training phase, the haar features will be evaluated and rescaled for all possible combinations. At the end, the feature with the smallest error will form the first stage (attached picture). My question is, during the detection phase when a sub-window is selected for evaluation, will the features be placed at a specific region (like in the attached picture again) ?
For example, for the top left feature, it must always be positioned in the center leaving an empty space of 10% (of the width) to the left and right and be 30% (of the height) below.
Or will evaluation start at the top left hand corner (assuming origin), similar to training ? i.e. the feature will be evaluated over all the regions in the subwindow.
2) Regarding adaboost, I have understood the steps but my question is, when the weights are updated after the nth iteration, is it possible that a feature that has been already selected, get selected again ? i.e. it has the smallest error again. Or will features/classifiers that have already been selected be "removed" from the subsequent selection process ?
I am really loving computer vision. I will be undergoing this module in 10 weeks when semester starts but, I can't wait for so long to officially start learning what I love haha. Thanks all.
Relevant answer
Answer
 For the question 1, I think what you should keep in mind is that, classical Haar filters are sensitive to position changes. So, for the evaluated windows, those Haar filters selected in training stages should be placed in the same positions with that in training images. However, because the sliding windows tend to have different scales from training images, so you need to 'align' the sliding window with training scale via some techniques before you place those Haar filters.
  • asked a question related to Object Recognition
Question
4 answers
I've implement sift
1. make Gaussian blurring in every octaves and find the DoGs.
2. find local extrema.
but at this step I am confuse what should I do to find the key point with those 6 extremas?
can some one explain the formula of keypoint localization D(x) ?
Relevant answer
Answer
and how about those formulas?
what is the formulas mean? what is is the correlation of each parts in notation with all of Dog that we count before?
  • asked a question related to Object Recognition
Question
10 answers
I need to know about object recognition steps can any one help me, with an example.
Relevant answer
Answer
example: Detect pedestrians.
You need to do two main blocks:
 1. Training
 2. Detection