Olga Veksler’s research while affiliated with University of Waterloo and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (7)


Sparse Non-Local CRF With Applications
  • Article

October 2024

·

3 Reads

IEEE Transactions on Pattern Analysis and Machine Intelligence

Olga Veksler

·

Yuri Boykov

CRFs model spatial coherence in classical and deep learning computer vision. The most common CRF is called pairwise, as it connects pixel pairs. There are two types of pairwise CRF: sparse and dense. A sparse CRF connects the nearby pixels, leading to a linear number of connections in the image size. A dense CRF connects all pixel pairs, leading to a quadratic number of connections. While dense CRF is a more general model, it is much less efficient than sparse CRF. In fact, only Gaussian edge dense CRF is used in practice, and even then with approximations. We propose a new pairwise CRF, which we call sparse non-local CRF. Like dense CRF, it has non-local connections, and, therefore, it is more general than sparse CRF. Like sparse CRF, the number of connections is linear, and, therefore, our model is efficient. Besides efficiency, another advantage is that our edge weights are unrestricted. We show that our sparse non-local CRF models properties similar to that of Gaussian dense CRF. We also discuss connections to other CRF models. We demonstrate the usefulness of our model on classical and deep learning applications, for two and multiple labels.


Regularized Loss With Hyperparameter Estimation for Weakly Supervised Single Class Segmentation

May 2024

·

5 Reads

IEEE Transactions on Pattern Analysis and Machine Intelligence

We propose a new image level weakly supervised segmentation approach for datasets with a single object class of interest. Our approach is based on a regularized loss function inspired by the classical Conditional Random Field (CRF) modeling. Our loss models properties of generic objects, and we use it to guide CNN towards segments that are more likely to correspond to the object, thus avoiding the need for pixel precise annotations. Training CNN with regularized loss is a difficult task for gradient descent. We develop an annealing algorithm which is crucial for a successful training. Furthermore, we develop an approach for hyperparameter setting for the most important components of our regularized loss. This is far from trivial, since there is no pixel precise ground truth for guidance. The advantage of our method is that we use a standard CNN architecture and an easy to interpret loss function, derived from classical CRF models. Furthermore, we apply the same loss function for any task/dataset. We first evaluate our approach for salient object segmentation and co-segmentation. These tasks naturally involve one object class of interest. Then we adapt our approach to image level weakly supervised multi-class semantic segmentation. We obtain state-of-the-art results.




Regularized Loss for Weakly Supervised Single Class Semantic Segmentation

October 2020

·

12 Reads

·

12 Citations

Lecture Notes in Computer Science

Fully supervised semantic segmentation is highly successful, but obtaining dense ground truth is expensive. Thus there is an increasing interest in weakly supervised approaches. We propose a new weakly supervised method for training CNNs to segment an object of a single class of interest. Instead of ground truth, we guide training with a regularized loss function. Regularized loss models prior knowledge about the likely object shape properties and thus guides segmentation towards the more plausible shapes. Training CNNs with regularized loss is difficult. We develop an annealing strategy that is crucial for successful training. The advantage of our method is simplicity: we use standard CNN architectures and intuitive and computationally efficient loss function. Furthermore, we apply the same loss function for any task/dataset, without any tailoring. We first evaluate our approach for salient object segmentation and co-segmentation. These tasks naturally involve one object class of interest. In some cases, our results are only a few points of standard performance measure behind those obtained training the same CNN with full supervision, and state-of-the art results in weakly supervised setting. Then we adapt our approach to weakly supervised multi-class semantic segmentation and obtain state-of-the-art results.


Efficient Graph Cut Optimization for Full CRFs with Quantized Edges

March 2019

·

14 Reads

·

13 Citations

IEEE Transactions on Pattern Analysis and Machine Intelligence

Fully connected pairwise Conditional Random Fields (Full-CRF) with Gaussian edge weights can achieve superior results compared to sparsely connected CRFs. However, traditional methods for Full-CRFs are too expensive. Previous work develops efficient approximate optimization based on mean field inference, which is a local optimization method and can be far from the optimum. We propose efficient and effective optimization based on graph cuts for Full-CRFs with quantized edge weights. To quantize edge weights, we partition the image into superpixels and assume that the weight of an edge between any two pixels depends only on the superpixels these pixels belong to. Our quantized edge CRF is an approximation to the Gaussian edge CRF, and gets closer to it as superpixel size decreases. Being an approximation, our model offers an intuition about the regularization properties of the Guassian edge Full-CRF. For efficient inference, we first consider the two-label case and develop an approximate method based on transforming the original problem into a smaller domain. Then we handle multi-label CRF by showing how to implement expansion moves. In both binary and multi-label cases, our solutions have significantly lower energy compared to that of mean field inference. We also show the effectiveness of our approach on semantic segmentation task.


Location Augmentation for CNN

July 2018

·

15 Reads

CNNs have made a tremendous impact on the field of computer vision in the last several years. The main component of any CNN architecture is the convolution operation, which is translation invariant by design. However, location in itself can be an important cue. For example, a salient object is more likely to be closer to the center of the image, the sky in the top part of an image, etc. To include the location cue for feature learning, we propose to augment the color image, the usual input to CNNs, with one or more channels that carry location information. We test two approaches for adding location information. In the first approach, we incorporate location directly, by including the row and column indexes as two additional channels to the input image. In the second approach, we add location less directly by adding distance transform from the center pixel as an additional channel to the input image. We perform experiments with both direct and indirect ways to encode location. We show the advantage of augmenting the standard color input with location related channels on the tasks of salient object segmentation, semantic segmentation, and scene parsing.

Citations (4)


... extra training data, instead opting for on-the-fly model adjustments to the model in response to the characteristic of test data with only a modest increase in computational overhead. Test-time adaptation has been proven effective in various domains such as image classification [26][27][28][29] , semantic segmentation [30][31][32][33] , and object detection [34][35][36][37] . Nevertheless, its application in predicting interatomic potentials remains unexplored. ...

Reference:

Online test-time adaptation for better generalization of interatomic potentials to out-of-distribution data
Test Time Adaptation with Regularized Loss for Weakly Supervised Salient Object Detection
  • Citing Conference Paper
  • June 2023

... We present two failure cases in Figure 8. In the left image, the predicted polygon fails to fit concave contours since the pairwise loss prefers the shorter length [73,6]. The right image shows that our model faces challenges distinguishing similar parts from different instances, as it is difficult to reason object ownership based on color alone. ...

Sparse Non-local CRF
  • Citing Conference Paper
  • June 2022

... The difficulty in segmentation data collection shapes two main families of solutions. One is weakly-supervised semantic segmentation [27], [37]- [39], [39]- [41] which utilizes image-level labels instead of the pixel-level annotations. The other one trains the segmentation model with synthetic data, of which ground-truth semantics can be obtained freely. ...

Regularized Loss for Weakly Supervised Single Class Semantic Segmentation
  • Citing Chapter
  • October 2020

Lecture Notes in Computer Science

... maximal cliques of size 2. In this case, we can write down the energy as in Eq. (3) [22] As the graph becomes denser the complexity of the algorithms to solve the energy minimization problem increases. Therefore, fully connected graphs can be efficiently addressed using mean-field algorithms, the meanfield approach estimate, and the distribution P (w│I) for the given image I with fully factorized distribution Q(x). ...

Efficient Graph Cut Optimization for Full CRFs with Quantized Edges
  • Citing Article
  • March 2019

IEEE Transactions on Pattern Analysis and Machine Intelligence