Abstract
Consensus maximization is one of the most widely used robust fitting paradigms in computer vision, and the development of algorithms for consensus maximization is an active research topic. In this paper, we propose an efficient deterministic optimization algorithm for consensus maximization. Given an initial solution, our method conducts a deterministic search that forcibly increases the consensus of the initial solution. We show how each iteration of the update can be formulated as an instance of biconvex programming, which we solve efficiently using a novel biconvex optimization algorithm. In contrast to our algorithm, previous consensus improvement techniques rely on random sampling or relaxations of the objective function, which reduce their ability to significantly improve the initial consensus. In fact, on challenging instances, the previous techniques may even return a worse off solution. Comprehensive experiments show that our algorithm can consistently and greatly improve the quality of the initial solution, without substantial cost. (Matlab demo program is available in the supplementary material)
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Due to the existence of noise and outliers in real-life data, robust model fitting is necessary to enable many computer vision applications. Arguably the most prevalent robust technique is random sample consensus (RANSAC) [11], which aims to find the model that has the largest consensus set. The RANSAC algorithm approximately solves this optimization problem, by repetitively sampling minimal subsets of the data, in the hope of “hitting” an all-inlier minimal subset that gives rise to a model hypothesis with high consensus.
Many variants of RANSAC have been proposed [7]. Most variants attempt to conduct guided sampling using various heuristics, so as to speed up the retrieval of all-inlier minimal subsets. Fundamentally, however, taking minimal subsets reduces the span of the data and produces biased model estimates [20, 27]. Thus, the best hypothesis found by RANSAC often has much lower consensus than the maximum achievable, especially on higher-dimensional problems. In reality, the RANSAC solution should only be taken as a rough initial estimate [9].
To “polish” a rough RANSAC solution, one can perform least squares (LS) on the consensus set of the RANSAC estimate (i.e. the Gold Standard Algorithm [12, Chap. 4]). Though justifiable from a maximum likelihood point of view, the efficacy of LS depends on having a sufficiently large consensus set to begin with.
A more useful approach is Locally Optimized RANSAC (LO-RANSAC) [9, 18], which attempts to enlarge the consensus set of an initial RANSAC estimate, by generating hypotheses from larger-than-minimal subsets of the consensus set.Footnote 1 The rationale is that hypotheses fitted on a larger number of inliers typically lead to better estimates with even higher support. Ultimately, however, LO-RANSAC is also a randomized algorithm. Although it conducts a more focused sampling, the algorithm cannot guarantee improvements to the initial estimate. As we will demonstrate in Sect. 5.2, often on more challenging datasets, LO-RANSAC is unable to significantly improve upon the RANSAC result.
Due to its combinatorial nature, consensus set maximization is NP-hard [4]. While this has not deterred the development of globally optimal algorithms [3, 5, 6, 10, 19, 21, 25, 30], the fundamental intractability of the problem means that global algorithms are essentially variants of exhaustive search-and-prune procedures, whose runtime scales exponentially in the general case. While global algorithms have their place in computer vision, currently they are mostly confined to problems with low-dimensions and/or small number of measurements.
1.1 Deterministic Algorithms—A New Class of Methods
Recently, efficient deterministic algorithms for consensus maximization are gaining attention [17, 22]. Different from random sampling, such algorithms begin with an initial solution (obtained using least squares or a random sampling method) and iteratively performs deterministic updates on the solution to improve its quality. While they do not strive for the global optimum, such algorithms are able to find excellent solutions due to the directed search.
To perform deterministic updating, the previous methods relax the objective function (Le et al. [17] use \(\ell _1\) penalization, and Purkait et al. [22] use a smooth surrogate function). Invariably this necessitates the setting of a smoothing parameter that controls the degree of relaxation, and the progressive tightening of the relaxation to ensure convergence to a good solution. As we will demonstrate in Sect. 5.4, incorrect settings of the smoothing parameter and/or its annealing rate may actually lead to a worse solution than the starting point.
1.2 Our Contributions
We propose a novel deterministic optimization algorithm for consensus maximization. The overall structure of our method is a bisection search to increase the consensus of the current solution. The key to the effectiveness of our method is to formulate the feasibility test in each iteration as a biconvex program, which we solve efficiently via a biconvex optimization algorithm. Unlike [17, 22], our method neither relaxes the objective function, nor requires tuning of smoothing parameters. On both synthetic and real datasets, we demonstrate the superior performance of our method over previous consensus improvement techniques.
2 Problem Definition
Given a set of N outlier contaminated measurements, consensus maximization aims to find the model \(\mathbf {x}\in D\) that is consistent with the largest data subset
where D is the domain of model parameters (more details later), and
counts the number of inliers (consensus) of \(\mathbf {x}\). Function \(r_i(\mathbf {x})\) gives the residual of the i-th measurement w.r.t. \(\mathbf {x}\), \(\epsilon \) is the inlier threshold and \(\mathbb {I}\) is the indicator function which returns 1 if the input statement is true and 0 otherwise.
Figure 1 illustrates the objective function \(\mathcal {I}(\mathbf {x})\). As can be appreciated from the inlier counting operations, \(\mathcal {I}(\mathbf {x})\) is a step function with uninformative gradients.
2.1 The Update Problem
Let \(\tilde{\mathbf {x}}\) be an initial solution to (1); we wish to improve \(\tilde{\mathbf {x}}\) to yield a better solution. We define this task formally as
where \(\delta > \mathcal {I}(\tilde{\mathbf {x}})\) is a target consensus value. See Fig. 1 for an illustration. For now, assume that \(\delta \) is given; later in Sect. 4 we will embed (3) in a broader algorithm to search over \(\delta \).
Also, although (3) does not demand that the revised solution be “close” to \(\tilde{\mathbf {x}}\), it is strategic to employ \(\tilde{\mathbf {x}}\) as a starting point to perform the update. In Sect. 3, we will propose such an algorithm that is able to efficiently solve (3).
2.2 Residual Functions and Solvable Models
Before embarking on a solution for (3), it is vital to first elaborate on the form of \(r_i(\mathbf {x})\) and the type of models that can be fitted by the proposed algorithm. Following previous works [5, 14, 15], we focus on residual functions of the form
where \(q_i(\mathbf {x})\) is convex quadratic and \(p_i(\mathbf {x})\) is linear. We also insist that \(p_i(\mathbf {x})\) positive. We call \(r_i(\mathbf {x})\) the quasiconvex geometric residual since it is quasiconvex [2, Sect. 3.4.1] in the domain
Note that D in the above form specifies a convex domain in \(\mathbb {R}^d\).
Many model fitting problems in computer vision have residuals of the type (4). For example, in multiple view triangulation where we aim to estimate the 3D point \(\mathbf {x}\in \mathbb {R}^3\) from multiple (possibly incorrect) 2D observations \(\{ \mathbf {u}_i \}^{N}_{i=1}\),
is the reprojection error in the i-th camera, where \(\bar{\mathbf {x}} = [\mathbf {x}^T~1]^T\),
is the i-th camera matrix with \(\mathbf {P}_i^{(1:2)}\) and \(\mathbf {P}_i^{(3)}\) respectively being the first-two rows and third row of \(\mathbf {P}\). Insisting that \(\mathbf {x}\) lies in the convex domain \(D = \{ \mathbf {x}\in \mathbb {R}^3 \mid \mathbf {P}^{(3)}_i\bar{\mathbf {x}} > 0, \forall i \}\) ensures that the estimated \(\mathbf {x}\) lies in front of all the cameras.
Other model fitting problems with quasiconvex geometric residuals include homography fitting, camera resectioning, and the known rotation problem; see [14] for details and other examples. However, note that fundamental matrix estimation is not a quasiconvex problem [14]; in Sect. 5, we will show how the proposed technique can be adapted to robustly estimate the fundamental matrix.
3 Solving the Update Problem
As the decision version of (1), the update problem (3) is NP-complete [4] and thus can only be approximately solved. In this section, we propose an algorithm that works well in practice, i.e., able to significantly improve \(\tilde{\mathbf {x}}\).
3.1 Reformulation as Continuous Optimization
With quasicovex geometric residuals (4), the inequality \(r_i(\mathbf {x}) \le \epsilon \) becomes
Since \(q_i(\mathbf {x})\) is convex and \(p_i(\mathbf {x})\) is linear, the constraint (8) specifies a convex region in D. Defining
and introducing for each \(r_i'(\mathbf {x})\) an indicator variable \(y_i\in [0,1]\) and a slack variable \(s_i \ge 0\), we can write (3) using complementarity constraints [13] as
Intuitively, \(y_i\) reflects whether the i-th datum is an inlier w.r.t. \(\mathbf {x}\). In the following, we establish the integrality of \(y_i\) and the equivalence between (10) and (3).
Lemma 1
Problems (10) and (3) are equivalent.
Proof
Observe that for any \(\mathbf {x}\),
-
a1: If \(r_i'(\mathbf {x}) > 0\), the i-th datum is outlying to \(\mathbf {x}\), and (10d) and (10e) will force \(s_i \ge r_i'(\mathbf {x})>0\) and \(y_i = 0\).
-
a2: If \(r_i'(\mathbf {x}) \le 0\), the i-th datum is inlying to \(\mathbf {x}\), and (10f) and (10d) allow \(s_i\) and \(y_i\) to have only one of the following settings: a2.1: \(s_i > 0\) and \(y_i = 0\); or a2.2: \(s_i = 0\) and \(y_i\) being indeterminate.
If \(\mathbf {x}\) is infeasible for (3), i.e., \(\mathcal {I}(\mathbf {x}) < \delta \), condition a1 ensures that (10b) is violated, hence \(\mathbf {x}\) is also infeasible for (10). Conversely, if \(\mathbf {x}\) is infeasible for (10), i.e., \(\sum _i{y_i} < \delta \), then \(\mathcal {I}(\mathbf {x}) < \delta \), hence \(\mathbf {x}\) is also infeasible for (3).
If \(\mathbf {x}\) is feasible for (3), we can always set \(y_i = 1\) and \(s_i = 0\) for all inliers to satisfy (10b), ensuring the feasibility of \(\mathbf {x}\) to (10). Conversely, if \(\mathbf {x}\) is feasible for (10), by a1 there are at least \(\delta \) inliers, thus \(\mathbf {x}\) is also feasible to (3). \(\square \)
From the computational standpoint, (10) is no easier to solve than (3). However, by constructing a cost function from the bilinear constraints (10d), we arrive at the following continuous optimization problem
where \(\mathbf {s}= \left[ s_1, \dots , s_N \right] ^T\) and \(\mathbf {y}= \left[ y_1, \dots , y_N \right] ^T\). The following lemma establishes the equivalence between (11) and (3).
Lemma 2
If the globally optimal value of (11) is zero, then there exists \(\mathbf {x}\) that satisfies the update problem (3).
Proof
Due to (11c) and (11e), the objective value of (11) is lower bounded by zero. Let \((\mathbf {x}^*, \mathbf {s}^*, \mathbf {y}^*)\) be a global minimizer of (11). If \(\sum _i y_i^*s_i^*= 0\), then \(\mathbf {x}^*\) satisfies all the constraints in (10) thus \(\mathbf {x}^*\) is feasible to (3). \(\square \)
3.2 Biconvex Optimization Algorithm
Although all the constraints in (11) are convex (including \(\mathbf {x}\in D\)), the objective function is not convex. Nonetheless, the primary value of formulation (11) is to enable the usage of convex solvers to approximately solve the update problem. Note also that (11) does not require any smoothing parameters.
To this end, observe that (11) is in fact an instance of biconvex programming [1]. If we fix \(\mathbf {x}\) and \(\mathbf {s}\), (11) reduces to the linear program (LP)
which can be solved in close form.Footnote 2 On the other hand, if we fix \(\mathbf {y}\), (11) reduces to the second order cone program (SOCP)
Note that \(s_i\) does not have influence if the corresponding \(y_i = 0\); these slack variables can be removed from the problem to speed up optimization.Footnote 3
The proposed algorithm (called Biconvex Optimization or BCO) is simple: we initialize \(\mathbf {x}\) as the starting \(\tilde{\mathbf {x}}\) from (3), and set the slacks as
Then, we alternate between solving the LP and SOCP until convergence. Since (11) is lower-bounded by zero, and each invocation of the LP and SOCP are guaranteed to reduce the cost, BCO will always converge to a local optimum \((\hat{\mathbf {x}}, \hat{\mathbf {s}}, \hat{\mathbf {y}})\).
In respect to solving the update problem (3), if the local optimum \((\hat{\mathbf {x}}, \hat{\mathbf {s}}, \hat{\mathbf {y}})\) turns out to be the global optimum (i.e., \(\sum _i \hat{y}_i \hat{s}_i = 0\)), then \(\hat{\mathbf {x}}\) is a solution to (3), i.e., \(\mathcal {I}(\hat{\mathbf {x}}) \ge \delta \). Else, \(\hat{\mathbf {x}}\) might still represent an improved solution over \(\tilde{\mathbf {x}}\). Compared to randomized search, our method is by design more capable of improving \(\tilde{\mathbf {x}}\). This is because optimizing (11) naturally reduces the residual of outliers that “should be” an inlier, i.e., with \(y_i = 1\), which may still lead to a local refinement, i.e., \(\mathcal {I}(\hat{\mathbf {x}})>\delta _l = \mathcal {I}(\tilde{\mathbf {x}})\), regardless of whether problem (3) is feasible or not. In the next section, we will construct an effective deterministic consensus maximization technique based on Algorithm 1.
4 Main Algorithm—Deterministic Consensus Maximization
Given an initial solution \(\mathbf {x}^{(0)}\) to (1), e.g., obtained using least squares or a random sampling heuristic, we wish to update \(\mathbf {x}^{(0)}\) to a better solution. The main structure of our proposed algorithm is simple: we conduct bisection over the consensus value to search for a better solution; see Algorithm 2.
A lower and upper bound \(\delta _l\) and \(\delta _h\) for the consensus, which are initialized respectively to \(\mathcal {I}(\mathbf {x}^{(0)})\) and N, are maintained and progressively tightened. Let \(\tilde{\mathbf {x}}\) be the current best solution (initialized to \(\mathbf {x}^{(0)}\)); then, the midpoint \(\delta = \lfloor 0.5(\delta _l + \delta _h) \rfloor \) is obtained and the update problem via the continuous biconvex formulation (11) is solved using Algorithm 1. If the solution \(\hat{\mathbf {x}}\) for (11) has a higher quality than the incumbent, \(\tilde{\mathbf {x}}\) is revised to become \(\hat{\mathbf {x}}\) and \(\delta _l\) is increased to \(\mathcal {I}(\hat{\mathbf {x}})\). And if \(\mathcal {I}(\hat{\mathbf {x}})<\delta \), \(\delta _h\) is decreased to \(\delta \). Algorithm 2 ends when \(\delta _h = \delta _l+1\).
Since the “feasibility test” in Algorithm 2 (Step 4) is solved via a non-convex subroutine, the bisection technique does not guarantee finding the global solution, i.e., the quality of the final solution may underestimate the maximum achievable quality. However, our technique is fundamentally advantageous compared to previous methods [9, 17, 22] since it is not subject to the vagaries of randomization or require tuning of hyperparameters. Empirical results in the next section will demonstrate the effectiveness of the proposed algorithm.
5 Results
We call the proposed algorithm IBCO (for iterative biconvex optimization). We compared IBCO against the following random sampling methods:
-
RANSAC (RS) [11] (baseline): the confidence \(\rho \) was set to 0.99 for computing the termination threshold.
-
PROSAC (PS) [8] and Guided MLESAC (GMS) [26] (RS variants with guided sampling): only tested for fundamental matrix and homography estimation since inlier priors like matching scores for correspondences were needed.
-
LO-RANSAC (LRS) [9]: subset size in inner sampling was set to half of the current consensus size, and the max number of inner iterations was set to 10.
-
Fixing LO-RANSAC (FLRS) [18]: subset size in inner sampling was set to \(7 \times \) minimal subset size, and the max number of inner iterations was set to 50.
-
USAC [23]: a modern technique that combines ideas from PS and LRS.Footnote 4 USAC was evaluated only on fundamental matrix and homography estimation since the available code only implements these models.
Except USAC which was implemented in C++, the other sampling methods were based on MATLAB [16]. Also, least squares was executed on the final consensus set to refine the results of all the random sampling methods.
In addition to the random sampling methods, we also compared IBCO against the following deterministic consensus maximization algorithms:
-
Exact Penalty (EP) method [17]: The methodFootnote 5 was retuned for best performance on our data: we set the penalty parameter \(\alpha \) to 1.5 for fundamental matrix estimation and 0.5 for all other problems. The annealing rate \(\kappa \) for the penalty parameter was set to 5 for linear regression and 2D homography estimation and 1.5 for triangulation and fundamental matrix estimation.
-
Smooth Surrogate (SS) method [22]: Using our own implementation. The smoothing parameter \(\gamma \) was set to 0.01 as suggested in [22].
For the deterministic methods, Table 1 lists the convex solvers used for their respective subproblems. Further, results for these methods with both FLRS and random initialization (\(\mathbf {x}^{(0)}\) was generated randomly) were provided, in order to show separately the performance with good (FLRS) and bad (random) initialization. We also tested least squares initialization, but under high outlier rates, its effectiveness was no better than random initialization. All experiments were executed on a laptop with Intel Core 2.60 GHz i7 CPU and 16GB RAM.
5.1 Robust Linear Regression on Synthetic Data
Data of size \(N = 1000\) for 8-dimensional linear regression (i.e., \(\mathbf {x}\in \mathbb {R}^8\)) were synthetically generated. In linear regression, the residual takes the form
which is a special case of (4) (set \(p_i(\mathbf {x}) = 1\)), and each datum is represented by {\(\mathbf {a}_i\in \mathbb {R}^8\), \(b_i \in \mathbb {R}\)}. First, the independent measurements \(\{ \mathbf {a}_i \}^{N}_{i=1}\) and parameter vector \(\mathbf {x}\) were randomly sampled. The dependent measurements were computed as \(b_i = \mathbf {a}_i^T \mathbf {x}\) and added with noise uniformly distributed between \([-0.3,0.3]\). A subset of \(\eta \%\) of the dependent measurements were then randomly selected and added with Gaussian noise of \(\sigma = 1.5\) to create outliers. To guarantee the outlier rate, each outlier is regenerated until the noise is not within [−0.3,0.3]. The inlier threshold \(\epsilon \) for (1) was set to 0.3.
Figure 2 shows the optimized consensus, runtime and model accuracy of the methods for \(\eta \in \{0, 5,...,70, 75\}\), averaged over 10 runs for each data instance. Note that the actual outlier rate was sometimes slightly lower than expected since the largest consensus set included some outliers with low noise value. For \(\eta = 75\) the actual outlier rate was around 72% (see Fig. 2(a)). To prevent inaccurate analysis caused by this phenomenon, results for \(\eta >75\) were not provided.
Figure 2(b) demonstrates for each method the relative consensus difference to RS. It is evident that both IBCO variants outperformed other methods in general. Unlike other methods, whose improvement to RS was low at high outlier rates, both IBCO variants were consistently better than RS by more than 11%. Though IBCO was only marginally better than EP for outlier rates lower than 65%, Fig. 2(a) shows that for most of the data instances, both IBCO variants found consensus very close or exactly equal to the maximum achievable. The cost of IBCO was fairly practical (less than 5 seconds for all data instances, see the data tip in Fig. 2(c)). Also the runtime of the random sampling methods (RS, LRS, FLRS) increased exponentially with \(\eta \). Hence, at high \(\eta \), the major cost of FLRS+EP, FLRS+SS and FLRS+IBCO came from FLRS.
To demonstrate the significance of having higher consensus, we further performed least squares fitting on the consensus set of each method. Given a least squares fitted model \(\mathbf {x}_{LS}\), define the average residual on ground truth inliers (the data assigned with less than 0.3 noise level) as:
where \(\mathcal {I}^*\) was the set of all ground truth inliers. Figure 2(d) shows \(e(\mathbf {x}_{LS})\) for all methods on all data instances. Generally, higher consensus led to a lower average residual, suggesting a more accurate model.
5.2 Homography Estimation
Five image pairs from the NYC Library dataset [29] were used for 2D homography estimation. On each image pair, SIFT correspondences were produced by the VLFeat toolbox [28] and used as inputs. Figure 3 depicts examples of inputs, as well as consensus sets from FLRS and FLRS+IBCO. The transfer error in one image [12, Sect. 4.2.2] was used as the distance measurement. The inlier threshold \(\epsilon \) was set to 4 pixels. The 4-Point algorithm [12, Sect. 4.7.1] was used in all random sampling approaches for model fitting on minimal samples.
Figure 4, shows the quantitative results, averaged over 50 runs. Though marginally costlier than SS and random approaches, both IBCO variants found considerably larger consensus sets than other methods for all data. Meanwhile, different from the linear regression case, EP no longer had similar result quality to IBCO. Also note that for challenging problems, e.g., Ceiling1 and Sign, the two IBCO variants were the only methods that returned much higher consensus than RS.
5.3 Triangulation
Five feature tracks from the NotreDame dataset [24] were selected for triangulation, i.e., estimating the 3D coordinates. The input from each feature track contained a set of camera matrices and the corresponding 2D feature coordinates. The re-projection error was used as the distance measurement [15] and the inlier threshold \(\epsilon \) was set to 1 pixel. The size of minimal samples was 2 (views) for all RANSAC variants. The results are demonstrated in Fig. 5. For triangulation, the quality of the initial solution largely affected the performance of EP, SS and IBCO. Initialized with FLRS, IBCO managed to find much larger consensus sets than all other methods.
5.4 Effectiveness of Refinement
Though all deterministic methods were provided with reliable initial FLRS solutions, IBCO was the only one that effectively refined all FLRS results. EP and SS sometimes even converged to worse than initial solutions. To illustrate these effects, Fig. 6 shows the solution quality during the iterations of the three deterministic methods (initialized by FLRS) on Ceiling1 for homography estimation and Point 16 for triangulation. In contrast to EP and SS which progressively made the initial solution worse, IBCO steadily improved the initial solution.
It may be possible to rectify the behaviour of EP and SS by choosing more appropriate smoothing parameters and/or their annealing rates. However, the need for data-dependent tuning makes EP and SS less attractive than IBCO.
5.5 Fundamental Matrix Estimation
Image pairs from the two-view geometry corpus of CMPFootnote 6 were used for fundamental matrix estimation. As in homography estimation, SIFT correspondences were used as the input data. Since the Sampson error [12, Sect. 11.4.3] and the reprojection error [12, Sect. 11.4.1] for fundamental matrix estimation are not linear or quasiconvex, the deterministic algorithms (EP, SS, IBCO) cannot be directly applied. Thus, we linearize the epipolar constraint and use the algebraic error [12, Sect. 11.3] as the residual. The inlier threshold \(\epsilon \) was set to 0.006 for all data.
Further, a valid fundamental matrix satisfies the rank-2 constraint [12, Sect. 11.1.1], which is non-convex. For EP, SS, IBCO, we impose the rank-2 constraint using SVD after each parameter vector updates (for IBCO, after each BCO run).
Figure 7 depicts sample image pairs and generated SIFT correspondences, as well as consensus sets from FLRS and FLRS+IBCO. The seven-point method [12, Sect. 11.1.2] was used in USAC and the normalized 8-point algorithm [12, Sect. 11.2] was used in all other RANSAC variants.
As shown in Fig. 8(a), unlike EP and SS who failed to refine the initial FLRS results for all the tested data, IBCO was still effective even though the problem contains non-convex constraints.
6 Conclusions
We proposed a novel deterministic algorithm for consensus maximization with non-linear residuals. The basis of our method lies in reformulating the decision version of consensus maximization into an instance of biconvex programming, which enables the use of bisection for efficient guided search. Compared to other deterministic methods, our method does not relax the objective of consensus maximization problem and is free from the tuning of smoothing parameters, which makes it much more effective at refining the initial solution. Experiments show that our method is able to greatly improve upon initial results from widely used random sampling heuristics.
Notes
- 1.
This is typically invoked from within a main RANSAC routine.
- 2.
Set \(y_i = 1\) if \(s_i\) is one of the \(\delta \)-smallest slacks, and \(y_i = 0\) otherwise.
- 3.
Given the optimal \(\hat{\mathbf {x}}\) for (13), the values of the slack variables that did not participate in the problem can be obtained as \(s_i = \max \{0,r_i'(\hat{\mathbf {x}})\}\).
- 4.
- 5.
Code from https://cs.adelaide.edu.au/~huu/.
- 6.
References
Biconvex optimization. https://en.wikipedia.org/wiki/Biconvex_optimization
Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)
Campbell, D., Petersson, L., Kneip, L., Li, H.: Globally-optimal inlier set maximisation for simultaneous camera pose and feature correspondence. arXiv preprint arXiv:1709.09384 (2017)
Chin, T.J., Cai, Z., Neumann, F.: Robust fitting in computer vision: easy or hard? arXiv preprint arXiv:1802.06464 (2018)
Chin, T.J., Heng Kee, Y., Eriksson, A., Neumann, F.: Guaranteed outlier removal with mixed integer linear programs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5858–5866 (2016)
Chin, T.J., Purkait, P., Eriksson, A., Suter, D.: Efficient globally optimal consensus maximisation with tree search. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2413–2421 (2015)
Choi, S., Kim, T., Yu, W.: Performance evaluation of RANSAC family. In: British Machine Vision Conference (BMVC) (2009)
Chum, O., Matas, J.: Matching with PROSAC-progressive sample consensus. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), vol. 1, pp. 220–226. IEEE (2005)
Chum, O., Matas, J., Kittler, J.: Locally optimized RANSAC. In: Michaelis, B., Krell, G. (eds.) DAGM 2003. LNCS, vol. 2781, pp. 236–243. Springer, Heidelberg (2003). https://doi.org/10.1007/978-3-540-45243-0_31
Enqvist, O., Ask, E., Kahl, F., Åström, K.: Tractable algorithms for robust model estimation. Int. J. Comput. Vis. 112(1), 115–129 (2015)
Fischler, M.A., Bolles, R.C.: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 24(6), 381–395 (1981)
Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision. Cambridge University Press, Cambridge (2003)
Hu, J., Mitchell, J.E., Pang, J.S., Yu, B.: On linear programs with linear complementarity constraints. J. Glob. Optim. 53(1), 29–51 (2012)
Kahl, F., Hartley, R.: Multiple-view geometry under the \(l_\infty \)-norm. IEEE Trans. Pattern Anal. Mach. Intell. 30(9), 1603–1617 (2008)
Ke, Q., Kanade, T.: Quasiconvex optimization for robust geometric reconstruction. IEEE Trans. Pattern Anal. Mach. Intell. 29(10), 1834–1847 (2007)
Kovesi, P.D.: MATLAB and octave functions for computer vision and image processing. http://www.peterkovesi.com/matlabfns/
Le, H., Chin, T.J., Suter, D.: An exact penalty method for locally convergent maximum consensus. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2017)
Lebeda, K., Matas, J., Chum, O.: Fixing the locally optimized RANSAC-full experimental evaluation. In: British Machine Vision Conference, pp. 1–11. Citeseer (2012)
Li, H.: Consensus set maximization with guaranteed global optimality for robust geometry estimation. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 1074–1080. IEEE (2009)
Meer, P.: Robust techniques for computer vision. In: Emerging Topics in Computer Vision pp. 107–190 (2004)
Olsson, C., Enqvist, O., Kahl, F.: A polynomial-time bound for matching and registration with outliers. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2008, pp. 1–8. IEEE (2008)
Purkait, P., Zach, C., Eriksson, A.: Maximum consensus parameter estimation by reweighted \(\ell _1\) methods. In: Pelillo, M., Hancock, E. (eds.) EMMCVPR 2017. LNCS, vol. 10746, pp. 312–327. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-78199-0_21
Raguram, R., Chum, O., Pollefeys, M., Matas, J., Frahm, J.M.: USAC: a universal framework for random sample consensus. IEEE Trans. Pattern Anal. Mach. Intell 35(8), 2022–2038 (2013)
Snavely, N., Seitz, S.M., Szeliski, R.: Photo tourism: exploring photo collections in 3D. ACM Trans. Graph. (TOG) 25, 835–846 (2006)
Speciale, P., Paudel, D.P., Oswald, M.R., Kroeger, T., Gool, L.V., Pollefeys, M.: Consensus maximization with linear matrix inequality constraints. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5048–5056. IEEE (2017)
Tordoff, B.J., Murray, D.W.: Guided-MLESAC: faster image transform estimation by using matching priors. IEEE Trans. Pattern Anal. Mach. Intell. 27(10), 1523–1535 (2005)
Tran, Q.H., Chin, T.J., Chojnacki, W., Suter, D.: Sampling minimal subsets with large spans for robust estimation. Int. J. Comput. Vis. 106(1), 93–112 (2014)
Vedaldi, A., Fulkerson, B.: VLFeat: an open and portable library of computer vision algorithms (2008). http://www.vlfeat.org/
Wilson, K., Snavely, N.: Robust global translations with 1DSfM. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8691, pp. 61–75. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10578-9_5
Zheng, Y., Sugimoto, S., Okutomi, M.: Deterministically maximizing feasible subsystem for robust model fitting with unit norm constraint. In: 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1825–1832. IEEE (2011)
Acknowledgements
This work was supported by the ARC grant DP160103490.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Cai, Z., Chin, TJ., Le, H., Suter, D. (2018). Deterministic Consensus Maximization with Biconvex Programming. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds) Computer Vision – ECCV 2018. ECCV 2018. Lecture Notes in Computer Science(), vol 11216. Springer, Cham. https://doi.org/10.1007/978-3-030-01258-8_42
Download citation
DOI: https://doi.org/10.1007/978-3-030-01258-8_42
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-01257-1
Online ISBN: 978-3-030-01258-8
eBook Packages: Computer ScienceComputer Science (R0)