Distributed calibration of pan-tilt camera network using multi-layered belief propagation
Summary (3 min read)
1. Introduction
- The authors present a distributed algorithm for selfcalibration of a pan-tilt camera network using multi-layered belief propagation.
- As the cameras can pan and tilt, the camera network contains various mutually exclusive sub-networks, where, all cameras in a sub-network view a common region.
- The authors then propagate belief between sub-networks to obtain the globally consistent and accurate estimates of the camera parameters for each camera in the network.
- That the camera network be calibrated with respect to a global WCS so that tasks such as 3D-tracking, recognition of objects, activities and events can be effectively performed.
- The authors distributed calibration also leads to making the system scalable, as large camera networks spanning a wide geographical area would contain mutually exclusive sub-networks, thereby, no communication and computation among the cameras of these sub-networks would be necessary for calibration.
3. Distributed calibration of pan-tilt camera
- The authors also assume that each camera has a processing unit attached with it and that there exists an underlying communication network such that each camera can communicate with every other camera.
- In a pan-tilt camera network, there may exist many such mutually exclusive graphs at any point in time.
- To perform multi-layered belief propagation between two graphs containing the same camera in different pan-tilt positions, the authors need to bring the cameras to their home (zero pan and zero tilt) position in both the graphs.
- The authors also propose a protocol in Section 9, for aligning all the cameras’ home positions to a global WCS, to get a globally consistent estimate of the camera’s home position (zero pan, zero tilt position).
- In the next section, the authors give a method for automatically finding correspondences between three images.
4. Automatically finding corresponding points
- The authors propose a method for automatically finding corresponding points in three images.
- But, as the number of images increase, the error in correspondences also increase.
- First, compute the SIFT features in all three images and then, compute the SIFT matches between the pairs I1−I2, I1−I3 and I2−I3.
- Figure 1 shows the common points found between three images taken by three different cameras.
5. Finding the graphs
- Starting with the camera with the smallest number that does not belong to any graph currently, say Ci, find the camera with the next smallest number, say Cj , that has an overlap with Ci and which does not belong to any graph.
- In general, there will be more than one graph in the pan-tilt camera network.
- Moreover, each graph will be a complete graph.
- In a wide area pan-tilt camera network it is possible that two sets of cameras are geographically so far apart that there will be no overlapping view between these two sets of cameras.
6. Camera calibration within a graph
- The authors assume that the cameras in a graph, say Gk, remain static for a certain time period.
- Thus, standard multi-camera self-calibration techniques can be used for calibrating the cameras within a graph.
- The crucial point here is to automatically find multiview correspondences at each node.
- The corresponding points between the nodes ofGik are found automatically as discussed in Section 4.
- Belief propagation (discussed in Section 7) between the nodes of Gik gives a consistent estimate of the camera parameters for each camera in Gik.
7. Belief Propagation within a graph
- For distributed calibration of cameras in a graph, sayGk, multi-camera self-calibration is carried out at each node, using the automatically found corresponding points.
- As has been shown in [3], belief propagation can be directly applied on a graph which has cameras viewing a common scene as its nodes.
- Μ̃i,k and Σ̃i,k are the estimates of the camera parameters after belief propagation within graph Gk.
- The covariance matrix is calculated based on the forward covariance propagation from bundle adjustment.
7.1. Multi-layered Belief Propagation
- Since the graphs are dynamic and the same camera Ci can be a part of two graphs, say Gk−1 and Gk, in different pan-tilt orientations at different points in time, the authors perform belief propagation between graphs at each node, Ci, which is common in both Gk−1 and Gk.
- Similarly, the authors can get to the pan-tilt position as: Pθφ = H−1 ∗ Phome.
- In case, the pan-tilt view of the camera does not have any overlap with the home position’s view, a sequence of homographies can be used, again calculated automatically, as shown in Figure 2.
- The home position is calculated in each graph using the imageto-image homography before applying the update equations for multi-layered belief propagation.
8. Forming new graphs
- The multi-layered belief propagation mechanism can be utilized only if the graphs change across time.
- Each camera will have information of all other cameras about the landmark they are viewing.
- This also makes their system scalable as the correspondences have to be calculated among only those cameras which view the same landmark and in step 3, the messages have to be passed only between those cameras which can have overlapping views in some pan-tilt configuration.
- In the current time period these cameras are not considered for calibration and therefore, remain idle.
- In the next time period, they shall repeat the above protocol and become part of graphs with ≥ 3 nodes and hence, will be used for calibration and multi-layered belief propagation.
9. Aligning cameras to a global world coordinate system
- The authors want the position and orientation of each camera’s home position with respect to a global WCS.
- Moreover, belief propagation can be carried out only if all the cameras are aligned with respect to a common coordinate system in the world.
- These two conditions establish a common coordinate system at the lowest numbered camera, say Ci, in each graph formed in the camera network.
- All other cameras in Gj are aligned to this common coordinate system.
- In case the global WCS is not pre-specified, the lowest numbered camera in the network may be assumed to be at the origin of the global WCS.
10. Results and Discussion
- The authors use 6 SONY EVI-D70 PTZ cameras for their experiments.
- If the authors randomly select one camera (all its parameters) from each node, for example, P1 from node C2, P2 from C3 and P3 from C1, then as seen in Figure 3(b) and (c) the reprojection error is high and vary based on which camera is selected from which node.
- The reprojection error statistics are given in Table 1.
- The authors consider five 3-cliques of the graph for calibrating this graph.
- Multi-layered belief propagation at the nodes of the graph results in consistent and accurate camera parameters as seen in Figure 4.
11. Conclusion
- The authors have presented a multi-layered belief propagation based distributed algorithm for self-calibration of a pantilt camera network.
- The authors have shown that by using multi- layered belief propagation it is possible to get accurate and globally consistent estimates of the camera parameters for each pan-tilt camera in the network with respect to a global world coordinate system.
- The authors system does not require that all the cameras should have overlapping views at all times.
- The authors method gives an accurate and globally consistent estimate of the camera parameters for the home position of each camera and using the method for automatically finding correspondences in two views, homographies between the home view and any pan/tilt view can be automatically computed.
- Therefore, it is possible to obtain accurate and globally consistent camera parameters for any pan/tilt position of the pan-tilt cameras in the network with respect to a global world coordinate system.
Did you find this useful? Give us your feedback
Citations
12 citations
6 citations
Cites background from "Distributed calibration of pan-tilt..."
...At the same time, metric auto calibration has been obtained in the stochastic framework [35], [36], [37]....
[...]
5 citations
4 citations
2 citations
References
46,906 citations
15,558 citations
14,708 citations
1,539 citations
1,532 citations
"Distributed calibration of pan-tilt..." refers methods in this paper
...In general, belief propagation algorithm is used for solving inference problems based on local message passing [11]....
[...]
Related Papers (5)
Frequently Asked Questions (8)
Q2. How many cameras can be used to calibrate a larger network?
The authors show that by using multi-layered belief propagation it is sufficient to have correspondences between only three cameras at a time for consistent calibration of a larger N > 3 static camera network.
Q3. How do the authors perform multi-layered belief propagation between two graphs?
To perform multi-layered belief propagation between two graphs containing the same camera in different pan-tilt positions, the authors need to bring the cameras to their home (zero pan and zero tilt) position in both the graphs.
Q4. How can the authors get the camera parameters for pan-tilt cameras?
The authors have shown that by using multi-layered belief propagation it is possible to get accurate and globally consistent estimates of the camera parameters for each pan-tilt camera in the network with respect to a global world coordinate system.
Q5. What is the method for finding the points between the views of the cameras in each graph?
Corresponding points between the views of the cameras in each graph are found automatically and multicamera self-calibration is performed at each node of the graph.
Q6. What is the use of the equations for the home position of Ci?
Homography or a sequence of homographies is used to calculate the camera parameters for the home position of Ci, denoted by Pihome,k.
Q7. How can the authors calculate the camera matrix for the home position of a camera?
The authors show that the camera matrix for the home position of the camera can be computed by automatically finding pairwise correspondences to compute the homography or a sequence of homographies between the camera’s pan-tilt view and the home view.
Q8. How do they calculate the parameters of a camera network?
Very recently, authors in [4], have proposed a distributed algorithm for calibration of a camera sensor network, where they assume that one of the cameras is calibrated and use epipolar geometry based algorithms at each node to obtain its calibration parameters.