Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A Novel Camera Calibration Method Based on Polar Coordinate

  • Shaoyan Gai ,

    qxxymm@163.com

    Affiliations Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education, School of Automation, Southeast University, Nanjing 210096, China, State Key Laboratory of Digital Manufacturing Equipment & Technology, Huazhong University of Science & Technology, Wuhan 430074, China, Key Laboratory of Modern Agricultural Equipment and Technology(Jiangsu University), Ministry of Education, Zhenjiang 212013, China

  • Feipeng Da,

    Affiliation Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education, School of Automation, Southeast University, Nanjing 210096, China

  • Xu Fang

    Affiliation Key Laboratory of Measurement and Control of Complex Systems of Engineering, Ministry of Education, School of Automation, Southeast University, Nanjing 210096, China

Abstract

A novel calibration method based on polar coordinate is proposed. The world coordinates are expressed in the form of polar coordinates, which are converted to world coordinates in the calibration process. In the beginning, the calibration points are obtained in polar coordinates. By transformation between polar coordinates and rectangular coordinates, the points turn into form of rectangular coordinates. Then, the points are matched with the corresponding image coordinates. At last, the parameters are obtained by objective function optimization. By the proposed method, the relationships between objects and cameras are expressed in polar coordinates easily. It is suitable for multi-camera calibration. Cameras can be calibrated with fewer points. The calibration images can be positioned according to the location of cameras. The experiment results demonstrate that the proposed method is an efficient calibration method. By the method, cameras are calibrated conveniently with high accuracy.

Introduction

With the development of photo-electronics, image processing, information sensing, signal processing, and electronics, digital camera is becoming increasingly relevant in science and technology [114]. Camera calibration is an essential step in computer vision, image processing, and optical measurement [114], which makes it possible to obtain metric information of the object from the projections on the image plane. The accuracy of vision system is very sensitive to the camera parameters [1518]. A tiny error in estimating camera parameters may adversely affect the whole system performance.

Camera calibration has been widely studied, which falls into several categories [1819]. One category of the methods is called coplanar approaches. These methods rely on calibration points, which are on a planar template in a single depth. These approaches are either computationally complex or fail to provide result solutions for one camera parameters or more, e.g., the image center, the scale factor, or lens distortion coefficients[19]. The category of calibration methods based on World-reference are classical approaches, which requires a set of calibration points with two dimensional image coordinates and corresponding of three dimensional world coordinates [2022]. Zhang [23] proposed a flexible calibration method, which only requires a few images of a two-dimensional calibration plate taken from different orientations. Based on this method, a set of optimal conditions is proposed to improve the calibration results accurately[24]. The disadvantage of these methods is that a complex and high-precise calibration template is used to achieve precise 3D measurements. One-dimension calibration method is also studied [25]. By which, a 1D calibration object is placed at several positions and in different orientations[25].

Many researchers focus on the camera modeling and analysis, which is very important for stereo vision and display[2631]. Yang and Song model two kinds of cameras, a parallel and a converged one, and analyze the difference between them in horizontal and vertical parallax[26]. The conclusion is that converged arrays are more suitable for short-distance. The work can be guidance to camera arrays applications. And it is pointed that, for future work on the camera array, it is necessary to focus on camera calibration and visual stereo-video evaluation. In reference[27], two types of stereo cameras are studied, which are parallel and toed-in cameras. The objective shooting quality evaluation criteria over short distance is proposed. Furthermore, three shooting conditions (macro shooting, short, and long distance shooting) are discussed in reference [28]. The shooting quality of stereo cameras can be evaluated effectively by the proposed approach. In reference [29], a full-reference metric for quality assessment of stereoscopic images is proposed. It is based on the binocular difference channel and binocular summation channel.

Reference [31] use point correspondences between model plane and image to calculate the homograph and distortion coefficients. A calibration process is applied, which is non-iterative with no risk of local minima. It’s a one-shot algorithm that can be solved by linear least-squares technique.

Generally, the 3D world coordinates and 2D image plan coordinates are related by calibration objects. They are in Cartesian coordinates usually. But some images are very complex, and cannot be represent in rectangular coordinate system. For example, rose line, Archimedes line, etc. For these images, the method based on rectangular coordinates cannot be used generally. But the characteristics of the curve is obvious. They are suitable for the feature point extraction and matching. Thus in this paper, polar coordinates are introduced to indicate the location of the calibration points. The coordinates of the points are obtained in polar coordinates, and then converted into rectangular coordinate system. The point can be easily expressed. The calibration can be applied without increasing system complex.

Polar coordinate images are suitable for multiple-camera calibration. The calibration object for multiple-camera has always been a problem: the 3D calibration objects may be hidden partly in one of the camera images; Two-dimension calibration plates are designed large and complex due to the scope of cameras in multiple locations. Two-dimensional flexible spliced calibration board are studied [32, 33]. Also one-dimensional calibration objects are studied [34, 35, 36] to solve the problems However there are movement restrictions in these solutions. Theodolite is used to obtain the precise and relative position [37]. However, additional equipment and complex operations is required.

In this paper, a whole solution of calibration by polar coordinates is presented. This method is in a novel view of calibration, that is the polar coordinates. The relationship in polar coordinates image is relatively simple, which can be easily used in the multiple-camera calibration plate. Furthermore, the layout of the polar coordinates calibration plate can be designed according to the camera position. Thus the manufacture difficulty of calibration objects is reduced. And the complexity of the calibration is simplified.

The method is useful in practice. Unlike the traditional methods based on rectangular coordinate system, the calibration board is designed in polar coordinates, and the difficulties in multiple-camera calibration is overcome. In the second section, the camera model and the calibration method is proposed in detail. In the third section, the simulation experiment and real experiment are presented. In the last section, the method of this paper is summarized.

Methods

Pinhole projection model

A camera is modeled by the usual pinhole model, in which case the relationship between a 3D point (Xw,Yw,Zw) and its image projection (u,v) is given by (1) where where (Xw,Yw,Zw) is the 3D coordinates of the point at the world coordinate system. In the camera image plan, the point image is produced at point (u,v). ρ is an arbitrary scale factor. fx, fy, u0, v0, θ are camera intrinsics, which is included in the camera intrinsic matrix M1, where fx, fy are the scale factors in image abscissa and vertical axes respectively, and u0, v0 define the coordinates of the principal point in the image plan of the camera, and θ is the parameter describing the skewness of the two image axes. R, T are the rotation and translation matrixes, which comprise the camera extrinsic matrix M2, which define the spatial relationship of camera and the world coordinate system. R is a general 3×3 orthogonal rotation matrix, and T is a 3×1 translation matrix. M1 and M2 are camera intrinsic matrix and extrinsic matrix respectively. H is the perspective projection matrix (3×4), which is the product of M1 and M2.

Calibration in polar coordinate

In view of expressing graphics in polar coordinates, there are two ideas:

  1. The world coordinates and computer image coordinates are both expressed in polar coordinates;
  2. The world coordinates are expressed in polar coordinates, which are converted to world coordinates in the calibration process. Computer image coordinates are in rectangular coordinates.

In the case of method 1, the system have to be modified to satisfy Eq (1). The left of Eq (1) should be in polar coordinates. However the image hardware implementation is a representation of pixel rows and columns, which is arranged according to the rectangular coordinate system. Thus we use the second method.

Some calibration image patterns, e.g. Archimedes line and rose line in Fig 1 and Fig 2, are suitable for polar coordinates. The characteristic point are in form of polar coordinates. Then the rectangular coordinates of the points are obtained by coordinate transformation. And the corresponding image coordinates are matched with the points. By this simple transformation, it will not increase the complexity of the system.

Computation of H

As we known it in Eq (1), H is a 3×4 matrix. The calibration image base on polar coordinate is in a plane. Without loss of generality, the plane is Zw = 0. Then Eq (1) becomes (2)

As can be seen from Eq (2), the third column of H is omitted since Zw = 0. Thus H turned into a 3×3 matrix, which contains 9 unknowns. Since there is a scaling factor ρ, there are only eight unknown parameters. By four (or more) groups of corresponding points (u,v)—(Xw,Yw,Zw), the matrix H can be calculated.

Computation of camera parameters

By Eq (1) and Eq (2), H can be expressed as a column-matrix form, as follows (3)

Where r1, r2 is the first and second columns of orthogonal rotation matrix R. T is the translation matrix. By the nature of orthogonal matrix, we have (4)

Where t denotes matrix transposition. From Eq (2) and Eq (3), we have (5)

According to Eq (5), two equations can be obtained from each image. There are five unknowns in the intrinsic matrix M1. If there are three images, the intrinsic matrix can be determined with Eq (5). If there are more than 3 images, an optimal solution can be worked out.

Then, the extrinsic matrix can be obtained by the result of M1 and Eq (4), as follows (6) where λ = 1/‖M1-1h1‖ = 1/‖M1-1h2‖. The result is a primary solution, without lens distortion.

Furthermore, we consider lens distortion, which can be express as (7)

Where (u,v) and (ud,vd) are the ideal pixel image coordinates and the real image coordinates, respectively. k1,k2 are coefficients of radial distortion.

Then, the optimal solution with radial distortion can be calculated. The objective function for optimization is defined as (8) where pij is the actual coordinates of the j-th point in i-th image, (udij,vdij) is the pixel image coordinates, is the calculation result of coordinates, n denotes the total number of images, and m denotes the number of points in each image.

Multiple-camera calibration

Due to the simple transform relationship in polar coordinates, it is convenient for multiple-camera calibration. Calibration board can be designed as the graphics shown in Fig 3. By the relationship between each board image and its center (ri,θi), it is easy to determine the absolute position of the points in polar coordinates.

thumbnail
Fig 3. Multi-camera calibration board based on cameras’ positions.

https://doi.org/10.1371/journal.pone.0165487.g003

The great advantage of polar coordinates in multiple-camera calibration is as follows.

It does not require common areas for all the cameras. For example, for camera 1, the center of the world coordinates is set to polar coordinates center 1. For camera 2, the center of the world coordinates is set to polar coordinates center 2. For camera 3, the center of the world coordinates is set to polar coordinates center 3. The center of the world coordinates of camera 2 and 3, is about the position of calibration board center. In this case, the calibration board can be changed according to the location of cameras. In other word, the location is not fixed in advance, but on or after the image is snapped. Thus it solves the problem of accumulated error caused by transfer matrix. As shown in Fig 4, center of polar coordinates 1 is (0, 0) in world coordinate, center of polar coordinates 2 is (24, 17) in world coordinate, and center of polar coordinates 3 is (48, -36) in world coordinate. Thus the world coordinate of the cameras is the same one. More than the traditional multiple camera calibration methods, where it requires public viewing area between two cameras [35,36], the proposed approach is free of public viewing area. It is not considered that data may be lost because cameras are of imperfect convergence configuration. The reason is that the calibration distance is shorter than the retrieval distance of the object. Thus there is no or few data-missing problem. The flexibility of calibration is enhanced to a large extent.

thumbnail
Fig 4. Multi-camera calibration board with known positions of centers.

https://doi.org/10.1371/journal.pone.0165487.g004

Experiment Results

Experiment 1 (Simulation)

In the following simulation experiments, the calibration objects is in form of Archimedes curve, as shown in Fig 1. The parameters are p = θ (θ = 0:6π), θ = 45° × k, k is integer, and θ ≥ 225°. A set of points, which are 12 points, are made as feature points.

The intrinsic matrix M1 is:

The extrinsic parameters of three images are:

1st image:

        α1 = 16.9°, β1 = -11.6°, γ1 = -27.3°, T1 = [3,3,10]

2nd image:

        α1 = 26.8°, β1 = -8.6°, γ1 = -38.5°, T1 = [2,2,10]

3rd image:

        α1 = 28.3°, β1 = -19.9°, γ1 = 40.1°, T1 = [1,2,10]

According to the above parameters, 3 calibration images are generated with a resolution of 2048 by 2048. In the projected image, gaussian noise with zero mean is added. The standard deviation of the noise is varying from 0 to 1 at an interval of 0.1 pixels. The calculation is performed 200 times under each noise level. The obtained parameters are shown. The Average of parameters under different noise levels are shown in Table 1 and Table 2, and the standard deviation is shown in Figs 5, 6 and 7. The horizontal ordinates of Figs 57 denote the noise level in pixel. The vertical ordinates denote the standard deviation of parameters. The standard deviation of u0, v0 is shown in Fig 5. The standard deviation of fx, fy is shown in Fig 6. The standard deviation of α,β,γ is shown in Fig 7.

thumbnail
Table 1. Average of intrinsic parameters under different noise levels.

https://doi.org/10.1371/journal.pone.0165487.t001

thumbnail
Table 2. Average of extrinsic parameters under different noise level.

https://doi.org/10.1371/journal.pone.0165487.t002

It can be clearly seen that, intrinsic parameters deviation is of the order of magnitude of 0.01 to 0.1, while extrinsic parameters deviation is of the order of magnitude of 1. Since the value of intrinsic parameters is large, the proportion of the deviation may be small. For example, fx = 1000, and the max deviation of fx is 0.6, thus the proportion of the deviation is 0.6%. In case of the extrinsic parameters, the max deviation of α is 0.225, and the proportion of the deviation is 1.3%. In conclusion, the noise impact on the results is relatively small. The algorithm is of high stability.

To study the inference of the numbers of pictures, we try different numbers of pictures. The results of 3 pictures and 10 pictures are shown in Figs 810. The horizontal ordinates of Figs 810 denote the noise level in pixel. The vertical ordinates denote the standard deviation of parameters. The Standard deviations for intrinsic parameters fx, u0 and extrinsic parameters α are shown in Fig 8, Fig 9 and Fig 10, respectively.

thumbnail
Fig 8. the standard deviations of fx with 3 and 10 pictures.

https://doi.org/10.1371/journal.pone.0165487.g008

thumbnail
Fig 9. the standard deviations of u0 with 3 and 10 pictures.

https://doi.org/10.1371/journal.pone.0165487.g009

thumbnail
Fig 10. the standard deviations of α with 3 and 10 pictures.

https://doi.org/10.1371/journal.pone.0165487.g010

As shown in Figs 810, the stability of 10 pictures is better than results of 3 pictures. But with 10 pictures, the calibration processing is complex significantly, while the improvement is not notable. In practice, 3–5 pictures are suitable for calibration applications.

Experiment 2 (Real data experiments)

The real data experiments were carried out in our laboratory. To demonstrate the efficiency of the proposed method, camera calibration based on proposed calibration image and accurate calibration plate are conducted. And zhang’smethod, which is the classical calibration method, is also peformed.

A UNIQ UP-1800 camera with fixed focal length is placed in front of the calibration boards. The image resolution is 1380×1030 pixels. The training data for Zhang’s calibration is obtained from the accurate calibration board. While the training data for the proposed method is obtained from the polar coordinates calibration board. The images of calibration boards are shown in Fig 11, where the accurate board is at the top, and the polar coordinates board is at the bottom.

thumbnail
Fig 11. Real Pictures—the accurate calibration board is at the top, the polar coordinates calibration board is at the bottom.

https://doi.org/10.1371/journal.pone.0165487.g011

The results are listed as follows

The calibration results of intrinsic parameters by the proposed method:

The calibration results of intrinsic parameters by accurate calibration plate:

Both of algorithms obtain accurate calibration results. From a practical point of view, the proposed method is more simple and convenient, and the broad is of high applicability.

The relationship between Polar coordinates images is simple. Thus Polar coordinates calibration images can be easily used for multiple camera calibration. The calibration can be performed by pre-made calibration board, as shown in Fig 12, the respective relationship of which is known before snapping images. Also the relationship can be calculated according to the snapped images, as shown in Fig 3. In this case, the calibration images can be positioned according to the location of the cameras.

The experiment results of multiple camera calibration are shown as follows.

The optimized parameters of left camera are

The optimized parameters of right camera are

The diagrams of retrieval results are shown in Figs 13 and 14.

As shown in Figs 13 and 14, 3D point cloud are obtained successfully by both the methods. There is little difference in the results. The retrieval result of the plate is consistent with the result by Zhang’s method, thus the calibration result is close to Zhang’s method. The distances between adjacent 3D points are calculated. The distance results are gathered along horizontal and vertical direction.By Zhang’s method, the average distance in horizontal and vertical direction is 2.553, and the standard deviation is 0.01619. By the proposed method, the corresponding average distance and deviation is 2.648, 0.016724. It is shown that the proposed method can be used for multiple camera calibration conveniently.

Conclusion

A novel study on camera calibration was carried out in the polar coordinate. There are several advantages of the proposed method. Some calibration image patterns are flexible in polar coordinate. Fewer calibration points are used. To calibrate multiple camera, the calibration images are placed according to the location of the camera. There is no accumulated error caused by transfer matrix. Zhang’s method, as well as the proposed method, are chosen for experimentation on both simulated and real data. The accuracy were evaluated. The efficiency of the proposed method is demonstrated in the experiments.

Supporting Information

S3 File. Data of parameters with 3 pictures.

https://doi.org/10.1371/journal.pone.0165487.s003

(XLSX)

S4 File. Data of parameters with 10 pictures.

https://doi.org/10.1371/journal.pone.0165487.s004

(XLSX)

Acknowledgments

This research is supported by National Natural Science Foundation of P.R.China(61628304, 51475092). Open Research Foundation of Key Laboratory of Modern Agricultural Equipment and Technology in Jiangsu University (NZ201301). Open Research Foundation of State Key Lab. of Digital Manufacturing Equipment & Technology in Huazhong University of Science & Technology (DMETKF2012002). A Project Funded by the Priority Academic Program Development of Jiangsu Higher Education Institutions.

Author Contributions

  1. Conceptualization: SG.
  2. Data curation: XF.
  3. Formal analysis: XF.
  4. Investigation: FD.
  5. Methodology: SG.
  6. Supervision: FD.
  7. Validation: XF.
  8. Visualization: XF SG.
  9. Writing – original draft: SG XF.
  10. Writing – review & editing: SG XF.

References

  1. 1. Chen Y, Zhang R, Shang L. A novel method of object detection from a moving camera based on image matching and frame coupling.[J]. Plos One. 2014; 9(10):e109809–e109809. pmid:25354301
  2. 2. Gai S, Da F. A novel fringe adaptation method for digital projector[J]. Optics & Lasers in Engineering. 2011; 49(4):547–552.
  3. 3. Jin T, Hou X, Li P, Zhou F. A Novel Method of Automatic Plant Species Identification Using Sparse Representation of Leaf Tooth Features.[J]. Plos One. 2015; 10(10).
  4. 4. Da F, Gai S. Flexible three-dimensional measurement technique based on a digital light processing projector.[J]. Applied Optics. 2008; 47(3):377–85. pmid:18204725
  5. 5. Tavoli R, Kozegar E, Shojafar M, Soleimani H, Pooranian Z. Weighted PCA for improving Document Image Retrieval System based on keyword spotting accuracy[C]. International Conference on Telecommunications and Signal Processing. 2013:773–777.
  6. 6. Keyvanpour M, Tavoli R. Document Image Retrieval: Algorithms, Analysis and Promising Directions[J]. International Journal of Software Engineering & Its Applications, 2013, 7:93–106.
  7. 7. Baccarelli E, Cordeschi N, Mei A, Panella M, Shojafar M,Stefa J. Energy-efficient dynamic traffic offloading and reconfiguration of networked data centers for big data stream mobile computing: review, challenges, and a case study[J]. IEEE Network, 2016, 30(2):54–61.
  8. 8. Pouryazdan M, Kantarci B, Soyata T, Song H. Anchor-Assisted and Vote-Based Trustworthiness Assurance in Smart City Crowdsensing[J]. IEEE Access, 2016, 4:1–1.
  9. 9. Xu Q, Ren P, Song H, Du Q. Security Enhancement for IoT Communications Exposed to Eavesdroppers With Uncertain Locations[J]. IEEE Access, 2016, 4:2840–2853.
  10. 10. Xiong J, Cai L, Jiang D, Song H, He X. Spectral Matrix Decomposition-Based Motion Artifacts Removal in Multi-Channel PPG Sensor Signals[J]. IEEE Access, 2016, 4:3076–3086.
  11. 11. Gai S, Da F, Dai X.Novel 3D measurement system based on speckle and fringe pattern projection. Optics Express[J]. 2016,24(16):17686–17697 pmid:27505737
  12. 12. Wang C, Da F. Dai X.Differential signal-assisted method for adaptive analysis of fringe pattern. Applied optics[J].2014, 53(27): 6222–6229. pmid:25322101
  13. 13. Deng X, Da F. Expression-robust 3D face recognition based on feature-level fusion and feature-region fusion. Multimedia Tools and Applications[J]. 2015: 1–19.
  14. 14. Zhi Y., Lu R., Wang B., Zhang Q., and Yao X.C.. Rapid super-resolution line scanning microscopy through virtually structured detection. Optics Letters[J]. 2015, 40:1683–1686 pmid:25872047
  15. 15. Zappa E, Busca G. Static and dynamic features of Fourier transform profilometry: A review[J]. Optics & Lasers in Engineering. 2012; 50(8):1140–1151.
  16. 16. Sun Q, Hou Y, Tan Q, Li G. A flexible calibration method using the planar target with a square pattern for line structured light vision system.[J]. Plos One. 2014; 9(9):e106911–e106911. pmid:25203507
  17. 17. Zhou Z, Huang J, Wang J, Zhang K, Kuang Z. Object-Oriented Classification of Sugarcane Using Time-Series Middle-Resolution Remote Sensing Data Based on AdaBoost[J]. Plos One. 2015; 10(11).
  18. 18. Busca G, Zappa E. Sensitivity analysis applied to an improved Fourier-transform profilometry[J]. Optics & Lasers in Engineering. 2011; 49(2):210–221.
  19. 19. Chatterjee C, Roychowdhury V P. Algorithms for coplanar camera calibration[J]. Machine Vision & Applications. 2000; 12(2):84–97.
  20. 20. Tsai R Y. A Versatile Camera Calibration Technique for High-Accuracy 3D Machine Vision Metrology Using Off-the-Shelf TV Cameras and Lenses[J]. IEEE Journal on Robotics and Automation. 1987; 3(4):323–344.
  21. 21. Heikkil J. Geometric Camera Calibration Using Circular Control Points[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence. 2000; 22(10):1066–1077.
  22. 22. Salvi J, Armangué X, Batlle J. A comparative review of camera calibrating methods with accuracy evaluation [J]. Pattern Recognition. 2010; 35(7):1617–1635.
  23. 23. Zhang Z. A Flexible New Technique for Camera Calibration[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence. 2000; 22(11):1330–1334.
  24. 24. Ricolfe-Viala C, Sanchez-Salmeron A J. Camera calibration under optimal conditions.[J]. Optics Express. 2011; 19(11):10769–75. pmid:21643333
  25. 25. Zhang Z. Camera Calibration with One-Dimensional Objects[J]. IEEE Transactions on Pattern Analysis & Machine Intelligence. 2004; 26(7):892–9.
  26. 26. Yang J, Xu R, Lv Z, Song H. Analysis of Camera Arrays Applicable to the Internet of Things:[J]. Sensors, 2016, 16(3):421,1–12.
  27. 27. Liu Y, Yang J, Chu R. Objective Evaluation Criteria for Shooting Quality of Stereo Cameras over Short Distance[J]. Radioengineering, 2015, 24(1):305–313.
  28. 28. Yang J, Liu Y, Meng Q, Chu R. Objective Evaluation Criteria for Stereo Camera Shooting Quality Under Different Shooting Parameters and Shooting Distances[J]. IEEE Sensors Journal, 2015, 15(8):1–1.
  29. 29. Yang J, Lin Y, Gao Z, Lv Z, Wei W. Quality Index for Stereoscopic Images by Separately Evaluating Adding and Subtracting[J]. Plos One, 2015, 10(12): e0145800. pmid:26717412
  30. 30. Wei W, Shen P, Peng X, Song H, Zhang L. Simultaneous enhancement and noise reduction of a single low-light image[J]. Iet Image Processing, 2016.
  31. 31. Hong Y, Ren G, Liu E. Non-iterative method for camera calibration.[J]. Optics Express. 2015; 23(18):23992–4003. pmid:26368490
  32. 32. Yang J, Yang Q X, Qin P L. Two-dimensional flexible target for calibrating camera[J]. Optics & Precision Engineering. 2011; 19(5):1134–1142.
  33. 33. Liu Y H, Jia Q X, Sun H X, Su J. Multi-camera calibration based on coplanar projection of points[C]// Computer Design and Applications (ICCDA), 2010 International Conference on. IEEE. 2010:V1-221–V1-224.
  34. 34. Kojima Y, Fujii T, Tanimoto M. New multiple-camera calibration method for a large number of cameras[J]. Proceedings of SPIE—The International Society for Optical Engineering. 2005; 5665:156–163.
  35. 35. Loaiza M E, Raposo A B, Gattass M. Multi-camera calibration based on an invariant pattern[J]. Computers & Graphics. 2011; 35(2):198–207.
  36. 36. Ueshiba T, Tomita F. Plane-based Calibration Algorithm for Multi-camera Systems via Factorization of Homography Matrices[C]// IEEE International Conference on Computer Vision, 2003. Proceedings. IEEE. 2003:966–973.
  37. 37. Zhang G. Unique world coordiantes based global calibration method for multi-vision inspection system[J]. Journal of Beijing University of Aeronautics & Astronautics. 2006; 32(11):1268–1272.