Home
last modified time | relevance | path

Searched refs:cameras (Results 1 – 25 of 37) sorted by relevance

12

/external/opencv3/modules/stitching/include/opencv2/stitching/detail/
Dmotion_estimators.hpp79 std::vector<CameraParams> &cameras) in operator ()() argument
80 { return estimate(features, pairwise_matches, cameras); } in operator ()()
93 std::vector<CameraParams> &cameras) = 0;
107 std::vector<CameraParams> &cameras);
148 std::vector<CameraParams> &cameras);
154 virtual void setUpInitialCameraParams(const std::vector<CameraParams> &cameras) = 0;
159 virtual void obtainRefinedCameraParams(std::vector<CameraParams> &cameras) const = 0;
210 void setUpInitialCameraParams(const std::vector<CameraParams> &cameras);
211 void obtainRefinedCameraParams(std::vector<CameraParams> &cameras) const;
230 void setUpInitialCameraParams(const std::vector<CameraParams> &cameras);
[all …]
/external/opencv3/modules/stitching/src/
Dmotion_estimators.cpp63 : num_images(_num_images), pairwise_matches(&_pairwise_matches[0]), cameras(&_cameras[0]) {} in CalcRotation()
70 K_from(0,0) = cameras[edge.from].focal; in operator ()()
71 K_from(1,1) = cameras[edge.from].focal * cameras[edge.from].aspect; in operator ()()
72 K_from(0,2) = cameras[edge.from].ppx; in operator ()()
73 K_from(1,2) = cameras[edge.from].ppy; in operator ()()
76 K_to(0,0) = cameras[edge.to].focal; in operator ()()
77 K_to(1,1) = cameras[edge.to].focal * cameras[edge.to].aspect; in operator ()()
78 K_to(0,2) = cameras[edge.to].ppx; in operator ()()
79 K_to(1,2) = cameras[edge.to].ppy; in operator ()()
82 cameras[edge.to].R = cameras[edge.from].R * R; in operator ()()
[all …]
/external/opencv3/samples/cpp/
Dstitching_detailed.cpp509 vector<CameraParams> cameras; in main() local
510 if (!estimator(features, pairwise_matches, cameras)) in main()
516 for (size_t i = 0; i < cameras.size(); ++i) in main()
519 cameras[i].R.convertTo(R, CV_32F); in main()
520 cameras[i].R = R; in main()
521 LOGLN("Initial intrinsics #" << indices[i]+1 << ":\n" << cameras[i].K()); in main()
540 if (!(*adjuster)(features, pairwise_matches, cameras)) in main()
549 for (size_t i = 0; i < cameras.size(); ++i) in main()
551 LOGLN("Camera #" << indices[i]+1 << ":\n" << cameras[i].K()); in main()
552 focals.push_back(cameras[i].focal); in main()
[all …]
/external/ceres-solver/examples/
Dbundle_adjuster.cc148 double* cameras = bal_problem->mutable_cameras(); in SetOrdering() local
155 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 0); in SetOrdering()
167 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 0); in SetOrdering()
176 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 1); in SetOrdering()
215 ordering->AddElementToGroup(cameras + camera_block_size * i, 1); in SetOrdering()
219 ordering->AddElementToGroup(cameras + camera_block_size * i + 4, 1); in SetOrdering()
254 double* cameras = bal_problem->mutable_cameras(); in BuildProblem() local
281 cameras + camera_block_size * bal_problem->camera_index()[i]; in BuildProblem()
303 problem->SetParameterization(cameras + camera_block_size * i, in BuildProblem()
Dbal_problem.cc248 double* cameras = mutable_cameras(); in Normalize() local
252 double* camera = cameras + camera_block_size() * i; in Normalize()
/external/robolectric/v1/src/main/java/com/xtremelabs/robolectric/shadows/
DShadowCamera.java28 … private static Map<Integer, Camera.CameraInfo> cameras = new HashMap<Integer,Camera.CameraInfo>(); field in ShadowCamera
104 Camera.CameraInfo foundCam = cameras.get( cameraId ); in getCameraInfo()
111 return cameras.size(); in getNumberOfCameras()
150 cameras.put(id, camInfo); in addCameraInfo()
154 cameras.clear(); in clearCameraInfo()
/external/autotest/client/cros/camera/
Dcamera_utils.py36 cameras = [os.path.basename(camera) for camera in
38 if not cameras:
40 camera = cameras[0]
/external/opencv3/samples/winrt/JavaScript/js/
DAdvancedCapture.js109 …ceInfo.findAllAsync(Windows.Devices.Enumeration.DeviceClass.videoCapture).done(function (cameras) { argument
110 if (cameras.length === 0) {
116 cameras.forEach(function (camera) {
/external/opencv3/doc/py_tutorials/py_calib3d/py_epipolar_geometry/
Dpy_epipolar_geometry.markdown17 it is an important question whether we can find the depth information using these cameras. And the
18 answer is to use more than one camera. Our eyes works in similar way where we use two cameras (two
25 two cameras taking the image of same scene.
62 cameras so that we can relate the two cameras in pixel coordinates. (If we are using rectified
/external/opencv3/modules/videoio/src/
Dcap_dc1394.cpp179 struct camnode {dc1394_cameracapture cam;int portnum;} cameras[MAX_CAMERAS]; variable
217 cameras[numCameras].cam.node = camera_nodes[i]; in icvInitCapture_DC1394()
218 cameras[numCameras].portnum = p; in icvInitCapture_DC1394()
243 pcap->device_name = videodev[cameras[index].portnum]; in icvCaptureFromCAM_DC1394()
244 pcap->handle = handles[cameras[index].portnum]; in icvCaptureFromCAM_DC1394()
245 pcap->camera = &cameras[index].cam; in icvCaptureFromCAM_DC1394()
Dcap_qtkit.mm102 * CvCaptureCAM is the instantiation of a capture source for cameras.
242 * CvCaptureCAM is the instantiation of a capture source for cameras.
Dcap_avfoundation.mm81 * CvCaptureCAM is the instantiation of a capture source for cameras.
226 * CvCaptureCAM is the instantiation of a capture source for cameras.
/external/ceres-solver/internal/ceres/
Dsystem_test.cc384 double* cameras = mutable_cameras(); in BuildProblem() local
397 double* camera = cameras + 9 * camera_index_[i]; in BuildProblem()
410 options_.linear_solver_ordering->AddElementToGroup(cameras + 9 * i, 1); in BuildProblem()
/external/libexif/
DREADME18 Fuji, Nikon, Olympus, Pentax and Sanyo cameras are also supported.
64 wide range of digital cameras (http://www.gphoto.org).
/external/opencv3/modules/stitching/include/opencv2/
Dstitching.hpp213 std::vector<detail::CameraParams> cameras() const { return cameras_; } in cameras() function in cv::Stitcher
/external/opencv3/doc/py_tutorials/py_calib3d/py_depthmap/
Dpy_depthmap.markdown26 their camera center. \f$B\f$ is the distance between two cameras (which we know) and \f$f\f$ is the…
/external/opencv3/doc/tutorials/calib3d/camera_calibration/
Dcamera_calibration.markdown5 cameras in the late 20th century, they became a common occurrence in our everyday life.
114 For some cameras we may need to flip the input image. Here we do this too.
129 Then again in case of cameras we only take camera images when an input delay time is passed.
277 quality cameras.
/external/ceres-solver/docs/source/
Dfaqs.rst141 cameras, use ``DENSE_SCHUR``.
148 5. For large bundle adjustment problems (a few thousand cameras or
/external/opencv3/doc/tutorials/highgui/
Dkinect_openni.markdown102 property. The following properties of cameras available through OpenNI interfaces are supported:
/external/opencv/
DREADME.rdoc53 … are two slightly different 'flavors' of running. Both are socket based cameras, but one is writt…
/external/opencv3/doc/tutorials/video/background_subtraction/
Dbackground_subtraction.markdown6 using static cameras.
/external/opencv3/modules/imgproc/doc/
Dcolors.markdown147 The Bayer pattern is widely used in CCD and CMOS cameras. It enables you to get color pictures
/external/opencv3/doc/py_tutorials/py_calib3d/py_calibration/
Dpy_calibration.markdown14 Today's cheap pinhole cameras introduces a lot of distortion to images. Two major distortions are
/external/libjpeg-turbo/
Dchange.log112 APP12 markers as text. Some digital cameras store useful text information
/external/opencv3/doc/tutorials/highgui/video-input-psnr-ssim/
Dvideo_input_psnr_ssim.markdown37 rate* specifying just how long is between two frames. While for the video cameras usually there is a

12