/external/opencv3/modules/stitching/include/opencv2/stitching/detail/ |
D | motion_estimators.hpp | 79 std::vector<CameraParams> &cameras) in operator ()() argument 80 { return estimate(features, pairwise_matches, cameras); } in operator ()() 93 std::vector<CameraParams> &cameras) = 0; 107 std::vector<CameraParams> &cameras); 148 std::vector<CameraParams> &cameras); 154 virtual void setUpInitialCameraParams(const std::vector<CameraParams> &cameras) = 0; 159 virtual void obtainRefinedCameraParams(std::vector<CameraParams> &cameras) const = 0; 210 void setUpInitialCameraParams(const std::vector<CameraParams> &cameras); 211 void obtainRefinedCameraParams(std::vector<CameraParams> &cameras) const; 230 void setUpInitialCameraParams(const std::vector<CameraParams> &cameras); [all …]
|
/external/opencv3/modules/stitching/src/ |
D | motion_estimators.cpp | 63 : num_images(_num_images), pairwise_matches(&_pairwise_matches[0]), cameras(&_cameras[0]) {} in CalcRotation() 70 K_from(0,0) = cameras[edge.from].focal; in operator ()() 71 K_from(1,1) = cameras[edge.from].focal * cameras[edge.from].aspect; in operator ()() 72 K_from(0,2) = cameras[edge.from].ppx; in operator ()() 73 K_from(1,2) = cameras[edge.from].ppy; in operator ()() 76 K_to(0,0) = cameras[edge.to].focal; in operator ()() 77 K_to(1,1) = cameras[edge.to].focal * cameras[edge.to].aspect; in operator ()() 78 K_to(0,2) = cameras[edge.to].ppx; in operator ()() 79 K_to(1,2) = cameras[edge.to].ppy; in operator ()() 82 cameras[edge.to].R = cameras[edge.from].R * R; in operator ()() [all …]
|
/external/opencv3/samples/cpp/ |
D | stitching_detailed.cpp | 509 vector<CameraParams> cameras; in main() local 510 if (!estimator(features, pairwise_matches, cameras)) in main() 516 for (size_t i = 0; i < cameras.size(); ++i) in main() 519 cameras[i].R.convertTo(R, CV_32F); in main() 520 cameras[i].R = R; in main() 521 LOGLN("Initial intrinsics #" << indices[i]+1 << ":\n" << cameras[i].K()); in main() 540 if (!(*adjuster)(features, pairwise_matches, cameras)) in main() 549 for (size_t i = 0; i < cameras.size(); ++i) in main() 551 LOGLN("Camera #" << indices[i]+1 << ":\n" << cameras[i].K()); in main() 552 focals.push_back(cameras[i].focal); in main() [all …]
|
/external/ceres-solver/examples/ |
D | bundle_adjuster.cc | 148 double* cameras = bal_problem->mutable_cameras(); in SetOrdering() local 155 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 0); in SetOrdering() 167 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 0); in SetOrdering() 176 options->inner_iteration_ordering->AddElementToGroup(cameras + camera_block_size * i, 1); in SetOrdering() 215 ordering->AddElementToGroup(cameras + camera_block_size * i, 1); in SetOrdering() 219 ordering->AddElementToGroup(cameras + camera_block_size * i + 4, 1); in SetOrdering() 254 double* cameras = bal_problem->mutable_cameras(); in BuildProblem() local 281 cameras + camera_block_size * bal_problem->camera_index()[i]; in BuildProblem() 303 problem->SetParameterization(cameras + camera_block_size * i, in BuildProblem()
|
D | bal_problem.cc | 248 double* cameras = mutable_cameras(); in Normalize() local 252 double* camera = cameras + camera_block_size() * i; in Normalize()
|
/external/robolectric/v1/src/main/java/com/xtremelabs/robolectric/shadows/ |
D | ShadowCamera.java | 28 … private static Map<Integer, Camera.CameraInfo> cameras = new HashMap<Integer,Camera.CameraInfo>(); field in ShadowCamera 104 Camera.CameraInfo foundCam = cameras.get( cameraId ); in getCameraInfo() 111 return cameras.size(); in getNumberOfCameras() 150 cameras.put(id, camInfo); in addCameraInfo() 154 cameras.clear(); in clearCameraInfo()
|
/external/autotest/client/cros/camera/ |
D | camera_utils.py | 36 cameras = [os.path.basename(camera) for camera in 38 if not cameras: 40 camera = cameras[0]
|
/external/opencv3/samples/winrt/JavaScript/js/ |
D | AdvancedCapture.js | 109 …ceInfo.findAllAsync(Windows.Devices.Enumeration.DeviceClass.videoCapture).done(function (cameras) { argument 110 if (cameras.length === 0) { 116 cameras.forEach(function (camera) {
|
/external/opencv3/doc/py_tutorials/py_calib3d/py_epipolar_geometry/ |
D | py_epipolar_geometry.markdown | 17 it is an important question whether we can find the depth information using these cameras. And the 18 answer is to use more than one camera. Our eyes works in similar way where we use two cameras (two 25 two cameras taking the image of same scene. 62 cameras so that we can relate the two cameras in pixel coordinates. (If we are using rectified
|
/external/opencv3/modules/videoio/src/ |
D | cap_dc1394.cpp | 179 struct camnode {dc1394_cameracapture cam;int portnum;} cameras[MAX_CAMERAS]; variable 217 cameras[numCameras].cam.node = camera_nodes[i]; in icvInitCapture_DC1394() 218 cameras[numCameras].portnum = p; in icvInitCapture_DC1394() 243 pcap->device_name = videodev[cameras[index].portnum]; in icvCaptureFromCAM_DC1394() 244 pcap->handle = handles[cameras[index].portnum]; in icvCaptureFromCAM_DC1394() 245 pcap->camera = &cameras[index].cam; in icvCaptureFromCAM_DC1394()
|
D | cap_qtkit.mm | 102 * CvCaptureCAM is the instantiation of a capture source for cameras. 242 * CvCaptureCAM is the instantiation of a capture source for cameras.
|
D | cap_avfoundation.mm | 81 * CvCaptureCAM is the instantiation of a capture source for cameras. 226 * CvCaptureCAM is the instantiation of a capture source for cameras.
|
/external/ceres-solver/internal/ceres/ |
D | system_test.cc | 384 double* cameras = mutable_cameras(); in BuildProblem() local 397 double* camera = cameras + 9 * camera_index_[i]; in BuildProblem() 410 options_.linear_solver_ordering->AddElementToGroup(cameras + 9 * i, 1); in BuildProblem()
|
/external/libexif/ |
D | README | 18 Fuji, Nikon, Olympus, Pentax and Sanyo cameras are also supported. 64 wide range of digital cameras (http://www.gphoto.org).
|
/external/opencv3/modules/stitching/include/opencv2/ |
D | stitching.hpp | 213 std::vector<detail::CameraParams> cameras() const { return cameras_; } in cameras() function in cv::Stitcher
|
/external/opencv3/doc/py_tutorials/py_calib3d/py_depthmap/ |
D | py_depthmap.markdown | 26 their camera center. \f$B\f$ is the distance between two cameras (which we know) and \f$f\f$ is the…
|
/external/opencv3/doc/tutorials/calib3d/camera_calibration/ |
D | camera_calibration.markdown | 5 cameras in the late 20th century, they became a common occurrence in our everyday life. 114 For some cameras we may need to flip the input image. Here we do this too. 129 Then again in case of cameras we only take camera images when an input delay time is passed. 277 quality cameras.
|
/external/ceres-solver/docs/source/ |
D | faqs.rst | 141 cameras, use ``DENSE_SCHUR``. 148 5. For large bundle adjustment problems (a few thousand cameras or
|
/external/opencv3/doc/tutorials/highgui/ |
D | kinect_openni.markdown | 102 property. The following properties of cameras available through OpenNI interfaces are supported:
|
/external/opencv/ |
D | README.rdoc | 53 … are two slightly different 'flavors' of running. Both are socket based cameras, but one is writt…
|
/external/opencv3/doc/tutorials/video/background_subtraction/ |
D | background_subtraction.markdown | 6 using static cameras.
|
/external/opencv3/modules/imgproc/doc/ |
D | colors.markdown | 147 The Bayer pattern is widely used in CCD and CMOS cameras. It enables you to get color pictures
|
/external/opencv3/doc/py_tutorials/py_calib3d/py_calibration/ |
D | py_calibration.markdown | 14 Today's cheap pinhole cameras introduces a lot of distortion to images. Two major distortions are
|
/external/libjpeg-turbo/ |
D | change.log | 112 APP12 markers as text. Some digital cameras store useful text information
|
/external/opencv3/doc/tutorials/highgui/video-input-psnr-ssim/ |
D | video_input_psnr_ssim.markdown | 37 rate* specifying just how long is between two frames. While for the video cameras usually there is a
|