/external/opencv3/modules/videostab/src/ |
D | stabilizer.cpp | 79 frames_.clear(); in reset() 122 at(curPos_, frames_) = frame; in doOneIteration() 145 at(curStabilizedPos_ + radius_, frames_) = at(curPos_, frames_); in doOneIteration() 164 inpainter_->setFrames(frames_); in setUp() 178 deblurer_->setFrames(frames_); in setUp() 198 at(curStabilizedPos_, frames_).copyTo(preProcessedFrame_); in stabilizeFrame() 202 preProcessedFrame_ = at(curStabilizedPos_, frames_); in stabilizeFrame() 273 frames_.resize(cacheSize); in setUp() 282 at(i, frames_) = firstFrame; in setUp() 285 at(0, frames_) = firstFrame; in setUp() [all …]
|
D | inpainting.cpp | 154 const Mat_<Point3_<uchar> > &framei = at(idx + i, *frames_); in inpaint() 349 float err = alignementError(motion0to1, frame, mask, at(idx + i, *frames_)); in inpaint() 375 frame1_ = at(neighbor, *frames_); in inpaint()
|
/external/libchrome/base/trace_event/ |
D | heap_profiler_stack_frame_deduplicator.cc | 44 frame_index = static_cast<int>(frames_.size()); in Insert() 51 frames_.push_back(frame_node); in Insert() 57 nodes = &frames_[frame_index].children; in Insert() 103 size_t frames_allocated = frames_.capacity() * sizeof(FrameNode); in EstimateTraceMemoryOverhead() 104 size_t frames_resident = frames_.size() * sizeof(FrameNode); in EstimateTraceMemoryOverhead() 106 for (const FrameNode& node : frames_) in EstimateTraceMemoryOverhead()
|
D | heap_profiler_stack_frame_deduplicator.h | 57 ConstIterator begin() const { return frames_.begin(); } in begin() 58 ConstIterator end() const { return frames_.end(); } in end() 71 std::vector<FrameNode> frames_; variable
|
/external/webrtc/webrtc/modules/desktop_capture/ |
D | screen_capture_frame_queue.cc | 32 assert(!frames_[current_].get() || !frames_[current_]->IsShared()); in MoveToNextFrame() 36 frames_[current_].reset(SharedDesktopFrame::Wrap(frame)); in ReplaceCurrentFrame() 41 frames_[i].reset(); in Reset()
|
D | screen_capture_frame_queue.h | 55 return frames_[current_].get(); in current_frame() 59 return frames_[(current_ + kQueueLength - 1) % kQueueLength].get(); in previous_frame() 67 rtc::scoped_ptr<SharedDesktopFrame> frames_[kQueueLength]; variable
|
/external/opencv3/modules/cudacodec/src/ |
D | video_reader.cpp | 81 std::deque< std::pair<CUVIDPARSERDISPINFO, CUVIDPROCPARAMS> > frames_; member in __anonbf6f1fc10111::VideoReaderImpl 133 if (frames_.empty()) in nextFrame() 165 frames_.push_back(std::make_pair(displayInfo, videoProcParams)); in nextFrame() 169 if (frames_.empty()) in nextFrame() 172 std::pair<CUVIDPARSERDISPINFO, CUVIDPROCPARAMS> frameInfo = frames_.front(); in nextFrame() 173 frames_.pop_front(); in nextFrame() 191 if (frames_.empty()) in nextFrame()
|
/external/opencv3/modules/videostab/include/opencv2/videostab/ |
D | deblurring.hpp | 62 DeblurerBase() : radius_(0), frames_(0), motions_(0), blurrinessRates_(0) {} in DeblurerBase() 74 virtual void setFrames(const std::vector<Mat> &val) { frames_ = &val; } in setFrames() 75 virtual const std::vector<Mat>& frames() const { return *frames_; } in frames() 85 const std::vector<Mat> *frames_; member in cv::videostab::DeblurerBase
|
D | inpainting.hpp | 65 : radius_(0), motionModel_(MM_UNKNOWN), frames_(0), motions_(0), in InpainterBase() 81 virtual void setFrames(const std::vector<Mat> &val) { frames_ = &val; } in setFrames() 82 virtual const std::vector<Mat>& frames() const { return *frames_; } in frames() 96 const std::vector<Mat> *frames_; member in cv::videostab::InpainterBase
|
/external/google-breakpad/src/processor/ |
D | stackwalker_address_list.cc | 55 frames_(frames), in StackwalkerAddressList() 66 frame->instruction = frames_[0]; in GetContextFrame() 87 frame->instruction = frames_[frame_index]; in GetCallerFrame()
|
D | call_stack.cc | 46 for (vector<StackFrame *>::const_iterator iterator = frames_.begin(); in Clear() 47 iterator != frames_.end(); in Clear()
|
D | stackwalker.cc | 172 stack->frames_.push_back(frame.release()); in Walk() 173 if (stack->frames_.size() > max_frames_) { in Walk()
|
D | stackwalker_address_list.h | 64 const uint64_t* frames_; variable
|
/external/google-breakpad/src/google_breakpad/processor/ |
D | call_stack.h | 65 const vector<StackFrame*>* frames() const { return &frames_; } in frames() 72 vector<StackFrame*> frames_; variable
|
/external/opencv3/modules/cudaoptflow/src/ |
D | farneback.cpp | 159 GpuMat frames_[2]; member in __anon899078620111::FarnebackOpticalFlowImpl 321 frame0.convertTo(frames_[0], CV_32F, streams[0]); in calcImpl() 322 frame1.convertTo(frames_[1], CV_32F, streams[1]); in calcImpl() 329 pyramid0_[0] = frames_[0]; in calcImpl() 330 pyramid1_[0] = frames_[1]; in calcImpl() 428 …frames_[i], smoothSize/2, blurredFrame[i], BORDER_REFLECT101, StreamAccessor::getStream(streams[i]… in calcImpl()
|
/external/opencv3/3rdparty/libwebp/demux/ |
D | demux.c | 73 Frame* frames_; member 429 if (dmux->frames_ != NULL) return PARSE_ERROR; in ParseSingleImage() 590 const Frame* const frame = dmux->frames_; in IsValidSimpleFormat() 609 if (dmux->state_ == WEBP_DEMUX_DONE && dmux->frames_ == NULL) return 0; in IsValidExtendedFormat() 611 for (f = dmux->frames_; f != NULL; f = f->next_) { in IsValidExtendedFormat() 664 dmux->frames_tail_ = &dmux->frames_; in InitDemux() 712 for (f = dmux->frames_; f != NULL;) { in WebPDemuxDelete() 748 for (f = dmux->frames_; f != NULL; f = f->next_) { in GetFrame()
|
/external/webp/src/demux/ |
D | demux.c | 70 Frame* frames_; member 397 if (dmux->frames_ != NULL) return PARSE_ERROR; in ParseSingleImage() 562 const Frame* const frame = dmux->frames_; in IsValidSimpleFormat() 594 const Frame* f = dmux->frames_; in IsValidExtendedFormat() 600 if (dmux->state_ == WEBP_DEMUX_DONE && dmux->frames_ == NULL) return 0; in IsValidExtendedFormat() 656 dmux->frames_tail_ = &dmux->frames_; in InitDemux() 760 for (f = dmux->frames_; f != NULL;) { in WebPDemuxDelete() 794 for (f = dmux->frames_; f != NULL; f = f->next_) { in GetFrame()
|
/external/libvpx/libvpx/test/ |
D | encode_perf_test.cc | 28 uint32_t bitrate_, int frames_) in EncodePerfTestVideo() 33 frames(frames_) {} in EncodePerfTestVideo()
|
D | decode_perf_test.cc | 198 uint32_t bitrate_, int frames_) in EncodePerfTestVideo() 203 frames(frames_) {} in EncodePerfTestVideo()
|
/external/libvpx/libvpx/third_party/libwebm/ |
D | mkvmuxer.cpp | 2235 frames_(NULL), in Segment() 2270 if (frames_) { in ~Segment() 2272 Frame* const frame = frames_[i]; in ~Segment() 2275 delete[] frames_; in ~Segment() 2999 const Frame* const f = frames_[0]; // earliest queued frame in MakeNewCluster() 3164 frames[i] = frames_[i]; in QueueFrame() 3167 delete[] frames_; in QueueFrame() 3168 frames_ = frames; in QueueFrame() 3172 frames_[frames_size_++] = frame; in QueueFrame() 3178 if (frames_ == NULL) in WriteFramesAll() [all …]
|
/external/opencv3/modules/superres/src/ |
D | btv_l1_cuda.cpp | 435 std::vector<GpuMat> frames_; member in __anon26178dda0111::BTVL1_CUDA 460 frames_.clear(); in collectGarbage() 478 frames_.resize(cacheSize); in initImpl() 531 curFrame_.convertTo(at(storePos_, frames_), CV_32F); in readNextFrame() 564 srcFrames_[k] = at(i, frames_); in processFrame()
|
/external/v8/src/ |
D | deoptimizer.h | 245 iterator begin() { return frames_.begin(); } in begin() 246 iterator end() { return frames_.end(); } in end() 249 const_iterator begin() const { return frames_.begin(); } in begin() 250 const_iterator end() const { return frames_.end(); } in end() 252 std::vector<TranslatedFrame>& frames() { return frames_; } in frames() 284 std::vector<TranslatedFrame> frames_; variable
|
D | deoptimizer.cc | 3305 DCHECK(frames_.empty()); in Init() 3316 frames_.reserve(count); in Init() 3323 frames_.push_back(CreateNextTranslatedFrame( in Init() 3325 TranslatedFrame& frame = frames_.back(); in Init() 3376 for (auto& frame : frames_) frame.Handlify(); in Prepare() 3387 TranslatedFrame* frame = &(frames_[frame_index]); in MaterializeAt() 3549 frames_[pos.frame_index_].values_[pos.value_index_].GetValue(); in MaterializeAt() 3585 Handle<JSFunction>::cast(frames_[frame_index].front().GetValue()); in GetAdaptedArguments() 3589 TranslatedFrame* previous_frame = &(frames_[frame_index]); in GetAdaptedArguments() 3617 for (size_t i = 0; i < frames_.size(); i++) { in GetArgumentsInfoFromJSFrameIndex() [all …]
|
/external/webrtc/webrtc/video/ |
D | video_quality_test.cc | 144 frames_.push_back(copy); in IncomingCapturedFrame() 195 while (frames_.front().timestamp() < send_timestamp) { in RenderFrame() 196 AddFrameComparison(frames_.front(), last_rendered_frame_, true, in RenderFrame() 198 frames_.pop_front(); in RenderFrame() 201 VideoFrame reference_frame = frames_.front(); in RenderFrame() 202 frames_.pop_front(); in RenderFrame() 589 std::deque<VideoFrame> frames_ GUARDED_BY(crit_);
|
/external/opencv3/modules/video/src/ |
D | optflowgf.cpp | 639 frame0.convertTo(frames_[0], CV_32F); in operator ()() 640 frame1.convertTo(frames_[1], CV_32F); in operator ()() 647 pyramid0_[0] = frames_[0]; in operator ()() 648 pyramid1_[0] = frames_[1]; in operator ()() 743 if (!gaussianBlurOcl(frames_[i], smoothSize/2, blurredFrame[i])) in operator ()() 781 frames_[0].release(); in releaseMemory() 782 frames_[1].release(); in releaseMemory() 828 UMat frames_[2]; member in cv::FarnebackOpticalFlow
|