/hardware/ti/omap4-aah/camera/ |
D | AppCallbackNotifier.cpp | 702 static void copyCroppedNV12(CameraFrame* frame, unsigned char *dst) in copyCroppedNV12() argument 708 CAMHAL_ASSERT(frame && dst); in copyCroppedNV12() 710 offset = frame->mOffset; in copyCroppedNV12() 711 stride = frame->mAlignment; in copyCroppedNV12() 712 width = frame->mWidth; in copyCroppedNV12() 713 height = frame->mHeight; in copyCroppedNV12() 714 size = frame->mLength; in copyCroppedNV12() 715 unsigned const char *src = (unsigned char *) frame->mBuffer->mapped; in copyCroppedNV12() 740 void AppCallbackNotifier::copyAndSendPictureFrame(CameraFrame* frame, int32_t msgType) in copyAndSendPictureFrame() argument 753 if (frame->mBuffer->format && in copyAndSendPictureFrame() [all …]
|
D | BaseCameraAdapter.cpp | 309 CameraFrame *frame = new CameraFrame; in addFramePointers() local 310 frame->mBuffer = frameBuf; in addFramePointers() 311 frame->mYuv[0] = pBuf[0]; in addFramePointers() 312 frame->mYuv[1] = pBuf[1]; in addFramePointers() 313 mFrameQueue.add(frameBuf, frame); in addFramePointers() 315 … CAMHAL_LOGVB("Adding Frame=0x%x Y=0x%x UV=0x%x", frame->mBuffer, frame->mYuv[0], frame->mYuv[1]); in addFramePointers() 327 CameraFrame *frame = (CameraFrame *)mFrameQueue.valueAt(i); in removeFramePointers() local 328 … CAMHAL_LOGVB("Free Frame=0x%x Y=0x%x UV=0x%x", frame->mBuffer, frame->mYuv[0], frame->mYuv[1]); in removeFramePointers() 329 delete frame; in removeFramePointers() 412 CameraFrame *frame = NULL; in sendCommand() local [all …]
|
/hardware/ti/omap4xxx/camera/ |
D | AppCallbackNotifier.cpp | 690 void AppCallbackNotifier::copyAndSendPictureFrame(CameraFrame* frame, int32_t msgType) in copyAndSendPictureFrame() argument 703 picture = mRequestMemory(-1, frame->mLength, 1, NULL); in copyAndSendPictureFrame() 708 src = (void *) ((unsigned int) frame->mBuffer + frame->mOffset); in copyAndSendPictureFrame() 709 memcpy(dest, src, frame->mLength); in copyAndSendPictureFrame() 715 mFrameProvider->returnFrame(frame->mBuffer, (CameraFrame::FrameType) frame->mFrameType); in copyAndSendPictureFrame() 726 void AppCallbackNotifier::copyAndSendPreviewFrame(CameraFrame* frame, int32_t msgType) in copyAndSendPreviewFrame() argument 739 if (!mPreviewMemory || !frame->mBuffer) { in copyAndSendPreviewFrame() 750 frame->mBuffer, in copyAndSendPreviewFrame() 751 frame->mWidth, in copyAndSendPreviewFrame() 752 frame->mHeight, in copyAndSendPreviewFrame() [all …]
|
D | BaseCameraAdapter.cpp | 230 CameraFrame *frame = new CameraFrame; in addFramePointers() local 231 frame->mBuffer = frameBuf; in addFramePointers() 232 frame->mYuv[0] = pBuf[0]; in addFramePointers() 233 frame->mYuv[1] = pBuf[1]; in addFramePointers() 234 mFrameQueue.add(frameBuf, frame); in addFramePointers() 236 … CAMHAL_LOGVB("Adding Frame=0x%x Y=0x%x UV=0x%x", frame->mBuffer, frame->mYuv[0], frame->mYuv[1]); in addFramePointers() 248 CameraFrame *frame = (CameraFrame *)mFrameQueue.valueAt(i); in removeFramePointers() local 249 … CAMHAL_LOGVB("Free Frame=0x%x Y=0x%x UV=0x%x", frame->mBuffer, frame->mYuv[0], frame->mYuv[1]); in removeFramePointers() 250 delete frame; in removeFramePointers() 332 CameraFrame *frame = NULL; in sendCommand() local [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/test/ |
D | superframe_test.cc | 39 if (video->frame() == 1) { in PreEncodeFrameHook() 49 const uint8_t *buffer = reinterpret_cast<uint8_t*>(pkt->data.frame.buf); in MutateEncoderOutputHook() 50 const uint8_t marker = buffer[pkt->data.frame.sz - 1]; in MutateEncoderOutputHook() 55 pkt->data.frame.sz >= index_sz && in MutateEncoderOutputHook() 56 buffer[pkt->data.frame.sz - index_sz] == marker) { in MutateEncoderOutputHook() 60 modified_buf_ = new uint8_t[pkt->data.frame.sz - index_sz]; in MutateEncoderOutputHook() 61 memcpy(modified_buf_, pkt->data.frame.buf, in MutateEncoderOutputHook() 62 pkt->data.frame.sz - index_sz); in MutateEncoderOutputHook() 64 modified_pkt_.data.frame.buf = modified_buf_; in MutateEncoderOutputHook() 65 modified_pkt_.data.frame.sz -= index_sz; in MutateEncoderOutputHook() [all …]
|
D | resize_test.cc | 72 pts = pkt->data.frame.pts; in write_ivf_frame_header() 73 mem_put_le32(header, static_cast<unsigned int>(pkt->data.frame.sz)); in write_ivf_frame_header() 84 unsigned int ScaleForFrameNumber(unsigned int frame, unsigned int val) { in ScaleForFrameNumber() argument 85 if (frame < 10) in ScaleForFrameNumber() 87 if (frame < 20) in ScaleForFrameNumber() 89 if (frame < 30) in ScaleForFrameNumber() 91 if (frame < 40) in ScaleForFrameNumber() 93 if (frame < 50) in ScaleForFrameNumber() 151 const unsigned int frame = static_cast<unsigned>(info->pts); in TEST_P() local 152 const unsigned int expected_w = ScaleForFrameNumber(frame, kInitialWidth); in TEST_P() [all …]
|
D | datarate_test.cc | 52 vpx_codec_pts_t duration = pkt->data.frame.pts - last_pts_; in FramePktHook() 70 const bool key_frame = (pkt->data.frame.flags & VPX_FRAME_IS_KEY) in FramePktHook() 74 << pkt->data.frame.pts; in FramePktHook() 77 const size_t frame_size_in_bits = pkt->data.frame.sz * 8; in FramePktHook() 90 last_pts_ = pkt->data.frame.pts; in FramePktHook() 277 if (video->frame() == 1) { in PreEncodeFrameHook() 281 if (video->frame() == 1) { in PreEncodeFrameHook() 286 frame_flags_ = SetFrameFlags(video->frame(), cfg_.ts_number_layers); in PreEncodeFrameHook() 287 layer_id.temporal_layer_id = SetLayerId(video->frame(), in PreEncodeFrameHook() 289 if (video->frame() > 0) { in PreEncodeFrameHook() [all …]
|
/hardware/qcom/camera/QCamera2/stack/mm-lib2d-interface/src/ |
D | mm_lib2d.c | 208 p_frame->frame[0].plane_cnt = 1; in lib2d_fill_img_frame() 209 p_frame->frame[0].plane[0].plane_type = PLANE_ARGB; in lib2d_fill_img_frame() 210 p_frame->frame[0].plane[0].addr = rgb_buffer->buffer; in lib2d_fill_img_frame() 211 p_frame->frame[0].plane[0].stride = rgb_buffer->stride; in lib2d_fill_img_frame() 212 p_frame->frame[0].plane[0].length = (rgb_buffer->stride * in lib2d_fill_img_frame() 214 p_frame->frame[0].plane[0].fd = rgb_buffer->fd; in lib2d_fill_img_frame() 215 p_frame->frame[0].plane[0].height = rgb_buffer->height; in lib2d_fill_img_frame() 216 p_frame->frame[0].plane[0].width = rgb_buffer->width; in lib2d_fill_img_frame() 217 p_frame->frame[0].plane[0].offset = 0; in lib2d_fill_img_frame() 218 p_frame->frame[0].plane[0].scanline = rgb_buffer->height; in lib2d_fill_img_frame() [all …]
|
/hardware/intel/common/libmix/mix_video/src/ |
D | mixsurfacepool.c | 267 MixVideoFrame *frame = NULL; in mix_surfacepool_initialize() local 272 frame = mix_videoframe_new(); in mix_surfacepool_initialize() 274 if (frame == NULL) { in mix_surfacepool_initialize() 283 mix_videoframe_set_frame_id(frame, surfaces[i]); in mix_surfacepool_initialize() 285 mix_videoframe_set_ci_frame_idx (frame, i); in mix_surfacepool_initialize() 288 mix_videoframe_set_pool(frame, obj); in mix_surfacepool_initialize() 291 obj->free_list = g_slist_append(obj->free_list, frame); in mix_surfacepool_initialize() 316 MIX_RESULT mix_surfacepool_put(MixSurfacePool * obj, MixVideoFrame * frame) { in mix_surfacepool_put() argument 319 if (obj == NULL || frame == NULL) in mix_surfacepool_put() 322 LOG_V( "Frame id: %d\n", frame->frame_id); in mix_surfacepool_put() [all …]
|
D | mixframemanager.c | 281 MixVideoFrame *frame = NULL; in mix_framemanager_flush() local 283 frame = (MixVideoFrame *) g_ptr_array_index(fm->frame_array, in mix_framemanager_flush() 285 if (frame) { in mix_framemanager_flush() 286 mix_videoframe_unref(frame); in mix_framemanager_flush() 297 MixVideoFrame *frame = NULL; in mix_framemanager_flush() local 298 while ((frame = (MixVideoFrame *) g_queue_pop_head(fm->frame_queue))) { in mix_framemanager_flush() 299 mix_videoframe_unref(frame); in mix_framemanager_flush() 328 MixVideoFrame *frame = NULL; in get_expected_frame_from_array() local 382 frame = _frame; in get_expected_frame_from_array() 385 return frame; in get_expected_frame_from_array() [all …]
|
D | mixvideo.h | 60 MixVideoFrame ** frame); 63 MixVideoFrame * frame); 66 MixVideoRenderParams * render_params, MixVideoFrame *frame); 187 MIX_RESULT mix_video_get_frame(MixVideo * mix, MixVideoFrame ** frame); 189 MIX_RESULT mix_video_release_frame(MixVideo * mix, MixVideoFrame * frame); 192 MixVideoRenderParams * render_params, MixVideoFrame *frame);
|
/hardware/qcom/camera/QCamera2/HAL/ |
D | QCamera2HWICallbacks.cpp | 146 mm_camera_super_buf_t* frame = in zsl_channel_cb() local 148 if (frame == NULL) { in zsl_channel_cb() 153 *frame = *recvd_frame; in zsl_channel_cb() 192 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() 193 QCameraStream *pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb() 196 pMetaFrame = frame->bufs[i]; //find the metadata in zsl_channel_cb() 233 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() 234 pStream = pChannel->getStreamByHandle(frame->bufs[i]->stream_id); in zsl_channel_cb() 237 pMetaFrame = frame->bufs[i]; in zsl_channel_cb() 253 for (uint32_t i = 0; i < frame->num_bufs; i++) { in zsl_channel_cb() [all …]
|
/hardware/ti/omap4-aah/camera/inc/ |
D | BufferSourceAdapter.h | 101 void addFrame(CameraFrame *frame) { in addFrame() argument 103 mFrames.add(new CameraFrame(*frame)); in addFrame() 114 CameraFrame *frame = mFrames.itemAt(0); in requestExit() local 116 frame->mMetaData.clear(); in requestExit() 117 delete frame; in requestExit() local 123 CameraFrame *frame = NULL; in threadLoop() local 128 frame = mFrames.itemAt(0); in threadLoop() 133 if (frame) { in threadLoop() 134 mBufferSourceAdapter->handleFrameCallback(frame); in threadLoop() 135 frame->mMetaData.clear(); in threadLoop() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/examples/ |
D | decode_with_partial_drops.c | 140 void throw_packets(unsigned char* frame, int* size, int loss_rate, in throw_packets() argument 157 tmp = (frame[2] << 16) | (frame[1] << 8) | frame[0]; in throw_packets() 178 memcpy(loss_frame + loss_pos, frame + pos, pkg_size); in throw_packets() 190 memcpy(frame, loss_frame, loss_pos); in throw_packets() 191 memset(frame + loss_pos, 0, *size - loss_pos); in throw_packets() 201 unsigned char frame[256*1024]; in main() local 258 if(frame_sz > sizeof(frame)) in main() 260 if(fread(frame, 1, frame_sz, infile) != frame_sz) in main() 282 throw_packets(frame, &frame_sz, n, &thrown_frame, &kept_frame); in main() 303 if(vpx_codec_decode(&codec, frame, frame_sz, NULL, 0)) in main()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libwebm/ |
D | mkvmuxer.cpp | 141 bool Frame::Init(const uint8* frame, uint64 length) { in Init() argument 151 memcpy(frame_, frame, static_cast<size_t>(length_)); in Init() 1490 bool Cluster::AddFrame(const uint8* frame, in AddFrame() argument 1495 return DoWriteBlock(frame, in AddFrame() 1503 bool Cluster::AddFrameWithAdditional(const uint8* frame, in AddFrameWithAdditional() argument 1511 return DoWriteBlockWithAdditional(frame, in AddFrameWithAdditional() 1522 bool Cluster::AddFrameWithDiscardPadding(const uint8* frame, in AddFrameWithDiscardPadding() argument 1528 return DoWriteBlockWithDiscardPadding(frame, in AddFrameWithDiscardPadding() 1537 bool Cluster::AddMetadata(const uint8* frame, in AddMetadata() argument 1542 return DoWriteBlock(frame, in AddMetadata() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/encoder/ |
D | temporal_filter.c | 231 int frame; in vp8_temporal_filter_iterate_c() local 283 for (frame = 0; frame < frame_count; frame++) in vp8_temporal_filter_iterate_c() 285 if (cpi->frames[frame] == NULL) in vp8_temporal_filter_iterate_c() 291 if (frame == alt_ref_index) in vp8_temporal_filter_iterate_c() 305 cpi->frames[frame], in vp8_temporal_filter_iterate_c() 322 cpi->frames[frame]->y_buffer + mb_y_offset, in vp8_temporal_filter_iterate_c() 323 cpi->frames[frame]->u_buffer + mb_uv_offset, in vp8_temporal_filter_iterate_c() 324 cpi->frames[frame]->v_buffer + mb_uv_offset, in vp8_temporal_filter_iterate_c() 325 cpi->frames[frame]->y_stride, in vp8_temporal_filter_iterate_c() 433 int frame = 0; in vp8_temporal_filter_prepare_c() local [all …]
|
/hardware/intel/common/libmix/mix_vbp/viddec_fw/fw/codecs/h264/parser/ |
D | h264parse_dpb.c | 113 if(active_fs->frame.used_for_reference) in h264_dpb_insert_ref_lists() 455 if((viddec_h264_get_is_used(active_fs) == 3)&&(active_fs->frame.used_for_reference == 3)) in h264_dpb_update_ref_lists() 462 active_fs->frame.pic_num = active_fs->frame_num_wrap; in h264_dpb_update_ref_lists() 468 list_sort_number[list0idx] = active_fs->frame.pic_num; in h264_dpb_update_ref_lists() 487 …s) == 3) && (viddec_h264_get_is_long_term(active_fs) == 3) && (active_fs->frame.used_for_reference… in h264_dpb_update_ref_lists() 489 active_fs->frame.long_term_pic_num = active_fs->frame.long_term_frame_idx; in h264_dpb_update_ref_lists() 494 list_sort_number[list0idx-p_dpb->listXsize[0]] = active_fs->frame.long_term_pic_num; in h264_dpb_update_ref_lists() 526 if (active_fs->frame.used_for_reference) in h264_dpb_update_ref_lists() 534 if ((active_fs->frame.used_for_reference)&0x1) { in h264_dpb_update_ref_lists() 538 if ((active_fs->frame.used_for_reference)&0x2) { in h264_dpb_update_ref_lists() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ |
D | debugmodes.c | 16 void vp8_print_modes_and_motion_vectors(MODE_INFO *mi, int rows, int cols, int frame) in vp8_print_modes_and_motion_vectors() argument 26 fprintf(mvs, "Mb Modes for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors() 45 fprintf(mvs, "Mb mv ref for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors() 65 fprintf(mvs, "UV Modes for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors() 85 fprintf(mvs, "Mbs for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors() 113 fprintf(mvs, "MVs for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors() 133 fprintf(mvs, "MVs for Frame %d\n", frame); in vp8_print_modes_and_motion_vectors()
|
/hardware/ti/omap4-aah/camera/V4LCameraAdapter/ |
D | V4LCameraAdapter.cpp | 541 CameraFrame frame; in takePicture() local 636 frame.mFrameType = CameraFrame::IMAGE_FRAME; in takePicture() 637 frame.mBuffer = buffer; in takePicture() 638 frame.mLength = yuv422i_buff_size; in takePicture() 639 frame.mWidth = width; in takePicture() 640 frame.mHeight = height; in takePicture() 641 frame.mAlignment = width*2; in takePicture() 642 frame.mOffset = 0; in takePicture() 643 frame.mTimestamp = systemTime(SYSTEM_TIME_MONOTONIC); in takePicture() 644 frame.mFrameMask = (unsigned int)CameraFrame::IMAGE_FRAME; in takePicture() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
D | vp9_temporal_filter.c | 196 int frame; in temporal_filter_iterate_c() local 252 for (frame = 0; frame < frame_count; frame++) { in temporal_filter_iterate_c() 253 if (cpi->frames[frame] == NULL) in temporal_filter_iterate_c() 259 if (frame == alt_ref_index) { in temporal_filter_iterate_c() 271 cpi->frames[frame]->y_buffer + mb_y_offset, in temporal_filter_iterate_c() 272 cpi->frames[frame]->y_stride); in temporal_filter_iterate_c() 285 cpi->frames[frame]->y_buffer + mb_y_offset, in temporal_filter_iterate_c() 286 cpi->frames[frame]->u_buffer + mb_uv_offset, in temporal_filter_iterate_c() 287 cpi->frames[frame]->v_buffer + mb_uv_offset, in temporal_filter_iterate_c() 288 cpi->frames[frame]->y_stride, in temporal_filter_iterate_c() [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/ |
D | usage_dx.dox | 6 frame of data. Packets \ref MUST be passed in decode order. If the 7 application wishes to associate some data with the frame, the 10 spend working on the frame. This is typically used to support adaptive 18 There are two methods for the application to access decoded frame data. Some 24 frame-based and slice-based variants. Frame based callbacks conform to the 26 frame has been decoded. Slice based callbacks conform to the signature of 27 #vpx_codec_put_slice_cb_fn_t and are invoked after a subsection of the frame 32 assembled into a full frame. For this reason, the application \ref MUST 50 Postprocessing is a process that is applied after a frame is decoded to 52 compression process. It is not required to properly decode the frame, and [all …]
|
/hardware/intel/common/libva/test/basic/ |
D | testplan.txt | 60 - Render single MPEG2 I-frame 64 - Render single MPEG2 I-frame, multiple buffer submission 68 - Render single MPEG2 I-frame, split buffers 72 - Render single MPEG2 I-frame, split buffers 77 - Render single MPEG2 I-frame, then check vaQuerySurfaceStatus, vaSyncSurface and vaQuerySurfaceSta… 89 - Render single MPEG2 I-frame, copy surface data to image, check resulting 95 - Render single MPEG2 I-frame, copy half of a VAImage to surface, copy
|
/hardware/qcom/camera/QCamera2/stack/mm-camera-test/src/ |
D | mm_qcamera_rdi.c | 39 static void mm_app_rdi_dump_frame(mm_camera_buf_def_t *frame, in mm_app_rdi_dump_frame() argument 48 if (frame != NULL) { in mm_app_rdi_dump_frame() 55 for (i = 0; i < frame->planes_buf.num_planes; i++) { in mm_app_rdi_dump_frame() 57 (uint8_t *)frame->buffer + frame->planes_buf.planes[i].data_offset, in mm_app_rdi_dump_frame() 71 mm_camera_buf_def_t *frame = bufs->bufs[0]; in mm_app_rdi_notify_cb() local 75 frame->frame_len, frame->frame_idx, frame->stream_id); in mm_app_rdi_notify_cb() 77 mm_app_rdi_dump_frame(frame, file_name, "raw", frame->frame_idx); in mm_app_rdi_notify_cb() 81 frame)) { in mm_app_rdi_notify_cb() 84 mm_app_cache_ops((mm_camera_app_meminfo_t *)frame->mem_info, in mm_app_rdi_notify_cb()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/decoder/ |
D | vp9_decoder.c | 70 void write_dx_frame_to_file(YV12_BUFFER_CONFIG *frame, int this_frame) { in write_dx_frame_to_file() argument 79 for (i = 0; i < frame->y_height; i++) in write_dx_frame_to_file() 80 fwrite(frame->y_buffer + i * frame->y_stride, in write_dx_frame_to_file() 81 frame->y_width, 1, yframe); in write_dx_frame_to_file() 87 for (i = 0; i < frame->uv_height; i++) in write_dx_frame_to_file() 88 fwrite(frame->u_buffer + i * frame->uv_stride, in write_dx_frame_to_file() 89 frame->uv_width, 1, yframe); in write_dx_frame_to_file() 95 for (i = 0; i < frame->uv_height; i++) in write_dx_frame_to_file() 96 fwrite(frame->v_buffer + i * frame->uv_stride, in write_dx_frame_to_file() 97 frame->uv_width, 1, yframe); in write_dx_frame_to_file()
|
/hardware/qcom/camera/QCamera2/HAL3/ |
D | QCamera3Stream.cpp | 504 int32_t QCamera3Stream::processDataNotify(mm_camera_super_buf_t *frame) in processDataNotify() argument 508 if (mDataQ.enqueue((void *)frame)) { in processDataNotify() 512 bufDone(frame->bufs[0]->buf_idx); in processDataNotify() 513 free(frame); in processDataNotify() 545 mm_camera_super_buf_t *frame = in dataNotifyCB() local 547 if (frame == NULL) { in dataNotifyCB() 552 *frame = *recvd_frame; in dataNotifyCB() 553 stream->processDataNotify(frame); in dataNotifyCB() 593 mm_camera_super_buf_t *frame = in dataProcRoutine() local 595 if (NULL != frame) { in dataProcRoutine() [all …]
|