Home
last modified time | relevance | path

Searched refs:chunk_index (Results 1 – 24 of 24) sorted by relevance

/external/libchrome/base/trace_event/
Dtrace_buffer.cc83 if (handle.chunk_index >= chunks_.size()) in GetEventByHandle()
85 TraceBufferChunk* chunk = chunks_[handle.chunk_index].get(); in GetEventByHandle()
96 size_t chunk_index = recyclable_chunks_queue_[current_iteration_index_]; in NextChunk() local
98 if (chunk_index >= chunks_.size()) // Skip uninitialized chunks. in NextChunk()
100 DCHECK(chunks_[chunk_index]); in NextChunk()
101 return chunks_[chunk_index].get(); in NextChunk()
111 size_t chunk_index = recyclable_chunks_queue_[queue_index]; in EstimateTraceMemoryOverhead() local
112 if (chunk_index >= chunks_.size()) // Skip uninitialized chunks. in EstimateTraceMemoryOverhead()
114 chunks_[chunk_index]->EstimateTraceMemoryOverhead(overhead); in EstimateTraceMemoryOverhead()
200 if (handle.chunk_index >= chunks_.size()) in GetEventByHandle()
[all …]
Dtrace_log.cc149 size_t chunk_index, in MakeHandle() argument
153 DCHECK(chunk_index <= TraceBufferChunk::kMaxChunkIndex); in MakeHandle()
155 DCHECK(chunk_index <= std::numeric_limits<uint16_t>::max()); in MakeHandle()
157 handle->chunk_index = static_cast<uint16_t>(chunk_index); in MakeHandle()
210 handle.chunk_index != chunk_index_) { in GetEventByHandle()
1494 << ", chunk_index " << handle.chunk_index << ", event_index " in UpdateTraceEventDurationExplicit()
1616 DCHECK(handle.chunk_index <= TraceBufferChunk::kMaxChunkIndex); in GetEventByHandleInternal()
1632 handle.chunk_index == thread_shared_chunk_index_) { in GetEventByHandleInternal()
Dtrace_event_unittest.cc2584 size_t chunk_index; in TEST_F() local
2590 chunks[i] = buffer->GetChunk(&chunk_index).release(); in TEST_F()
2592 EXPECT_EQ(i, chunk_index); in TEST_F()
2608 chunks[i] = buffer->GetChunk(&chunk_index).release(); in TEST_F()
2610 EXPECT_EQ(i, chunk_index); in TEST_F()
2623 chunks[i] = buffer->GetChunk(&chunk_index).release(); in TEST_F()
2625 EXPECT_EQ(num_chunks - i - 1, chunk_index); in TEST_F()
2643 size_t chunk_index; in TEST_F() local
2652 chunks[i] = buffer->GetChunk(&chunk_index).release(); in TEST_F()
2654 EXPECT_EQ(i, chunk_index); in TEST_F()
[all …]
Dtrace_event_impl.h71 unsigned chunk_index : 26; member
/external/swiftshader/third_party/SPIRV-Tools/source/fuzz/
Dshrinker.cpp45 uint32_t chunk_index, uint32_t chunk_size) { in RemoveChunk() argument
46 uint32_t lower = chunk_index * chunk_size; in RemoveChunk()
47 uint32_t upper = std::min((chunk_index + 1) * chunk_size, in RemoveChunk()
180 for (int chunk_index = num_chunks - 1; in Run() local
181 attempt < step_limit_ && chunk_index >= 0; chunk_index--) { in Run()
186 static_cast<uint32_t>(chunk_index), chunk_size); in Run()
209 chunk_index * chunk_size && in Run()
/external/deqp-deps/SPIRV-Tools/source/fuzz/
Dshrinker.cpp45 uint32_t chunk_index, uint32_t chunk_size) { in RemoveChunk() argument
46 uint32_t lower = chunk_index * chunk_size; in RemoveChunk()
47 uint32_t upper = std::min((chunk_index + 1) * chunk_size, in RemoveChunk()
180 for (int chunk_index = num_chunks - 1; in Run() local
181 attempt < step_limit_ && chunk_index >= 0; chunk_index--) { in Run()
186 static_cast<uint32_t>(chunk_index), chunk_size); in Run()
209 chunk_index * chunk_size && in Run()
/external/perfetto/src/protozero/test/
Dfake_scattered_buffer.cc54 std::string FakeScatteredBuffer::GetChunkAsString(size_t chunk_index) { in GetChunkAsString() argument
55 return ToHex(chunks_[chunk_index].get(), chunk_size_); in GetChunkAsString()
61 size_t chunk_index = (start + pos) / chunk_size_; in GetBytes() local
63 buf[pos] = chunks_[chunk_index].get()[chunk_offset]; in GetBytes()
Dfake_scattered_buffer.h38 std::string GetChunkAsString(size_t chunk_index);
/external/tensorflow/tensorflow/python/distribute/v1/
Dall_reduce.py345 chunk_index = (seg_index * num_subchunks) + s
346 new_partial_reductions[chunk_index] = red_op(
347 chunks_by_dev[pred_dev][chunk_index],
348 chunks_by_dev[d][chunk_index])
354 chunk_index = (seg_index * num_subchunks) + s
355 chunks_by_dev[d][chunk_index] = new_partial_reductions[chunk_index]
409 chunk_index = (seg_index * num_subchunks) + s
410 passed_values[chunk_index] = array_ops.identity(
411 chunks_by_dev[pred_dev][chunk_index])
416 chunk_index = (seg_index * num_subchunks) + s
[all …]
/external/pigweed/pw_kvs/
Dflash_partition_test.cc52 for (size_t chunk_index = 0; chunk_index < chunks_per_sector; in WriteData() local
53 chunk_index++) { in WriteData()
69 for (size_t chunk_index = 0; chunk_index < chunks_per_sector; in WriteData() local
70 chunk_index++) { in WriteData()
/external/tensorflow/tensorflow/python/debug/lib/
Dgrpc_debug_server.py253 chunk_index = debugger_plugin_metadata["chunkIndex"]
267 chunks[chunk_index] = value.tensor
269 chunks[chunk_index] = event
306 chunk_index = int(event.graph_def[index_bar_0 + 1 : index_bar_1])
311 chunk_index] = event.graph_def[index_bar_2 + 1:]
/external/tensorflow/tensorflow/core/kernels/
Dconv_ops_using_gemm.cc330 for (int64 chunk_index = 0; chunk_index < chunk_count; ++chunk_index) { in operator ()() local
331 const int64 patch_index_start = chunk_index * patches_per_chunk; in operator ()()
Dquantized_conv_ops.cc296 for (int64 chunk_index = 0; chunk_index < chunk_count; ++chunk_index) { in operator ()() local
297 const int64 patch_index_start = chunk_index * patches_per_chunk; in operator ()()
Dtensor_cord.h155 explicit ChunkIterator(const TensorCord* cord, int chunk_index);
/external/tensorflow/tensorflow/core/debug/
Ddebugger_event_metadata.proto10 int32 chunk_index = 4; field
Ddebug_io_utils.cc63 const size_t chunk_index, in PrepareChunkEventProto() argument
85 metadata.set_chunk_index(chunk_index); in PrepareChunkEventProto()
/external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/strings/internal/str_format/
Dfloat_conversion.cc173 int chunk_index = exp / 32; in BinaryToDecimal() local
177 data_[chunk_index] = static_cast<uint32_t>(v << offset); in BinaryToDecimal()
179 data_[++chunk_index] = static_cast<uint32_t>(v); in BinaryToDecimal()
181 while (chunk_index >= 0) { in BinaryToDecimal()
186 for (int i = chunk_index; i >= 0; --i) { in BinaryToDecimal()
193 if (data_[chunk_index] == 0) --chunk_index; in BinaryToDecimal()
196 assert(decimal_start_ != chunk_index); in BinaryToDecimal()
/external/openscreen/third_party/abseil/src/absl/strings/internal/str_format/
Dfloat_conversion.cc187 int chunk_index = exp / 32; in BinaryToDecimal() local
191 data_[chunk_index] = static_cast<uint32_t>(v << offset); in BinaryToDecimal()
193 data_[++chunk_index] = static_cast<uint32_t>(v); in BinaryToDecimal()
195 while (chunk_index >= 0) { in BinaryToDecimal()
200 for (int i = chunk_index; i >= 0; --i) { in BinaryToDecimal()
207 if (data_[chunk_index] == 0) --chunk_index; in BinaryToDecimal()
210 assert(decimal_start_ != chunk_index); in BinaryToDecimal()
/external/libtextclassifier/abseil-cpp/absl/strings/internal/str_format/
Dfloat_conversion.cc173 int chunk_index = exp / 32; in BinaryToDecimal() local
177 data_[chunk_index] = static_cast<uint32_t>(v << offset); in BinaryToDecimal()
179 data_[++chunk_index] = static_cast<uint32_t>(v); in BinaryToDecimal()
181 while (chunk_index >= 0) { in BinaryToDecimal()
186 for (int i = chunk_index; i >= 0; --i) { in BinaryToDecimal()
193 if (data_[chunk_index] == 0) --chunk_index; in BinaryToDecimal()
196 assert(decimal_start_ != chunk_index); in BinaryToDecimal()
/external/mesa3d/src/amd/vulkan/
Dradv_rgp.c541 int32_t chunk_index, in radv_sqtt_fill_sqtt_desc() argument
546 chunk->header.chunk_id.index = chunk_index; in radv_sqtt_fill_sqtt_desc()
572 int32_t chunk_index, int32_t offset, int32_t size) in radv_sqtt_fill_sqtt_data() argument
575 chunk->header.chunk_id.index = chunk_index; in radv_sqtt_fill_sqtt_data()
/external/gemmlowp/meta/
Dtest_streams_correctness.cc71 int chunk_index = i * rows * 8; in check() local
75 if (result[chunk_index + j * 8 + k] != chunk_start_value + k) { in check()
/external/protobuf/src/google/protobuf/compiler/cpp/
Dcpp_message.cc2839 for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { in GenerateClear() local
2840 std::vector<const FieldDescriptor*>& chunk = chunks[chunk_index]; in GenerateClear()
2855 cold_skipper.OnStartChunk(chunk_index, cached_has_bit_index, "", printer); in GenerateClear()
2955 if (cold_skipper.OnEndChunk(chunk_index, printer)) { in GenerateClear()
3142 for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { in GenerateMergeFrom() local
3143 const std::vector<const FieldDescriptor*>& chunk = chunks[chunk_index]; in GenerateMergeFrom()
3159 cold_skipper.OnStartChunk(chunk_index, cached_has_bit_index, "from.", in GenerateMergeFrom()
3231 if (cold_skipper.OnEndChunk(chunk_index, printer)) { in GenerateMergeFrom()
4182 for (int chunk_index = 0; chunk_index < chunks.size(); chunk_index++) { in GenerateByteSize() local
4183 const std::vector<const FieldDescriptor*>& chunk = chunks[chunk_index]; in GenerateByteSize()
[all …]
/external/rust/crates/grpcio-sys/grpc/tools/interop_matrix/
Drun_interop_matrix_tests.py261 for chunk_index in range(chunk_count):
262 chunk_start = chunk_index * max_chunk_size
/external/vixl/test/aarch64/
Dtest-utils-aarch64.h199 int chunk_index = (lane * p_bits_per_lane) / kChunkSizeInBits; in preg_lane() local
201 Chunk chunk = dump_.p_[code].GetLane<Chunk>(chunk_index); in preg_lane()