/hardware/google/aemu/base/include/aemu/base/ |
D | address_space.h | 55 struct address_block *blocks; member 182 address_space_assert(size < allocator->blocks[i].size); in address_space_allocator_split_block() 186 allocator->blocks = in address_space_allocator_split_block() 189 allocator->blocks, in address_space_allocator_split_block() 191 address_space_assert(allocator->blocks); in address_space_allocator_split_block() 195 struct address_block *blocks = allocator->blocks; in address_space_allocator_split_block() local 201 memmove(&blocks[i + 2], &blocks[i + 1], in address_space_allocator_split_block() 204 struct address_block *to_borrow_from = &blocks[i]; in address_space_allocator_split_block() 249 address_space_assert(size < allocator->blocks[i].size); in address_space_allocator_split_block_at_offset() 250 address_space_assert(offset >= allocator->blocks[i].offset); in address_space_allocator_split_block_at_offset() [all …]
|
/hardware/google/gfxstream/guest/android-emu/aemu/base/ |
D | address_space.h | 55 struct address_block *blocks; member 156 address_space_assert(size < allocator->blocks[i].size); in address_space_allocator_split_block() 160 allocator->blocks = in address_space_allocator_split_block() 163 allocator->blocks, in address_space_allocator_split_block() 165 address_space_assert(allocator->blocks); in address_space_allocator_split_block() 169 struct address_block *blocks = allocator->blocks; in address_space_allocator_split_block() local 175 memmove(&blocks[i + 2], &blocks[i + 1], in address_space_allocator_split_block() 178 struct address_block *to_borrow_from = &blocks[i]; in address_space_allocator_split_block() 207 struct address_block *blocks = allocator->blocks; in address_space_allocator_release_block() local 215 blocks[i].available = 1; in address_space_allocator_release_block() [all …]
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/scripts/Retired/ |
D | findBalance.py | 32 blocks = [] 37 blockDepth = len(blocks) 39 thisBlock = blocks[blockDepth-1] 54 blocks.pop() 59 blocks.append([ '--', line, None ]) 82 blocks.append([ 'ifdef', line, condition ]) 105 blocks.pop() 127 blockDepth = len(blocks) 131 'block type:', blocks[blockDepth-1][0])
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/config/chunkindex/ |
D | chunked.js | 144 var blocks = document.getElementsByClassName("listingblock") 146 for(var i=0; i < blocks.length; i++) { 147 if(blocks[i].id.length > 0) { 151 a.setAttribute('href', '#' + blocks[i].id); 153 blocks[i].insertBefore(a, blocks[i].childNodes[0]);
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/appendices/ |
D | VK_EXT_inline_uniform_block.adoc | 23 This extension introduces the ability to back uniform blocks directly with 46 1) Do we need a new storage class for inline uniform blocks vs. uniform 47 blocks? 51 both uniform buffers and inline uniform blocks. 66 sname:VkDescriptorUpdateTemplateEntry is ignored for inline uniform blocks 70 3) What layout rules apply for uniform blocks corresponding to inline 76 by `VK_EXT_descriptor_indexing` for inline uniform blocks? 78 *RESOLVED*: No, because inline uniform blocks are not allowed to be 84 level of support for update-after-bind inline uniform blocks. 87 by `VK_EXT_descriptor_indexing` supported for inline uniform blocks?
|
D | VK_EXT_scalar_block_layout.adoc | 22 This extension enables C-like structure layout for SPIR-V blocks.
|
D | VK_KHR_8bit_storage.adoc | 25 storage buffers, and push constant blocks.
|
D | VK_KHR_16bit_storage.adoc | 32 input and output interfaces, and push constant blocks.
|
/hardware/google/aemu/host-common/testing/ |
D | HostAddressSpace.cpp | 100 auto& mem = entry.blocks[off]; in setHostAddr() 110 for (auto &it2 : it.second.blocks) { in setHostAddrByPhysAddr() 130 for (auto &it2 : it.second.blocks) { in unsetHostAddrByPhysAddr() 160 for (const auto &it2 : it.second.blocks) { in getHostAddr() 206 if (entry.blocks.find(off) != entry.blocks.end()) { in claimShared() 219 auto& entryBlock = entry.blocks[off]; in claimShared() 227 if (entry.blocks.find(off) == entry.blocks.end()) { in unclaimShared() 233 if (!enclosingSharedRegionExists(mSharedRegions, off, entry.blocks[off].size)) { in unclaimShared() 236 (unsigned long long)off + entry.blocks[off].size); in unclaimShared() 240 entry.blocks.erase(off); in unclaimShared() [all …]
|
/hardware/google/gfxstream/guest/mesa/src/util/blake3/ |
D | blake3_dispatch.c | 195 size_t blocks, const uint32_t key[8], uint64_t counter, in blake3_hash_many() argument 203 blake3_hash_many_avx512(inputs, num_inputs, blocks, key, counter, in blake3_hash_many() 211 blake3_hash_many_avx2(inputs, num_inputs, blocks, key, counter, in blake3_hash_many() 219 blake3_hash_many_sse41(inputs, num_inputs, blocks, key, counter, in blake3_hash_many() 227 blake3_hash_many_sse2(inputs, num_inputs, blocks, key, counter, in blake3_hash_many() 236 blake3_hash_many_neon(inputs, num_inputs, blocks, key, counter, in blake3_hash_many() 241 blake3_hash_many_portable(inputs, num_inputs, blocks, key, counter, in blake3_hash_many()
|
D | blake3_impl.h | 191 size_t blocks, const uint32_t key[8], uint64_t counter, 210 size_t blocks, const uint32_t key[8], 226 size_t blocks, const uint32_t key[8], 241 size_t blocks, const uint32_t key[8], 248 size_t blocks, const uint32_t key[8], 265 size_t blocks, const uint32_t key[8], 274 size_t blocks, const uint32_t key[8],
|
D | blake3_portable.c | 125 INLINE void hash_one_portable(const uint8_t *input, size_t blocks, in hash_one_portable() argument 132 while (blocks > 0) { in hash_one_portable() 133 if (blocks == 1) { in hash_one_portable() 139 blocks -= 1; in hash_one_portable() 146 size_t blocks, const uint32_t key[8], in blake3_hash_many_portable() argument 151 hash_one_portable(inputs[0], blocks, key, counter, flags, flags_start, in blake3_hash_many_portable()
|
D | blake3_neon.c | 231 static void blake3_hash4_neon(const uint8_t *const *inputs, size_t blocks, in blake3_hash4_neon() argument 244 for (size_t block = 0; block < blocks; block++) { in blake3_hash4_neon() 245 if (block + 1 == blocks) { in blake3_hash4_neon() 303 INLINE void hash_one_neon(const uint8_t *input, size_t blocks, in hash_one_neon() argument 310 while (blocks > 0) { in hash_one_neon() 311 if (blocks == 1) { in hash_one_neon() 320 blocks -= 1; in hash_one_neon() 327 size_t blocks, const uint32_t key[8], in blake3_hash_many_neon() argument 332 blake3_hash4_neon(inputs, blocks, key, counter, increment_counter, flags, in blake3_hash_many_neon() 342 hash_one_neon(inputs[0], blocks, key, counter, flags, flags_start, in blake3_hash_many_neon()
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/config/vu-to-json/ |
D | extension.rb | 47 block.blocks.each do |list| 53 list.blocks.each do |item|
|
/hardware/interfaces/media/c2/aidl/aidl_api/android.hardware.media.c2/current/android/hardware/media/c2/ |
D | Buffer.aidl | 38 android.hardware.media.c2.Block[] blocks;
|
/hardware/interfaces/media/c2/aidl/aidl_api/android.hardware.media.c2/1/android/hardware/media/c2/ |
D | Buffer.aidl | 38 android.hardware.media.c2.Block[] blocks;
|
/hardware/interfaces/media/c2/aidl/android/hardware/media/c2/ |
D | Buffer.aidl | 36 Block[] blocks;
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/katex/contrib/ |
D | copy-tex.css | 1 /* Force selection of entire .katex/.katex-display blocks, so that we can
|
D | copy-tex.mjs | 16 // Remove .katex-html blocks that are preceded by .katex-mathml blocks
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/style/ |
D | revisions.adoc | 29 * 2023-04-30 - Update <<markup-blocks, Blocks>> to remove tilde markup for 30 source blocks inside refpages, and use `[open]`-tagged listing blocks for 31 <<markup-blocks-nested-open, continuation blocks inside refpages>> 67 tilde-delimited source blocks <<markup-blocks-source, inside reference 68 page open blocks>>, and rewrite the <<sample-command, Sample Command 158 Disallow use of standalone `+` except in latexmath and source blocks, in
|
/hardware/interfaces/bluetooth/audio/aidl/default/ |
D | A2dpOffloadCodecSbc.cpp | 226 int blocks = GetBlockLengthValue(configuration); in GetFrameSize() local 230 ((blocks * bitpool) << configuration.get(kChannelModeDualChannel)) + in GetFrameSize() 239 int blocks = GetBlockLengthValue(configuration); in GetBitrate() local 242 return (bits * sampling_frequency) / (blocks * subbands); in GetBitrate()
|
/hardware/google/interfaces/media/c2/1.0/ |
D | types.hal | 398 * NATIVE and POOLED. Each type has its own way of identifying blocks. 406 * There are two types of blocks: NATIVE and POOLED. 423 * extra information. One BaseBlock may occur in multiple blocks in one 453 vec<Block> blocks; 554 * BaseBlocks in the member `blocks` of the containing `WorkBundle`. 559 * BaseBlocks in the member `blocks` of the containing `WorkBundle`. 578 * A list of blocks indexed by elements of #works.
|
/hardware/interfaces/drm/1.0/ |
D | types.hal | 276 * A crypto Pattern is a repeating sequence of encrypted and clear blocks 284 * The number of blocks to be encrypted in the pattern. If zero, 290 * The number of blocks to be skipped (left clear) in the pattern. If
|
/hardware/google/gfxstream/codegen/vulkan/vulkan-docs-next/chapters/ |
D | sparsemem.adoc | 225 regions of the image called sparse image blocks to be bound to specific 349 used sparse blocks. 350 Any bound partially-used-sparse-blocks must: still have their full sparse 365 Each mip tail is logically a single array of sparse blocks with an 366 implementation-dependent mapping of texels or compressed texel blocks to 367 sparse blocks. 371 mip level that would contain partially used sparse blocks begins the mip 385 It is logically a single array of sparse blocks with an 386 implementation-dependent mapping of texels or compressed texel blocks to 387 sparse blocks. [all …]
|
/hardware/interfaces/tv/tuner/1.1/ |
D | types.hal | 879 * All blocks are coded as I blocks. 883 * Blocks are coded as I or P blocks. 887 * Blocks are coded as I, P or B blocks.
|