/external/gemmlowp/meta/generators/ |
D | zip_Nx8_neon.py | 41 lanes = [] 45 lanes.append(ZipLane(input_address, 50 lanes.append(ZipLane(address_register, 55 return lanes 67 def GenerateClearAggregators(emitter, lanes): argument 68 for lane in lanes: 72 def GenerateLoadAggregateStore(emitter, lanes, output_address, alignment): argument 77 for lane in lanes: 83 for lane in lanes: 92 emitter, leftovers, lanes, output_address): argument [all …]
|
D | qnt_Nx8_neon.py | 26 def BuildName(lanes, leftovers, aligned): argument 27 name = 'qnt_%dx8' % lanes 35 def LoadAndDuplicateOffsets(emitter, registers, lanes, offsets): argument 36 if lanes == 1 or lanes == 2 or lanes == 3: 38 for unused_i in range(0, lanes): 47 raise ConfigurationError('Unsupported number of lanes: %d' % lanes) 62 lanes = [] 67 lanes.append(QntLane(source, 75 lanes.append(QntLane(input_register, 84 return lanes [all …]
|
D | mul_Nx8_Mx8_neon.py | 22 self.lanes = [] 25 self.lanes.append(lane) 28 for i in range(0, len(self.lanes)): 29 registers.FreeRegister(self.lanes[i]) 30 self.lanes[i] = None 34 lanes = MulLanes(address) 36 lanes.AddLane(registers.DoubleRegister()) 37 return lanes 41 lanes = MulLanes(address) 42 lanes.AddLane(registers.Low(quad_register)) [all …]
|
/external/v8/src/runtime/ |
D | runtime-simd.cc | 173 #define CONVERT_SIMD_LANE_ARG_CHECKED(name, index, lanes) \ argument 180 if (number < 0 || number >= lanes || !IsInt32Double(number)) { \ 199 lane_type lanes[kLaneCount]; \ 201 lanes[i] = op(a->get_lane(i)); \ 203 Handle<type> result = isolate->factory()->New##type(lanes); 210 lane_type lanes[kLaneCount]; \ 212 lanes[i] = op(a->get_lane(i), b->get_lane(i)); \ 214 Handle<type> result = isolate->factory()->New##type(lanes); 221 bool lanes[kLaneCount]; \ 223 lanes[i] = a->get_lane(i) op b->get_lane(i); \ [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | detect-dead-lanes.mir | 1 # RUN: llc -march=amdgcn -run-pass detect-dead-lanes -o - %s | FileCheck %s 53 # Check defined lanes transfer; Includes checking for some special cases like 134 # Check used lanes transfer; Includes checking for some special cases like 216 # Check that copies to physregs use all lanes, copies from physregs define all 217 # lanes. So we should not get a dead/undef flag here. 315 ; let's swiffle some lanes around for fun... 327 # for the used lanes. The example reads sub3 lane at the end, however with each 371 ; rotate lanes, but skip sub2 lane... 381 # Similar to loop1 test, but check for fixpoint of defined lanes. 417 ; rotate subreg lanes, skipping sub1
|
/external/llvm/lib/Target/AArch64/ |
D | AArch64SchedVulcan.td | 669 // ASIMD load, 1 element, all lanes, D-form, B/H/S 670 // ASIMD load, 1 element, all lanes, D-form, D 671 // ASIMD load, 1 element, all lanes, Q-form 691 // ASIMD load, 2 element, all lanes, D-form, B/H/S 692 // ASIMD load, 2 element, all lanes, D-form, D 693 // ASIMD load, 2 element, all lanes, Q-form 714 // ASIMD load, 3 element, all lanes, D-form, B/H/S 715 // ASIMD load, 3 element, all lanes, D-form, D 716 // ASIMD load, 3 element, all lanes, Q-form, B/H/S 717 // ASIMD load, 3 element, all lanes, Q-form, D [all …]
|
D | AArch64RegisterInfo.td | 474 class TypedVecListAsmOperand<int count, int regsize, int lanes, string kind> 476 let Name = "TypedVectorList" # count # "_" # lanes # kind; 479 = "isTypedVectorList<" # count # ", " # lanes # ", '" # kind # "'>"; 483 class TypedVecListRegOperand<RegisterClass Reg, int lanes, string kind> 484 : RegisterOperand<Reg, "printTypedVectorList<" # lanes # ", '"
|
D | AArch64Schedule.td | 97 // Read the unwritten lanes of the VLD's destination registers.
|
D | AArch64CallingConvention.td | 30 // their lanes are in a consistent order. 95 // their lanes are in a consistent order.
|
/external/swiftshader/third_party/LLVM/test/MC/Disassembler/ARM/ |
D | invalid-VLD3DUPd32_UPD-thumb.txt | 9 # A8.6.315 VLD3 (single 3-element structure to all lanes)
|
/external/vixl/src/aarch64/ |
D | operands-aarch64.h | 310 VRegister(unsigned code, unsigned size, unsigned lanes = 1) 311 : CPURegister(code, size, kVRegister), lanes_(lanes) { in CPURegister() 401 VIXL_DEPRECATED("GetLanes", int lanes() const) { return GetLanes(); } in lanes() function
|
D | logic-aarch64.cc | 1384 int lanes = LaneCountFromFormat(vform); in sminmaxp() local 1388 for (int i = 0; i < lanes; i += 2) { in sminmaxp() 1397 VIXL_ASSERT(((i >> 1) + (j * lanes / 2)) < kMaxLanesPerVector); in sminmaxp() 1398 result[(i >> 1) + (j * lanes / 2)] = dst_val; in sminmaxp() 1564 int lanes = LaneCountFromFormat(vform); in uminmaxp() local 1577 VIXL_ASSERT(((i >> 1) + (j * lanes / 2)) < kMaxLanesPerVector); in uminmaxp() 1578 result[(i >> 1) + (j * lanes / 2)] = dst_val; in uminmaxp()
|
/external/mesa3d/src/gallium/drivers/nv50/codegen/ |
D | nv50_ir_emit_nv50.cpp | 602 code[1] = 0x00200000 | (i->lanes << 14); in emitLOAD() 619 code[1] = 0x00200000 | (i->lanes << 14); in emitLOAD() 759 code[1] |= (i->lanes << 14); in emitMOV() 1690 emitQUADOP(insn, insn->lanes, insn->subOp); in emitInstruction() 1759 if (i->join || i->lanes != 0xf || i->exit) in getMinEncodingSize()
|
D | nv50_ir.cpp | 573 lanes = 0xf; in init() 737 i->lanes = lanes; in clone()
|
D | nv50_ir_build_util.cpp | 265 quadop->lanes = l; in mkQuadop()
|
D | nv50_ir.h | 696 unsigned lanes : 4; variable
|
/external/llvm/test/CodeGen/X86/ |
D | vshift-4.ll | 16 ; shift1b can't use a packed shift but can shift lanes separately and shuffle back together
|
/external/clang/include/clang/Basic/ |
D | arm_neon.td | 78 // - "H" - Halve the number of lanes in the type. 79 // - "D" - Double the number of lanes in the type. 94 // all lanes. The type of the vector is the base type of the intrinsic. 159 // is a width in bits to reverse. The lanes this maps to is determined 164 // mask0 - The initial sequence of lanes for shuffle ARG0 166 // mask0 - The initial sequence of lanes for shuffle ARG1 661 // E.3.16 Extract lanes from a vector 667 // E.3.17 Set lanes within a vector 679 // E.3.19 Set all lanes to same value 1115 // Set all lanes to same value
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_luma_dc.s | 454 vdup.16 q12, d11[0] @3*dc + 2 (moved to all lanes)
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | quantize_avx_x86_64.asm | 61 pcmpeqw m4, m4 ; All word lanes -1 208 pcmpeqw m4, m4 ; All lanes -1
|
/external/llvm/docs/ |
D | BigEndianNEON.rst | 66 Big endian vector load using ``LD1``. Note that the lanes retain the correct ordering. 109 … ``ST1``. ``LDR`` and ``STR`` are oblivious to the size of the individual lanes of a vector. ``LD1…
|
/external/llvm/test/CodeGen/ARM/ |
D | vdiv_combine.ll | 139 ; Don't combine with 8 lanes. Just make sure things don't crash.
|
/external/vixl/doc/aarch64/ |
D | supported-instructions-aarch64.md | 2557 One-element single structure load to all lanes. 2584 Two-element single structure load to all lanes. 2614 Three-element single structure load to all lanes. 2647 Four-element single structure load to all lanes. 3818 Two-element single structure store from two lanes. 3837 Three-element single structure store from three lanes. 3858 Four-element single structure store from four lanes.
|
/external/llvm/test/MC/Disassembler/ARM/ |
D | invalid-thumbv7.txt | 259 # A8.6.315 VLD3 (single 3-element structure to all lanes)
|
/external/mesa3d/src/gallium/drivers/nvc0/codegen/ |
D | nv50_ir_emit_nvc0.cpp | 1508 opc |= i->lanes << 5; in emitMOV() 1720 emitQUADOP(insn, insn->subOp, insn->lanes); in emitInstruction() 1780 if (i->op == OP_MOV && i->lanes != 0xf) { in getMinEncodingSize()
|