Home
last modified time | relevance | path

Searched refs:lanes (Results 1 – 25 of 45) sorted by relevance

12

/external/gemmlowp/meta/generators/
Dzip_Nx8_neon.py41 lanes = []
45 lanes.append(ZipLane(input_address,
50 lanes.append(ZipLane(address_register,
55 return lanes
67 def GenerateClearAggregators(emitter, lanes): argument
68 for lane in lanes:
72 def GenerateLoadAggregateStore(emitter, lanes, output_address, alignment): argument
77 for lane in lanes:
83 for lane in lanes:
92 emitter, leftovers, lanes, output_address): argument
[all …]
Dqnt_Nx8_neon.py26 def BuildName(lanes, leftovers, aligned): argument
27 name = 'qnt_%dx8' % lanes
35 def LoadAndDuplicateOffsets(emitter, registers, lanes, offsets): argument
36 if lanes == 1 or lanes == 2 or lanes == 3:
38 for unused_i in range(0, lanes):
47 raise ConfigurationError('Unsupported number of lanes: %d' % lanes)
62 lanes = []
67 lanes.append(QntLane(source,
75 lanes.append(QntLane(input_register,
84 return lanes
[all …]
Dmul_Nx8_Mx8_neon.py22 self.lanes = []
25 self.lanes.append(lane)
28 for i in range(0, len(self.lanes)):
29 registers.FreeRegister(self.lanes[i])
30 self.lanes[i] = None
34 lanes = MulLanes(address)
36 lanes.AddLane(registers.DoubleRegister())
37 return lanes
41 lanes = MulLanes(address)
42 lanes.AddLane(registers.Low(quad_register))
[all …]
/external/v8/src/runtime/
Druntime-simd.cc173 #define CONVERT_SIMD_LANE_ARG_CHECKED(name, index, lanes) \ argument
180 if (number < 0 || number >= lanes || !IsInt32Double(number)) { \
199 lane_type lanes[kLaneCount]; \
201 lanes[i] = op(a->get_lane(i)); \
203 Handle<type> result = isolate->factory()->New##type(lanes);
210 lane_type lanes[kLaneCount]; \
212 lanes[i] = op(a->get_lane(i), b->get_lane(i)); \
214 Handle<type> result = isolate->factory()->New##type(lanes);
221 bool lanes[kLaneCount]; \
223 lanes[i] = a->get_lane(i) op b->get_lane(i); \
[all …]
/external/llvm/test/CodeGen/AMDGPU/
Ddetect-dead-lanes.mir1 # RUN: llc -march=amdgcn -run-pass detect-dead-lanes -o - %s | FileCheck %s
53 # Check defined lanes transfer; Includes checking for some special cases like
134 # Check used lanes transfer; Includes checking for some special cases like
216 # Check that copies to physregs use all lanes, copies from physregs define all
217 # lanes. So we should not get a dead/undef flag here.
315 ; let's swiffle some lanes around for fun...
327 # for the used lanes. The example reads sub3 lane at the end, however with each
371 ; rotate lanes, but skip sub2 lane...
381 # Similar to loop1 test, but check for fixpoint of defined lanes.
417 ; rotate subreg lanes, skipping sub1
/external/llvm/lib/Target/AArch64/
DAArch64SchedVulcan.td669 // ASIMD load, 1 element, all lanes, D-form, B/H/S
670 // ASIMD load, 1 element, all lanes, D-form, D
671 // ASIMD load, 1 element, all lanes, Q-form
691 // ASIMD load, 2 element, all lanes, D-form, B/H/S
692 // ASIMD load, 2 element, all lanes, D-form, D
693 // ASIMD load, 2 element, all lanes, Q-form
714 // ASIMD load, 3 element, all lanes, D-form, B/H/S
715 // ASIMD load, 3 element, all lanes, D-form, D
716 // ASIMD load, 3 element, all lanes, Q-form, B/H/S
717 // ASIMD load, 3 element, all lanes, Q-form, D
[all …]
DAArch64RegisterInfo.td474 class TypedVecListAsmOperand<int count, int regsize, int lanes, string kind>
476 let Name = "TypedVectorList" # count # "_" # lanes # kind;
479 = "isTypedVectorList<" # count # ", " # lanes # ", '" # kind # "'>";
483 class TypedVecListRegOperand<RegisterClass Reg, int lanes, string kind>
484 : RegisterOperand<Reg, "printTypedVectorList<" # lanes # ", '"
DAArch64Schedule.td97 // Read the unwritten lanes of the VLD's destination registers.
DAArch64CallingConvention.td30 // their lanes are in a consistent order.
95 // their lanes are in a consistent order.
/external/swiftshader/third_party/LLVM/test/MC/Disassembler/ARM/
Dinvalid-VLD3DUPd32_UPD-thumb.txt9 # A8.6.315 VLD3 (single 3-element structure to all lanes)
/external/vixl/src/aarch64/
Doperands-aarch64.h310 VRegister(unsigned code, unsigned size, unsigned lanes = 1)
311 : CPURegister(code, size, kVRegister), lanes_(lanes) { in CPURegister()
401 VIXL_DEPRECATED("GetLanes", int lanes() const) { return GetLanes(); } in lanes() function
Dlogic-aarch64.cc1384 int lanes = LaneCountFromFormat(vform); in sminmaxp() local
1388 for (int i = 0; i < lanes; i += 2) { in sminmaxp()
1397 VIXL_ASSERT(((i >> 1) + (j * lanes / 2)) < kMaxLanesPerVector); in sminmaxp()
1398 result[(i >> 1) + (j * lanes / 2)] = dst_val; in sminmaxp()
1564 int lanes = LaneCountFromFormat(vform); in uminmaxp() local
1577 VIXL_ASSERT(((i >> 1) + (j * lanes / 2)) < kMaxLanesPerVector); in uminmaxp()
1578 result[(i >> 1) + (j * lanes / 2)] = dst_val; in uminmaxp()
/external/mesa3d/src/gallium/drivers/nv50/codegen/
Dnv50_ir_emit_nv50.cpp602 code[1] = 0x00200000 | (i->lanes << 14); in emitLOAD()
619 code[1] = 0x00200000 | (i->lanes << 14); in emitLOAD()
759 code[1] |= (i->lanes << 14); in emitMOV()
1690 emitQUADOP(insn, insn->lanes, insn->subOp); in emitInstruction()
1759 if (i->join || i->lanes != 0xf || i->exit) in getMinEncodingSize()
Dnv50_ir.cpp573 lanes = 0xf; in init()
737 i->lanes = lanes; in clone()
Dnv50_ir_build_util.cpp265 quadop->lanes = l; in mkQuadop()
Dnv50_ir.h696 unsigned lanes : 4; variable
/external/llvm/test/CodeGen/X86/
Dvshift-4.ll16 ; shift1b can't use a packed shift but can shift lanes separately and shuffle back together
/external/clang/include/clang/Basic/
Darm_neon.td78 // - "H" - Halve the number of lanes in the type.
79 // - "D" - Double the number of lanes in the type.
94 // all lanes. The type of the vector is the base type of the intrinsic.
159 // is a width in bits to reverse. The lanes this maps to is determined
164 // mask0 - The initial sequence of lanes for shuffle ARG0
166 // mask0 - The initial sequence of lanes for shuffle ARG1
661 // E.3.16 Extract lanes from a vector
667 // E.3.17 Set lanes within a vector
679 // E.3.19 Set all lanes to same value
1115 // Set all lanes to same value
/external/libhevc/common/arm/
Dihevc_intra_pred_luma_dc.s454 vdup.16 q12, d11[0] @3*dc + 2 (moved to all lanes)
/external/libvpx/libvpx/vpx_dsp/x86/
Dquantize_avx_x86_64.asm61 pcmpeqw m4, m4 ; All word lanes -1
208 pcmpeqw m4, m4 ; All lanes -1
/external/llvm/docs/
DBigEndianNEON.rst66 Big endian vector load using ``LD1``. Note that the lanes retain the correct ordering.
109 … ``ST1``. ``LDR`` and ``STR`` are oblivious to the size of the individual lanes of a vector. ``LD1…
/external/llvm/test/CodeGen/ARM/
Dvdiv_combine.ll139 ; Don't combine with 8 lanes. Just make sure things don't crash.
/external/vixl/doc/aarch64/
Dsupported-instructions-aarch64.md2557 One-element single structure load to all lanes.
2584 Two-element single structure load to all lanes.
2614 Three-element single structure load to all lanes.
2647 Four-element single structure load to all lanes.
3818 Two-element single structure store from two lanes.
3837 Three-element single structure store from three lanes.
3858 Four-element single structure store from four lanes.
/external/llvm/test/MC/Disassembler/ARM/
Dinvalid-thumbv7.txt259 # A8.6.315 VLD3 (single 3-element structure to all lanes)
/external/mesa3d/src/gallium/drivers/nvc0/codegen/
Dnv50_ir_emit_nvc0.cpp1508 opc |= i->lanes << 5; in emitMOV()
1720 emitQUADOP(insn, insn->subOp, insn->lanes); in emitInstruction()
1780 if (i->op == OP_MOV && i->lanes != 0xf) { in getMinEncodingSize()

12