/external/jemalloc_new/test/unit/ |
D | extent_quantize.c | 60 size_t lextent_size, extent_size, floor, ceil; in TEST_BEGIN() local 69 ceil = extent_size_quantize_ceil(extent_size); in TEST_BEGIN() 75 assert_zu_eq(extent_size, ceil, in TEST_BEGIN() 109 size_t extent_size, floor, ceil; in TEST_BEGIN() local 113 ceil = extent_size_quantize_ceil(extent_size); in TEST_BEGIN() 117 floor, extent_size, ceil); in TEST_BEGIN() 118 assert_zu_ge(ceil, extent_size, in TEST_BEGIN() 120 "ceil=%zu)", floor, extent_size, ceil); in TEST_BEGIN() 124 floor_prev, floor, extent_size, ceil); in TEST_BEGIN() 125 assert_zu_le(ceil_prev, ceil, "Ceiling should be monotonic " in TEST_BEGIN() [all …]
|
/external/jemalloc/test/unit/ |
D | run_quantize.c | 60 size_t lrun_size, run_size, floor, ceil; in TEST_BEGIN() local 68 ceil = run_quantize_ceil(run_size); in TEST_BEGIN() 73 assert_zu_eq(run_size, ceil, in TEST_BEGIN() 115 size_t run_size, floor, ceil; in TEST_BEGIN() local 119 ceil = run_quantize_ceil(run_size); in TEST_BEGIN() 123 floor, run_size, ceil); in TEST_BEGIN() 124 assert_zu_ge(ceil, run_size, in TEST_BEGIN() 126 floor, run_size, ceil); in TEST_BEGIN() 130 floor_prev, floor, run_size, ceil); in TEST_BEGIN() 131 assert_zu_le(ceil_prev, ceil, "Ceiling should be monotonic " in TEST_BEGIN() [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | ceil.ll | 3 declare float @llvm.ceil.f32(float) #0 4 declare double @llvm.ceil.f64(double) #0 5 declare <4 x float> @llvm.ceil.v4f32(<4 x float>) #0 10 %x = call float @llvm.ceil.f32(float 1.00) #0 17 %x = call float @llvm.ceil.f32(float 1.25) #0 24 %x = call float @llvm.ceil.f32(float -1.25) #0 31 …%x = call <4 x float> @llvm.ceil.v4f32(<4 x float> <float 1.00, float 1.25, float -1.25, float -1.… 38 %x = call double @llvm.ceil.f64(double 1.0) #0 45 %x = call double @llvm.ceil.f64(double 1.3) #0 52 %x = call double @llvm.ceil.f64(double -1.75) #0
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | ceil.ll | 3 declare float @llvm.ceil.f32(float) #0 4 declare double @llvm.ceil.f64(double) #0 5 declare <4 x float> @llvm.ceil.v4f32(<4 x float>) #0 10 %x = call float @llvm.ceil.f32(float 1.00) #0 17 %x = call float @llvm.ceil.f32(float 1.25) #0 24 %x = call float @llvm.ceil.f32(float -1.25) #0 31 …%x = call <4 x float> @llvm.ceil.v4f32(<4 x float> <float 1.00, float 1.25, float -1.25, float -1.… 38 %x = call double @llvm.ceil.f64(double 1.0) #0 45 %x = call double @llvm.ceil.f64(double 1.3) #0 52 %x = call double @llvm.ceil.f64(double -1.75) #0
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | fceil64.ll | 5 declare double @llvm.ceil.f64(double) nounwind readnone 6 declare <2 x double> @llvm.ceil.v2f64(<2 x double>) nounwind readnone 7 declare <3 x double> @llvm.ceil.v3f64(<3 x double>) nounwind readnone 8 declare <4 x double> @llvm.ceil.v4f64(<4 x double>) nounwind readnone 9 declare <8 x double> @llvm.ceil.v8f64(<8 x double>) nounwind readnone 10 declare <16 x double> @llvm.ceil.v16f64(<16 x double>) nounwind readnone 33 %y = call double @llvm.ceil.f64(double %x) nounwind readnone 42 %y = call <2 x double> @llvm.ceil.v2f64(<2 x double> %x) nounwind readnone 52 ; %y = call <3 x double> @llvm.ceil.v3f64(<3 x double> %x) nounwind readnone 63 %y = call <4 x double> @llvm.ceil.v4f64(<4 x double> %x) nounwind readnone [all …]
|
D | fceil.ll | 5 declare float @llvm.ceil.f32(float) nounwind readnone 6 declare <2 x float> @llvm.ceil.v2f32(<2 x float>) nounwind readnone 7 declare <3 x float> @llvm.ceil.v3f32(<3 x float>) nounwind readnone 8 declare <4 x float> @llvm.ceil.v4f32(<4 x float>) nounwind readnone 9 declare <8 x float> @llvm.ceil.v8f32(<8 x float>) nounwind readnone 10 declare <16 x float> @llvm.ceil.v16f32(<16 x float>) nounwind readnone 17 %y = call float @llvm.ceil.f32(float %x) nounwind readnone 29 %y = call <2 x float> @llvm.ceil.v2f32(<2 x float> %x) nounwind readnone 45 %y = call <3 x float> @llvm.ceil.v3f32(<3 x float> %x) nounwind readnone 61 %y = call <4 x float> @llvm.ceil.v4f32(<4 x float> %x) nounwind readnone [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | fceil64.ll | 5 declare double @llvm.ceil.f64(double) nounwind readnone 6 declare <2 x double> @llvm.ceil.v2f64(<2 x double>) nounwind readnone 7 declare <3 x double> @llvm.ceil.v3f64(<3 x double>) nounwind readnone 8 declare <4 x double> @llvm.ceil.v4f64(<4 x double>) nounwind readnone 9 declare <8 x double> @llvm.ceil.v8f64(<8 x double>) nounwind readnone 10 declare <16 x double> @llvm.ceil.v16f64(<16 x double>) nounwind readnone 35 %y = call double @llvm.ceil.f64(double %x) nounwind readnone 44 %y = call <2 x double> @llvm.ceil.v2f64(<2 x double> %x) nounwind readnone 54 ; %y = call <3 x double> @llvm.ceil.v3f64(<3 x double> %x) nounwind readnone 65 %y = call <4 x double> @llvm.ceil.v4f64(<4 x double> %x) nounwind readnone [all …]
|
D | fceil.ll | 5 declare float @llvm.ceil.f32(float) nounwind readnone 6 declare <2 x float> @llvm.ceil.v2f32(<2 x float>) nounwind readnone 7 declare <3 x float> @llvm.ceil.v3f32(<3 x float>) nounwind readnone 8 declare <4 x float> @llvm.ceil.v4f32(<4 x float>) nounwind readnone 9 declare <8 x float> @llvm.ceil.v8f32(<8 x float>) nounwind readnone 10 declare <16 x float> @llvm.ceil.v16f32(<16 x float>) nounwind readnone 17 %y = call float @llvm.ceil.f32(float %x) nounwind readnone 29 %y = call <2 x float> @llvm.ceil.v2f32(<2 x float> %x) nounwind readnone 45 %y = call <3 x float> @llvm.ceil.v3f32(<3 x float> %x) nounwind readnone 61 %y = call <4 x float> @llvm.ceil.v4f32(<4 x float> %x) nounwind readnone [all …]
|
/external/llvm/test/CodeGen/Mips/ |
D | call-optimization.ll | 56 ; O32-DAG: lw $25, %call16(ceil) 60 ; O32: lw $25, %call16(ceil) 64 ; O32: lw $25, %call16(ceil) 69 ; O32-LOADTGT-DAG: lw $25, %call16(ceil) 83 %call = tail call double @ceil(double %d) 84 %call1 = tail call double @ceil(double %call) 86 %call2 = tail call double @ceil(double %call1) 91 declare double @ceil(double)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/Mips/ |
D | call-optimization.ll | 56 ; O32-DAG: lw $25, %call16(ceil) 60 ; O32: lw $25, %call16(ceil) 64 ; O32: lw $25, %call16(ceil) 69 ; O32-LOADTGT-DAG: lw $25, %call16(ceil) 83 %call = tail call double @ceil(double %d) 84 %call1 = tail call double @ceil(double %call) 86 %call2 = tail call double @ceil(double %call1) 91 declare double @ceil(double)
|
/external/libchrome/ui/gfx/geometry/ |
D | safe_integer_conversions.h | 21 return base::saturated_cast<int>(std::ceil(value)); in ToCeiledInt() 29 return base::saturated_cast<int>(std::ceil(value)); in ToCeiledInt() 37 rounded = std::ceil(value - 0.5f); in ToRoundedInt() 46 rounded = std::ceil(value - 0.5); in ToRoundedInt()
|
D | rect.h | 285 std::ceil(rect.right() * x_scale))); in ScaleToEnclosingRect() 287 std::ceil(rect.bottom() * y_scale))); in ScaleToEnclosingRect() 291 x : static_cast<int>(std::ceil(rect.right() * x_scale)); in ScaleToEnclosingRect() 293 y : static_cast<int>(std::ceil(rect.bottom() * y_scale)); in ScaleToEnclosingRect() 310 int w = base::saturated_cast<int>(std::ceil(rect.width() * x_scale)); in ScaleToEnclosingRectSafe() 311 int h = base::saturated_cast<int>(std::ceil(rect.height() * y_scale)); in ScaleToEnclosingRectSafe() 325 std::ceil(rect.x() * x_scale))); in ScaleToEnclosedRect() 327 std::ceil(rect.y() * y_scale))); in ScaleToEnclosedRect() 332 int x = static_cast<int>(std::ceil(rect.x() * x_scale)); in ScaleToEnclosedRect() 333 int y = static_cast<int>(std::ceil(rect.y() * y_scale)); in ScaleToEnclosedRect()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SLPVectorizer/X86/ |
D | alternate-calls.ll | 19 ; CHECK-NEXT: [[AB0:%.*]] = call float @llvm.ceil.f32(float [[A0]]) 22 ; CHECK-NEXT: [[AB3:%.*]] = call float @llvm.ceil.f32(float [[A3]]) 23 ; CHECK-NEXT: [[AB4:%.*]] = call float @llvm.ceil.f32(float [[A4]]) 24 ; CHECK-NEXT: [[AB5:%.*]] = call float @llvm.ceil.f32(float [[A5]]) 45 %ab0 = call float @llvm.ceil.f32(float %a0) 48 %ab3 = call float @llvm.ceil.f32(float %a3) 49 %ab4 = call float @llvm.ceil.f32(float %a4) 50 %ab5 = call float @llvm.ceil.f32(float %a5) 64 declare float @llvm.ceil.f32(float)
|
/external/libcxx/test/std/utilities/time/time.duration/time.duration.cast/ |
D | ceil.pass.cpp | 30 typedef decltype(std::chrono::ceil<ToDuration>(f)) R; in test() 32 assert(std::chrono::ceil<ToDuration>(f) == d); in test() 46 …constexpr std::chrono::hours h1 = std::chrono::ceil<std::chrono::hours>(std::chrono::milliseconds(… in main() 48 …constexpr std::chrono::hours h2 = std::chrono::ceil<std::chrono::hours>(std::chrono::milliseconds(… in main()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/CostModel/X86/ |
D | fround.ll | 16 define i32 @ceil(i32 %arg) { 17 ; SSE2-LABEL: 'ceil' 18 …Model: Found an estimated cost of 10 for instruction: %F32 = call float @llvm.ceil.f32(float undef) 19 … an estimated cost of 43 for instruction: %V4F32 = call <4 x float> @llvm.ceil.v4f32(<4 x float> u… 20 … an estimated cost of 86 for instruction: %V8F32 = call <8 x float> @llvm.ceil.v8f32(<8 x float> u… 21 … estimated cost of 172 for instruction: %V16F32 = call <16 x float> @llvm.ceil.v16f32(<16 x float>… 22 …del: Found an estimated cost of 10 for instruction: %F64 = call double @llvm.ceil.f64(double undef) 23 …an estimated cost of 21 for instruction: %V2F64 = call <2 x double> @llvm.ceil.v2f64(<2 x double> … 24 …an estimated cost of 42 for instruction: %V4F64 = call <4 x double> @llvm.ceil.v4f64(<4 x double> … 25 …an estimated cost of 84 for instruction: %V8F64 = call <8 x double> @llvm.ceil.v8f64(<8 x double> … [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | no-tail-call.ll | 8 declare float @llvm.ceil.f32(float) 10 ; Check that we are not emitting a tail call for the last call to ceil. 38 %5 = call float @llvm.ceil.f32(float 5.000000e+00) 42 %7 = call float @llvm.ceil.f32(float %6) 48 ; Check that we are not emitting a tail call for the last call to ceil. 76 %5 = call float @llvm.ceil.f32(float 5.000000e+00) 80 %7 = call float @llvm.ceil.f32(float %6)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | no-tail-call.ll | 8 declare float @llvm.ceil.f32(float) 10 ; Check that we are not emitting a tail call for the last call to ceil. 38 %5 = call float @llvm.ceil.f32(float 5.000000e+00) 42 %7 = call float @llvm.ceil.f32(float %6) 48 ; Check that we are not emitting a tail call for the last call to ceil. 76 %5 = call float @llvm.ceil.f32(float 5.000000e+00) 80 %7 = call float @llvm.ceil.f32(float %6)
|
/external/libnl/python/netlink/route/qdisc/ |
D | htb.py | 65 def ceil(self): member in HTBClass 66 ceil = capi.rtnl_htb_get_ceil(self._class._rtnl_class) 67 return util.Rate(ceil) 69 @ceil.setter 70 def ceil(self, value): member in HTBClass 125 if self.rate != self.ceil:
|
/external/tensorflow/tensorflow/core/kernels/ |
D | cwise_op_ceil.cc | 19 REGISTER3(UnaryOp, CPU, "Ceil", functor::ceil, float, Eigen::half, double); 22 REGISTER3(UnaryOp, GPU, "Ceil", functor::ceil, float, Eigen::half, double); 26 REGISTER2(UnaryOp, SYCL, "Ceil", functor::ceil, float, double);
|
/external/llvm/test/CodeGen/SystemZ/ |
D | fp-round-01.ll | 83 ; Test ceil for f32. 84 declare float @llvm.ceil.f32(float %f) 89 %res = call float @llvm.ceil.f32(float %f) 93 ; Test ceil for f64. 94 declare double @llvm.ceil.f64(double %f) 97 ; CHECK: brasl %r14, ceil@PLT 99 %res = call double @llvm.ceil.f64(double %f) 103 ; Test ceil for f128: omitted for now because we cannot handle
|
D | fp-round-02.ll | 105 ; Test ceil for f32. 106 declare float @llvm.ceil.f32(float %f) 111 %res = call float @llvm.ceil.f32(float %f) 115 ; Test ceil for f64. 116 declare double @llvm.ceil.f64(double %f) 121 %res = call double @llvm.ceil.f64(double %f) 125 ; Test ceil for f128. 126 declare fp128 @llvm.ceil.f128(fp128 %f) 132 %res = call fp128 @llvm.ceil.f128(fp128 %src)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/SystemZ/ |
D | fp-round-01.ll | 83 ; Test ceil for f32. 84 declare float @llvm.ceil.f32(float %f) 89 %res = call float @llvm.ceil.f32(float %f) 93 ; Test ceil for f64. 94 declare double @llvm.ceil.f64(double %f) 97 ; CHECK: brasl %r14, ceil@PLT 99 %res = call double @llvm.ceil.f64(double %f) 103 ; Test ceil for f128: omitted for now because we cannot handle
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/PowerPC/ |
D | vec_rounding.ll | 31 declare <2 x double> @llvm.ceil.v2f64(<2 x double> %p) 34 %t = call <2 x double> @llvm.ceil.v2f64(<2 x double> %p) 41 declare <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 44 %t = call <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 117 declare <4 x float> @llvm.ceil.v4f32(<4 x float> %p) 120 %t = call <4 x float> @llvm.ceil.v4f32(<4 x float> %p) 126 declare <8 x float> @llvm.ceil.v8f32(<8 x float> %p) 129 %t = call <8 x float> @llvm.ceil.v8f32(<8 x float> %p)
|
/external/llvm/test/CodeGen/PowerPC/ |
D | vec_rounding.ll | 31 declare <2 x double> @llvm.ceil.v2f64(<2 x double> %p) 34 %t = call <2 x double> @llvm.ceil.v2f64(<2 x double> %p) 41 declare <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 44 %t = call <4 x double> @llvm.ceil.v4f64(<4 x double> %p) 117 declare <4 x float> @llvm.ceil.v4f32(<4 x float> %p) 120 %t = call <4 x float> @llvm.ceil.v4f32(<4 x float> %p) 126 declare <8 x float> @llvm.ceil.v8f32(<8 x float> %p) 129 %t = call <8 x float> @llvm.ceil.v8f32(<8 x float> %p)
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/tests/ |
D | cpu_fusion_test.cc | 100 auto ceil = builder.AddInstruction( in TEST_F() local 103 HloInstruction::CreateUnary(vshape, HloOpcode::kExp, ceil)); in TEST_F() 150 auto ceil = builder.AddInstruction( in TEST_F() local 155 HloInstruction::CreateConcatenate(cshape, {ceil, ceil}, /*dimension=*/0)); in TEST_F() 247 auto ceil = builder.AddInstruction( in TEST_F() local 251 HloInstruction::CreateBinary(vshape, HloOpcode::kMultiply, negate, ceil)); in TEST_F() 253 HloInstruction::CreateBinary(vshape, HloOpcode::kMultiply, ceil, negate)); in TEST_F()
|