/external/deqp/modules/gles2/performance/ |
D | es2pShaderOptimizationTests.cpp | 186 virtual ProgramData generateProgramData (bool optimized) const = 0; 202 …ProgramData& programData (bool optimized) { return optimized ? m_optimizedData : m_unoptimi… in programData() argument 203 …SharedPtr<const ShaderProgram>& program (bool optimized) { return optimized ? m_optimizedProgra… in program() argument 204 …ShaderPerformanceMeasurer::Result& result (bool optimized) { return optimized ? m_optimizedResul… in result() argument 226 const bool optimized = ndx == 1; in init() local 228 programData(optimized) = generateProgramData(optimized); in init() 230 for (int i = 0; i < (int)programData(optimized).attributes.size(); i++) in init() 231 …DE_ASSERT(programData(optimized).attributes[i].name != "a_position"); // \note Position attribute … in init() 233 …program(optimized) = SharedPtr<const ShaderProgram>(new ShaderProgram(renderCtx, programData(optim… in init() 236 const tcu::ScopedLogSection section(log, optimized ? "OptimizedProgram" : "UnoptimizedProgram", in init() [all …]
|
/external/deqp/modules/gles3/performance/ |
D | es3pShaderOptimizationTests.cpp | 189 virtual ProgramData generateProgramData (bool optimized) const = 0; 205 …ProgramData& programData (bool optimized) { return optimized ? m_optimizedData : m_unoptimi… in programData() argument 206 …SharedPtr<const ShaderProgram>& program (bool optimized) { return optimized ? m_optimizedProgra… in program() argument 207 …ShaderPerformanceMeasurer::Result& result (bool optimized) { return optimized ? m_optimizedResul… in result() argument 229 const bool optimized = ndx == 1; in init() local 231 programData(optimized) = generateProgramData(optimized); in init() 233 for (int i = 0; i < (int)programData(optimized).attributes.size(); i++) in init() 234 …DE_ASSERT(programData(optimized).attributes[i].name != "a_position"); // \note Position attribute … in init() 236 …program(optimized) = SharedPtr<const ShaderProgram>(new ShaderProgram(renderCtx, programData(optim… in init() 239 const tcu::ScopedLogSection section(log, optimized ? "OptimizedProgram" : "UnoptimizedProgram", in init() [all …]
|
/external/v8/test/mjsunit/regress/ |
D | regress-1079.js | 33 function optimized() { function 40 return optimized.arguments; 45 assertEquals(3, optimized(1, 2, 3).length); 47 %OptimizeFunctionOnNextCall(optimized); 48 assertEquals(3, optimized(1, 2, 3).length);
|
D | regress-2499.js | 39 var optimized = foo(word, 1); variable 40 assertEquals(expected, optimized)
|
/external/guice/extensions/assistedinject/src/com/google/inject/assistedinject/ |
D | FactoryProvider2.java | 134 final boolean optimized; field in FactoryProvider2.AssistData 143 boolean optimized, List<ThreadLocalProvider> providers) { in AssistData() argument 150 this.optimized = optimized; in AssistData() 163 .add("optimized", optimized) in toString() 296 boolean optimized = false; in FactoryProvider2() 308 optimized = true; in FactoryProvider2() 317 optimized, in FactoryProvider2() 671 if(!data.optimized) { in initialize() 702 if(!data.optimized) { in getBindingFromNewInjector() 729 if(data.optimized) { in getBindingFromNewInjector()
|
/external/llvm/test/Transforms/InstCombine/ |
D | udiv-simplify-bug-1.ll | 6 ; The udiv instructions shouldn't be optimized away, and the 7 ; sext instructions should be optimized to zext.
|
D | 2008-07-16-fsub.ll | 5 ; fsub of self can't be optimized away.
|
/external/llvm/docs/HistoricalNotes/ |
D | 2003-06-25-Reoptimizer1.txt | 87 When we finally generate an optimized trace we first copy the code 89 original code, the instrumented code, and the optimized trace. The 90 optimized trace does not have instrumentation. The original code and 92 cache, where the optimized traces are kept. 98 generating the final optimized trace; we copy the same basic blocks 121 the beginning of the original code to the beginning of the optimized 132 We generate instrumentation traces and optimized traces into separate
|
/external/llvm/docs/ |
D | Packaging.rst | 23 LLVM runs much more quickly when it's optimized and assertions are removed. 26 in user code. We recommend allowing users to install both optimized and debug 40 ``--enable-optimized``
|
D | BuildingLLVMWithAutotools.rst | 36 * ``--enable-optimized`` --- Compile with optimizations enabled (default 44 The ``--enable-optimized`` configure option is used to specify a Release 82 ``--enable-optimized`` 84 Enables optimized compilation (debugging symbols are removed and GCC 141 types ``gmake`` (unless the ``--enable-optimized`` option was used during 144 ``--disable-optimized`` option must be passed to ``configure``. 148 These builds are enabled with the ``--enable-optimized`` option to
|
/external/gemmlowp/ |
D | README.txt | 34 while we have optional optimized code paths for specific architectures. 47 See "Architecture-specific optimized code paths" below. 49 Architecture-specific optimized code paths 52 We have some optimized code paths for specific instruction sets. 56 At the moment, we have a full set of optimized code paths (kernels, 60 We also have a partial set of optimized code paths (only kernels 256 2. Most GEMMs are optimized primarily for large dense matrix sizes (>= 1000).
|
/external/llvm/test/Transforms/GlobalOpt/ |
D | 2009-11-16-MallocSingleStoreToGlobalVar.ll | 1 ; Test ensures that non-optimizable array mallocs are not optimized; specifically 4 ; element global. The global object @TOP in this test should not be optimized.
|
/external/llvm/utils/valgrind/ |
D | x86_64-pc-linux-gnu.supp | 77 suppress optimized strcasecmp, to be fixed in valgrind 3.6.1 83 suppress optimized strcasecmp, to be fixed in valgrind 3.6.1
|
/external/proguard/examples/annotations/lib/ |
D | annotations.pro | 63 # shrunk, optimized, or obfuscated as entry points. 80 # optimized or obfuscated as entry points. 96 # optimized, or obfuscated as entry points.
|
/external/llvm/test/Transforms/LoadCombine/ |
D | load-combine-assume.ll | 7 ; 'load' before the 'call' gets optimized: 26 ; 'call' before the 'load' doesn't get optimized:
|
/external/llvm/ |
D | llvm.spec.in | 34 --enable-optimized \ 62 - Make the build be optimized+assertions
|
D | Makefile | 224 $(Echo) '*****' optimized build. Use 'make ENABLE_OPTIMIZED=1' to 225 $(Echo) '*****' make an optimized build. Alternatively you can 226 $(Echo) '*****' configure with --enable-optimized.
|
/external/gemmlowp/meta/ |
D | README | 6 - providing the optimized codepaths for as many possible user scenarios without 72 optimized, unrolled and curried pack/unpack, and multiply procedures and 73 a single gemm function that switches between the optimized versions based
|
/external/e2fsprogs/tests/e_icount_opt/ |
D | name | 1 inode counting abstraction optimized for counting
|
/external/e2fsprogs/tests/e_icount_normal/ |
D | name | 1 inode counting structure optimized for low counts
|
/external/opencv3/doc/py_tutorials/py_core/py_optimization/ |
D | py_optimization.markdown | 57 Many of the OpenCV functions are optimized using SSE2, AVX etc. It contains unoptimized code also. 59 support them). It is enabled by default while compiling. So OpenCV runs the optimized code if it is 79 See, optimized median filtering is \~2x faster than unoptimized version. If you check its source, 80 you can see median filtering is SIMD optimized. So you can use this to enable optimization at the 152 optimized for vector operations.
|
/external/zlib/src/contrib/masmx86/ |
D | match686.asm | 1 ; match686.asm -- Asm portion of the optimized longest_match for 32 bits x86 17 ; this longest_match was written by Brian raiter (1998), optimized for Pentium Pro 31 ; this file is not optimized for old pentium, but it compatible with all x86 32 bits 162 ;;; match686.asm -- Pentium-Pro-optimized version of longest_match()
|
/external/protobuf/src/google/protobuf/ |
D | unittest_embed_optimize_for.proto | 41 // We optimize for speed here, but we are importing a proto that is optimized 47 // one optimized for speed works.
|
/external/protobuf/python/google/protobuf/pyext/ |
D | python.proto | 39 // Protos optimized for SPEED use a strict superset of the generated code 40 // of equivalent ones optimized for CODE_SIZE, so we should optimize all our
|
/external/clang/test/CodeGenOpenCL/ |
D | vectorLoadStore.cl | 5 // Check for optimized vec3 load/store which treats vec3 as vec4.
|