Home
last modified time | relevance | path

Searched refs:cubin (Results 1 – 11 of 11) sorted by relevance

/external/llvm-project/mlir/test/Conversion/GPUToCUDA/
Dlower-nvvm-kernel-to-cubin.mlir1 // RUN: mlir-opt %s --test-kernel-to-cubin -split-input-file | FileCheck %s
3 // CHECK: attributes {nvvm.cubin = "CUBIN"}
/external/tensorflow/tensorflow/stream_executor/cuda/
Dcuda_asm_compiler.cc49 std::vector<uint8> cubin(static_cast<uint8*>(cubin_out), in LinkGpuAsm() local
52 return std::move(cubin); in LinkGpuAsm()
Dcuda_gpu_executor.cc220 port::Status GpuExecutor::LoadModuleFromCuBin(const char* cubin, in LoadModuleFromCuBin() argument
223 std::tie(*module, module_refcount) = gpu_binary_to_module_[cubin]; in LoadModuleFromCuBin()
226 TF_RETURN_IF_ERROR(GpuDriver::LoadCubin(context_, cubin, module)); in LoadModuleFromCuBin()
228 VLOG(3) << "Loaded CUBIN " << static_cast<const void*>(cubin) in LoadModuleFromCuBin()
232 VLOG(3) << "CUBIN " << static_cast<const void*>(cubin) in LoadModuleFromCuBin()
235 gpu_binary_to_module_[cubin] = {*module, module_refcount}; in LoadModuleFromCuBin()
274 const char* cubin = spec.cuda_cubin_in_memory().bytes(); in GetKernel() local
275 TF_RETURN_IF_ERROR(LoadModuleFromCuBin(cubin, &module)); in GetKernel()
276 kernel_to_gpu_binary_[kernel] = cubin; in GetKernel()
/external/tensorflow/third_party/nccl/
Dbuild_defs.bzl.tpl91 cubin = ctx.actions.declare_file("%s_%s.cubin" % (name, arch))
94 outputs = [register_h, cubin],
100 "--output-file=%s" % cubin.path,
104 cubins.append(cubin)
105 images.append("--image=profile=%s,file=%s" % (arch, cubin.path))
277 into cubin files, one per GPU architecture. It also produces a header file
290 register.h xy.cubin
/external/llvm-project/mlir/test/Conversion/GPUCommon/
Dlower-launch-func-to-gpu-runtime-calls.mlir1 // RUN: mlir-opt -allow-unregistered-dialect %s --gpu-to-llvm="gpu-binary-annotation=nvvm.cubin" | …
11 nvvm.cubin = "CUBIN", rocdl.hsaco = "HSACO"
/external/tensorflow/tensorflow/stream_executor/gpu/
Dasm_compiler.cc278 std::string cubin; in CompileGpuAsm() local
280 cubin_path, &cubin)); in CompileGpuAsm()
281 std::vector<uint8> cubin_vector(cubin.begin(), cubin.end()); in CompileGpuAsm()
/external/skqp/src/compute/hs/cuda/sm_35/u64/
Dgen.bat39 nvcc -I ../../../.. -use_fast_math -res-usage -cubin -arch sm_35 %HS_SRC%
/external/skqp/src/compute/hs/cuda/sm_35/u32/
Dgen.bat39 nvcc -I ../../../.. -use_fast_math -res-usage -cubin -arch sm_35 %HS_SRC%
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dnvptx_compiler.cc333 std::vector<uint8> cubin = CompileGpuAsmOrGetCachedResult( in CompileTargetBinary() local
338 std::move(cubin)); in CompileTargetBinary()
/external/tensorflow/tensorflow/core/kernels/mlir_generated/
Dbuild_defs.bzl1 """Generates cubin headers for TF dialect ops."""
29 "GPU binaries in either cubin format or hsaco format",
/external/tensorflow/tensorflow/stream_executor/rocm/
Drocm_gpu_executor.cc386 port::Status GpuExecutor::LoadModuleFromCuBin(const char* cubin, in LoadModuleFromCuBin() argument