diff options
author | Justin Holewinski <jholewinski@nvidia.com> | 2012-05-04 20:18:50 +0000 |
---|---|---|
committer | Justin Holewinski <jholewinski@nvidia.com> | 2012-05-04 20:18:50 +0000 |
commit | 49683f3c961379fbc088871a5d6304950f1f1cbc (patch) | |
tree | 830fa1ee9c992ef4645863d128be912ce2bfc987 /test/CodeGen/NVPTX/ld-addrspace.ll | |
parent | 2c7e5c714c8675f757c4936a3a2132c2466a626c (diff) | |
download | llvm-49683f3c961379fbc088871a5d6304950f1f1cbc.tar.gz llvm-49683f3c961379fbc088871a5d6304950f1f1cbc.tar.bz2 llvm-49683f3c961379fbc088871a5d6304950f1f1cbc.tar.xz |
This patch adds a new NVPTX back-end to LLVM which supports code generation for NVIDIA PTX 3.0. This back-end will (eventually) replace the current PTX back-end, while maintaining compatibility with it.
The new target machines are:
nvptx (old ptx32) => 32-bit PTX
nvptx64 (old ptx64) => 64-bit PTX
The sources are based on the internal NVIDIA NVPTX back-end, and
contain more functionality than the current PTX back-end currently
provides.
NV_CONTRIB
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@156196 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test/CodeGen/NVPTX/ld-addrspace.ll')
-rw-r--r-- | test/CodeGen/NVPTX/ld-addrspace.ll | 173 |
1 files changed, 173 insertions, 0 deletions
diff --git a/test/CodeGen/NVPTX/ld-addrspace.ll b/test/CodeGen/NVPTX/ld-addrspace.ll new file mode 100644 index 0000000000..d1f5093df2 --- /dev/null +++ b/test/CodeGen/NVPTX/ld-addrspace.ll @@ -0,0 +1,173 @@ +; RUN: llc < %s -march=nvptx -mcpu=sm_10 | FileCheck %s --check-prefix=PTX32 +; RUN: llc < %s -march=nvptx -mcpu=sm_20 | FileCheck %s --check-prefix=PTX32 +; RUN: llc < %s -march=nvptx64 -mcpu=sm_10 | FileCheck %s --check-prefix=PTX64 +; RUN: llc < %s -march=nvptx64 -mcpu=sm_20 | FileCheck %s --check-prefix=PTX64 + + +;; i8 +define i8 @ld_global_i8(i8 addrspace(1)* %ptr) { +; PTX32: ld.global.u8 %rc{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.u8 %rc{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i8 addrspace(1)* %ptr + ret i8 %a +} + +define i8 @ld_shared_i8(i8 addrspace(3)* %ptr) { +; PTX32: ld.shared.u8 %rc{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.u8 %rc{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i8 addrspace(3)* %ptr + ret i8 %a +} + +define i8 @ld_local_i8(i8 addrspace(5)* %ptr) { +; PTX32: ld.local.u8 %rc{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.u8 %rc{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i8 addrspace(5)* %ptr + ret i8 %a +} + +;; i16 +define i16 @ld_global_i16(i16 addrspace(1)* %ptr) { +; PTX32: ld.global.u16 %rs{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.u16 %rs{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i16 addrspace(1)* %ptr + ret i16 %a +} + +define i16 @ld_shared_i16(i16 addrspace(3)* %ptr) { +; PTX32: ld.shared.u16 %rs{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.u16 %rs{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i16 addrspace(3)* %ptr + ret i16 %a +} + +define i16 @ld_local_i16(i16 addrspace(5)* %ptr) { +; PTX32: ld.local.u16 %rs{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.u16 %rs{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i16 addrspace(5)* %ptr + ret i16 %a +} + +;; i32 +define i32 @ld_global_i32(i32 addrspace(1)* %ptr) { +; PTX32: ld.global.u32 %r{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.u32 %r{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i32 addrspace(1)* %ptr + ret i32 %a +} + +define i32 @ld_shared_i32(i32 addrspace(3)* %ptr) { +; PTX32: ld.shared.u32 %r{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.u32 %r{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i32 addrspace(3)* %ptr + ret i32 %a +} + +define i32 @ld_local_i32(i32 addrspace(5)* %ptr) { +; PTX32: ld.local.u32 %r{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.u32 %r{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i32 addrspace(5)* %ptr + ret i32 %a +} + +;; i64 +define i64 @ld_global_i64(i64 addrspace(1)* %ptr) { +; PTX32: ld.global.u64 %rl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.u64 %rl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i64 addrspace(1)* %ptr + ret i64 %a +} + +define i64 @ld_shared_i64(i64 addrspace(3)* %ptr) { +; PTX32: ld.shared.u64 %rl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.u64 %rl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i64 addrspace(3)* %ptr + ret i64 %a +} + +define i64 @ld_local_i64(i64 addrspace(5)* %ptr) { +; PTX32: ld.local.u64 %rl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.u64 %rl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load i64 addrspace(5)* %ptr + ret i64 %a +} + +;; f32 +define float @ld_global_f32(float addrspace(1)* %ptr) { +; PTX32: ld.global.f32 %f{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.f32 %f{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load float addrspace(1)* %ptr + ret float %a +} + +define float @ld_shared_f32(float addrspace(3)* %ptr) { +; PTX32: ld.shared.f32 %f{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.f32 %f{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load float addrspace(3)* %ptr + ret float %a +} + +define float @ld_local_f32(float addrspace(5)* %ptr) { +; PTX32: ld.local.f32 %f{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.f32 %f{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load float addrspace(5)* %ptr + ret float %a +} + +;; f64 +define double @ld_global_f64(double addrspace(1)* %ptr) { +; PTX32: ld.global.f64 %fl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.global.f64 %fl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load double addrspace(1)* %ptr + ret double %a +} + +define double @ld_shared_f64(double addrspace(3)* %ptr) { +; PTX32: ld.shared.f64 %fl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.shared.f64 %fl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load double addrspace(3)* %ptr + ret double %a +} + +define double @ld_local_f64(double addrspace(5)* %ptr) { +; PTX32: ld.local.f64 %fl{{[0-9]+}}, [%r{{[0-9]+}}] +; PTX32: ret +; PTX64: ld.local.f64 %fl{{[0-9]+}}, [%rl{{[0-9]+}}] +; PTX64: ret + %a = load double addrspace(5)* %ptr + ret double %a +} |