aboutsummaryrefslogtreecommitdiff
path: root/test/CodeGen/X86/fold-vector-sext-zext.ll
diff options
context:
space:
mode:
authorDimitry Andric <dim@FreeBSD.org>2017-01-02 19:17:04 +0000
committerDimitry Andric <dim@FreeBSD.org>2017-01-02 19:17:04 +0000
commitb915e9e0fc85ba6f398b3fab0db6a81a8913af94 (patch)
tree98b8f811c7aff2547cab8642daf372d6c59502fb /test/CodeGen/X86/fold-vector-sext-zext.ll
parent6421cca32f69ac849537a3cff78c352195e99f1b (diff)
downloadsrc-b915e9e0fc85ba6f398b3fab0db6a81a8913af94.tar.gz
src-b915e9e0fc85ba6f398b3fab0db6a81a8913af94.zip
Vendor import of llvm trunk r290819:vendor/llvm/llvm-trunk-r290819
Notes
Notes: svn path=/vendor/llvm/dist/; revision=311116 svn path=/vendor/llvm/llvm-trunk-r290819/; revision=311117; tag=vendor/llvm/llvm-trunk-r290819
Diffstat (limited to 'test/CodeGen/X86/fold-vector-sext-zext.ll')
-rw-r--r--test/CodeGen/X86/fold-vector-sext-zext.ll312
1 files changed, 211 insertions, 101 deletions
diff --git a/test/CodeGen/X86/fold-vector-sext-zext.ll b/test/CodeGen/X86/fold-vector-sext-zext.ll
index 6299280eb98d..3f502efa753e 100644
--- a/test/CodeGen/X86/fold-vector-sext-zext.ll
+++ b/test/CodeGen/X86/fold-vector-sext-zext.ll
@@ -1,5 +1,6 @@
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
-; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mcpu=corei7-avx -mattr=+avx | FileCheck %s
+; RUN: llc < %s -mtriple=i686-unknown-linux-gnu -mattr=+avx | FileCheck %s --check-prefix=X32
+; RUN: llc < %s -mtriple=x86_64-unknown-linux-gnu -mattr=+avx | FileCheck %s --check-prefix=X64
; Verify that the backend correctly folds a sign/zero extend of a vector where
; elements are all constant values or UNDEFs.
@@ -7,11 +8,16 @@
; simple loads from constant pool of the result. That is because the resulting
; vector should be known at static time.
-define <4 x i16> @test1() {
-; CHECK-LABEL: test1:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
-; CHECK-NEXT: retq
+define <4 x i16> @test_sext_4i8_4i16() {
+; X32-LABEL: test_sext_4i8_4i16:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i16:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -20,11 +26,16 @@ define <4 x i16> @test1() {
ret <4 x i16> %5
}
-define <4 x i16> @test2() {
-; CHECK-LABEL: test2:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
-; CHECK-NEXT: retq
+define <4 x i16> @test_sext_4i8_4i16_undef() {
+; X32-LABEL: test_sext_4i8_4i16_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i16_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 undef, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 undef, i32 2
@@ -33,11 +44,16 @@ define <4 x i16> @test2() {
ret <4 x i16> %5
}
-define <4 x i32> @test3() {
-; CHECK-LABEL: test3:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
-; CHECK-NEXT: retq
+define <4 x i32> @test_sext_4i8_4i32() {
+; X32-LABEL: test_sext_4i8_4i32:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i32:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -46,11 +62,16 @@ define <4 x i32> @test3() {
ret <4 x i32> %5
}
-define <4 x i32> @test4() {
-; CHECK-LABEL: test4:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
-; CHECK-NEXT: retq
+define <4 x i32> @test_sext_4i8_4i32_undef() {
+; X32-LABEL: test_sext_4i8_4i32_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i32_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 undef, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 undef, i32 2
@@ -59,11 +80,17 @@ define <4 x i32> @test4() {
ret <4 x i32> %5
}
-define <4 x i64> @test5() {
-; CHECK-LABEL: test5:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = [0,18446744073709551615,2,18446744073709551613]
-; CHECK-NEXT: retq
+define <4 x i64> @test_sext_4i8_4i64() {
+; X32-LABEL: test_sext_4i8_4i64:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,0,4294967295,4294967295]
+; X32-NEXT: vinsertf128 $1, {{\.LCPI.*}}, %ymm0, %ymm0
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i64:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = [0,18446744073709551615,2,18446744073709551613]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -72,11 +99,17 @@ define <4 x i64> @test5() {
ret <4 x i64> %5
}
-define <4 x i64> @test6() {
-; CHECK-LABEL: test6:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <u,18446744073709551615,u,18446744073709551613>
-; CHECK-NEXT: retq
+define <4 x i64> @test_sext_4i8_4i64_undef() {
+; X32-LABEL: test_sext_4i8_4i64_undef:
+; X32: # BB#0:
+; X32-NEXT: vpcmpeqd %xmm0, %xmm0, %xmm0
+; X32-NEXT: vinsertf128 $1, {{\.LCPI.*}}, %ymm0, %ymm0
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_4i8_4i64_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = <u,18446744073709551615,u,18446744073709551613>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 undef, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 undef, i32 2
@@ -85,11 +118,16 @@ define <4 x i64> @test6() {
ret <4 x i64> %5
}
-define <8 x i16> @test7() {
-; CHECK-LABEL: test7:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <0,65535,2,65533,u,u,u,u>
-; CHECK-NEXT: retq
+define <8 x i16> @test_sext_8i8_8i16() {
+; X32-LABEL: test_sext_8i8_8i16:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <0,65535,2,65533,u,u,u,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_8i8_8i16:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <0,65535,2,65533,u,u,u,u>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2
@@ -102,11 +140,16 @@ define <8 x i16> @test7() {
ret <8 x i16> %9
}
-define <8 x i32> @test8() {
-; CHECK-LABEL: test8:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <0,4294967295,2,4294967293,u,u,u,u>
-; CHECK-NEXT: retq
+define <8 x i32> @test_sext_8i8_8i32() {
+; X32-LABEL: test_sext_8i8_8i32:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} ymm0 = <0,4294967295,2,4294967293,u,u,u,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_8i8_8i32:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = <0,4294967295,2,4294967293,u,u,u,u>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2
@@ -119,11 +162,16 @@ define <8 x i32> @test8() {
ret <8 x i32> %9
}
-define <8 x i16> @test9() {
-; CHECK-LABEL: test9:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,65535,u,65533,u,u,u,u>
-; CHECK-NEXT: retq
+define <8 x i16> @test_sext_8i8_8i16_undef() {
+; X32-LABEL: test_sext_8i8_8i16_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,65535,u,65533,u,u,u,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_8i8_8i16_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,65535,u,65533,u,u,u,u>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 undef, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 undef, i32 2
@@ -136,11 +184,16 @@ define <8 x i16> @test9() {
ret <8 x i16> %9
}
-define <8 x i32> @test10() {
-; CHECK-LABEL: test10:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,u,u,u,u,u>
-; CHECK-NEXT: retq
+define <8 x i32> @test_sext_8i8_8i32_undef() {
+; X32-LABEL: test_sext_8i8_8i32_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,u,u,u,u,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_sext_8i8_8i32_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,u,u,u,u,u>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 undef, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2
@@ -153,11 +206,16 @@ define <8 x i32> @test10() {
ret <8 x i32> %9
}
-define <4 x i16> @test11() {
-; CHECK-LABEL: test11:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
-; CHECK-NEXT: retq
+define <4 x i16> @test_zext_4i8_4i16() {
+; X32-LABEL: test_zext_4i8_4i16:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i16:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -166,11 +224,16 @@ define <4 x i16> @test11() {
ret <4 x i16> %5
}
-define <4 x i32> @test12() {
-; CHECK-LABEL: test12:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
-; CHECK-NEXT: retq
+define <4 x i32> @test_zext_4i8_4i32() {
+; X32-LABEL: test_zext_4i8_4i32:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i32:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -179,11 +242,18 @@ define <4 x i32> @test12() {
ret <4 x i32> %5
}
-define <4 x i64> @test13() {
-; CHECK-LABEL: test13:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,2,253]
-; CHECK-NEXT: retq
+define <4 x i64> @test_zext_4i8_4i64() {
+; X32-LABEL: test_zext_4i8_4i64:
+; X32: # BB#0:
+; X32-NEXT: vpmovzxbq {{.*#+}} xmm0 = mem[0],zero,zero,zero,zero,zero,zero,zero,mem[1],zero,zero,zero,zero,zero,zero,zero
+; X32-NEXT: vpmovzxbq {{.*#+}} xmm1 = mem[0],zero,zero,zero,zero,zero,zero,zero,mem[1],zero,zero,zero,zero,zero,zero,zero
+; X32-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i64:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,2,253]
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -192,11 +262,16 @@ define <4 x i64> @test13() {
ret <4 x i64> %5
}
-define <4 x i16> @test14() {
-; CHECK-LABEL: test14:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253>
-; CHECK-NEXT: retq
+define <4 x i16> @test_zext_4i8_4i16_undef() {
+; X32-LABEL: test_zext_4i8_4i16_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i16_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 undef, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 undef, i32 2
@@ -205,11 +280,16 @@ define <4 x i16> @test14() {
ret <4 x i16> %5
}
-define <4 x i32> @test15() {
-; CHECK-LABEL: test15:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <0,u,2,u>
-; CHECK-NEXT: retq
+define <4 x i32> @test_zext_4i8_4i32_undef() {
+; X32-LABEL: test_zext_4i8_4i32_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <0,u,2,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i32_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <0,u,2,u>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 0, i32 0
%2 = insertelement <4 x i8> %1, i8 undef, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -218,11 +298,21 @@ define <4 x i32> @test15() {
ret <4 x i32> %5
}
-define <4 x i64> @test16() {
-; CHECK-LABEL: test16:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <u,255,2,u>
-; CHECK-NEXT: retq
+define <4 x i64> @test_zext_4i8_4i64_undef() {
+; X32-LABEL: test_zext_4i8_4i64_undef:
+; X32: # BB#0:
+; X32-NEXT: vpcmpeqd %xmm0, %xmm0, %xmm0
+; X32-NEXT: vpmovzxbq {{.*#+}} xmm0 = xmm0[0],zero,zero,zero,zero,zero,zero,zero,xmm0[1],zero,zero,zero,zero,zero,zero,zero
+; X32-NEXT: movl $2, %eax
+; X32-NEXT: vmovd %eax, %xmm1
+; X32-NEXT: vpmovzxbq {{.*#+}} xmm1 = xmm1[0],zero,zero,zero,zero,zero,zero,zero,xmm1[1],zero,zero,zero,zero,zero,zero,zero
+; X32-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_4i8_4i64_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = <u,255,2,u>
+; X64-NEXT: retq
%1 = insertelement <4 x i8> undef, i8 undef, i32 0
%2 = insertelement <4 x i8> %1, i8 -1, i32 1
%3 = insertelement <4 x i8> %2, i8 2, i32 2
@@ -231,11 +321,16 @@ define <4 x i64> @test16() {
ret <4 x i64> %5
}
-define <8 x i16> @test17() {
-; CHECK-LABEL: test17:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253,4,251,6,249]
-; CHECK-NEXT: retq
+define <8 x i16> @test_zext_8i8_8i16() {
+; X32-LABEL: test_zext_8i8_8i16:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253,4,251,6,249]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_8i8_8i16:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,255,2,253,4,251,6,249]
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2
@@ -248,11 +343,16 @@ define <8 x i16> @test17() {
ret <8 x i16> %9
}
-define <8 x i32> @test18() {
-; CHECK-LABEL: test18:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,2,253,4,251,6,249]
-; CHECK-NEXT: retq
+define <8 x i32> @test_zext_8i8_8i32() {
+; X32-LABEL: test_zext_8i8_8i32:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,2,253,4,251,6,249]
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_8i8_8i32:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = [0,255,2,253,4,251,6,249]
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2
@@ -265,11 +365,16 @@ define <8 x i32> @test18() {
ret <8 x i32> %9
}
-define <8 x i16> @test19() {
-; CHECK-LABEL: test19:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253,u,251,u,249>
-; CHECK-NEXT: retq
+define <8 x i16> @test_zext_8i8_8i16_undef() {
+; X32-LABEL: test_zext_8i8_8i16_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253,u,251,u,249>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_8i8_8i16_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,255,u,253,u,251,u,249>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 undef, i32 0
%2 = insertelement <8 x i8> %1, i8 -1, i32 1
%3 = insertelement <8 x i8> %2, i8 undef, i32 2
@@ -282,11 +387,16 @@ define <8 x i16> @test19() {
ret <8 x i16> %9
}
-define <8 x i32> @test20() {
-; CHECK-LABEL: test20:
-; CHECK: # BB#0:
-; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,253,4,u,6,u>
-; CHECK-NEXT: retq
+define <8 x i32> @test_zext_8i8_8i32_undef() {
+; X32-LABEL: test_zext_8i8_8i32_undef:
+; X32: # BB#0:
+; X32-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,253,4,u,6,u>
+; X32-NEXT: retl
+;
+; X64-LABEL: test_zext_8i8_8i32_undef:
+; X64: # BB#0:
+; X64-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,253,4,u,6,u>
+; X64-NEXT: retq
%1 = insertelement <8 x i8> undef, i8 0, i32 0
%2 = insertelement <8 x i8> %1, i8 undef, i32 1
%3 = insertelement <8 x i8> %2, i8 2, i32 2