1; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
2; RUN: opt < %s -loop-vectorize -S | FileCheck %s
3target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128"
4target triple = "x86_64-apple-macosx10.11.0"
5
6; This test checks vector GEP before scatter.
7; The code bellow crashed due to destroyed SSA while incorrect vectorization of
8; the GEP.
9
10@d = global [10 x [10 x i32]] zeroinitializer, align 16
11@c = external global i32, align 4
12@a = external global i32, align 4
13@b = external global i64, align 8
14
15; Function Attrs: norecurse nounwind ssp uwtable
16define void @_Z3fn1v() #0 {
17; CHECK-LABEL: @_Z3fn1v(
18; CHECK:       vector.body:
19; CHECK-NEXT:    [[INDEX:%.*]] = phi i64 [ 0, %vector.ph ], [ [[INDEX_NEXT:%.*]], %vector.body ]
20; CHECK-NEXT:    [[VEC_IND:%.*]] = phi <16 x i64> [ <i64 8, i64 10, i64 12, i64 14, i64 16, i64 18, i64 20, i64 22, i64 24, i64 26, i64 28, i64 30, i64 32, i64 34, i64 36, i64 38>, %vector.ph ], [ [[VEC_IND_NEXT:%.*]], %vector.body ]
21; CHECK-NEXT:    [[VEC_IND3:%.*]] = phi <16 x i64> [ <i64 0, i64 2, i64 4, i64 6, i64 8, i64 10, i64 12, i64 14, i64 16, i64 18, i64 20, i64 22, i64 24, i64 26, i64 28, i64 30>, %vector.ph ], [ [[VEC_IND_NEXT4:%.*]], %vector.body ]
22; CHECK-NEXT:    [[TMP10:%.*]] = sub nsw <16 x i64> <i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8, i64 8>, [[VEC_IND]]
23; CHECK-NEXT:    [[TMP11:%.*]] = getelementptr inbounds [10 x [10 x i32]], [10 x [10 x i32]]* @d, i64 0, <16 x i64> [[VEC_IND]]
24; CHECK-NEXT:    [[TMP12:%.*]] = add nsw <16 x i64> [[TMP10]], [[VEC_IND3]]
25; CHECK-NEXT:    [[TMP13:%.*]] = getelementptr inbounds [10 x i32], <16 x [10 x i32]*> [[TMP11]], <16 x i64> [[TMP12]], i64 0
26; CHECK-NEXT:    call void @llvm.masked.scatter.v16i32.v16p0i32(<16 x i32> <i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8>, <16 x i32*> [[TMP13]], i32 16, <16 x i1> <i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true>)
27; CHECK-NEXT:    [[TMP14:%.*]] = or <16 x i64> [[VEC_IND3]], <i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1, i64 1>
28; CHECK-NEXT:    [[TMP15:%.*]] = add nsw <16 x i64> [[TMP10]], [[TMP14]]
29; CHECK-NEXT:    [[TMP16:%.*]] = getelementptr inbounds [10 x i32], <16 x [10 x i32]*> [[TMP11]], <16 x i64> [[TMP15]], i64 0
30; CHECK-NEXT:    call void @llvm.masked.scatter.v16i32.v16p0i32(<16 x i32> <i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8, i32 8>, <16 x i32*> [[TMP16]], i32 8, <16 x i1> <i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true, i1 true>)
31; CHECK-NEXT:    [[INDEX_NEXT]] = add nuw i64 [[INDEX]], 16
32; CHECK-NEXT:    [[VEC_IND_NEXT]] = add <16 x i64> [[VEC_IND]], <i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32>
33; CHECK-NEXT:    [[VEC_IND_NEXT4]] = add <16 x i64> [[VEC_IND3]], <i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32, i64 32>
34; CHECK:         br i1 {{.*}}, label %middle.block, label %vector.body
35;
36entry:
37  %0 = load i32, i32* @c, align 4
38  %cmp34 = icmp sgt i32 %0, 8
39  br i1 %cmp34, label %for.body.lr.ph, label %for.cond.cleanup
40
41for.body.lr.ph:                                   ; preds = %entry
42  %1 = load i32, i32* @a, align 4
43  %tobool = icmp eq i32 %1, 0
44  %2 = load i64, i64* @b, align 8
45  %mul = mul i64 %2, 4063299859190
46  %tobool6 = icmp eq i64 %mul, 0
47  %3 = sext i32 %0 to i64
48  br i1 %tobool, label %for.body.us.preheader, label %for.body.preheader
49
50for.body.preheader:                               ; preds = %for.body.lr.ph
51  br label %for.body
52
53for.body.us.preheader:                            ; preds = %for.body.lr.ph
54  br label %for.body.us
55
56for.body.us:                                      ; preds = %for.body.us.preheader, %for.cond.cleanup4.us-lcssa.us.us
57  %indvars.iv78 = phi i64 [ %indvars.iv.next79, %for.cond.cleanup4.us-lcssa.us.us ], [ 8, %for.body.us.preheader ]
58  %indvars.iv70 = phi i64 [ %indvars.iv.next71, %for.cond.cleanup4.us-lcssa.us.us ], [ 0, %for.body.us.preheader ]
59  %4 = sub nsw i64 8, %indvars.iv78
60  %add.ptr.us = getelementptr inbounds [10 x [10 x i32]], [10 x [10 x i32]]* @d, i64 0, i64 %indvars.iv78
61  %5 = add nsw i64 %4, %indvars.iv70
62  %arraydecay.us.us.us = getelementptr inbounds [10 x i32], [10 x i32]* %add.ptr.us, i64 %5, i64 0
63  br i1 %tobool6, label %for.body5.us.us.us.preheader, label %for.body5.us.us48.preheader
64
65for.body5.us.us48.preheader:                      ; preds = %for.body.us
66  store i32 8, i32* %arraydecay.us.us.us, align 16
67  %indvars.iv.next66 = or i64 %indvars.iv70, 1
68  %6 = add nsw i64 %4, %indvars.iv.next66
69  %arraydecay.us.us55.1 = getelementptr inbounds [10 x i32], [10 x i32]* %add.ptr.us, i64 %6, i64 0
70  store i32 8, i32* %arraydecay.us.us55.1, align 8
71  br label %for.cond.cleanup4.us-lcssa.us.us
72
73for.body5.us.us.us.preheader:                     ; preds = %for.body.us
74  store i32 7, i32* %arraydecay.us.us.us, align 16
75  %indvars.iv.next73 = or i64 %indvars.iv70, 1
76  %7 = add nsw i64 %4, %indvars.iv.next73
77  %arraydecay.us.us.us.1 = getelementptr inbounds [10 x i32], [10 x i32]* %add.ptr.us, i64 %7, i64 0
78  store i32 7, i32* %arraydecay.us.us.us.1, align 8
79  br label %for.cond.cleanup4.us-lcssa.us.us
80
81for.cond.cleanup4.us-lcssa.us.us:                 ; preds = %for.body5.us.us48.preheader, %for.body5.us.us.us.preheader
82  %indvars.iv.next79 = add nuw nsw i64 %indvars.iv78, 2
83  %cmp.us = icmp slt i64 %indvars.iv.next79, %3
84  %indvars.iv.next71 = add nuw nsw i64 %indvars.iv70, 2
85  br i1 %cmp.us, label %for.body.us, label %for.cond.cleanup.loopexit
86
87for.cond.cleanup.loopexit:                        ; preds = %for.cond.cleanup4.us-lcssa.us.us
88  br label %for.cond.cleanup
89
90for.cond.cleanup.loopexit99:                      ; preds = %for.body
91  br label %for.cond.cleanup
92
93for.cond.cleanup:                                 ; preds = %for.cond.cleanup.loopexit99, %for.cond.cleanup.loopexit, %entry
94  ret void
95
96for.body:                                         ; preds = %for.body.preheader, %for.body
97  %indvars.iv95 = phi i64 [ %indvars.iv.next96, %for.body ], [ 8, %for.body.preheader ]
98  %indvars.iv87 = phi i64 [ %indvars.iv.next88, %for.body ], [ 0, %for.body.preheader ]
99  %8 = sub nsw i64 8, %indvars.iv95
100  %add.ptr = getelementptr inbounds [10 x [10 x i32]], [10 x [10 x i32]]* @d, i64 0, i64 %indvars.iv95
101  %9 = add nsw i64 %8, %indvars.iv87
102  %arraydecay.us31 = getelementptr inbounds [10 x i32], [10 x i32]* %add.ptr, i64 %9, i64 0
103  store i32 8, i32* %arraydecay.us31, align 16
104  %indvars.iv.next90 = or i64 %indvars.iv87, 1
105  %10 = add nsw i64 %8, %indvars.iv.next90
106  %arraydecay.us31.1 = getelementptr inbounds [10 x i32], [10 x i32]* %add.ptr, i64 %10, i64 0
107  store i32 8, i32* %arraydecay.us31.1, align 8
108  %indvars.iv.next96 = add nuw nsw i64 %indvars.iv95, 2
109  %cmp = icmp slt i64 %indvars.iv.next96, %3
110  %indvars.iv.next88 = add nuw nsw i64 %indvars.iv87, 2
111  br i1 %cmp, label %for.body, label %for.cond.cleanup.loopexit99
112}
113
114attributes #0 = { norecurse nounwind ssp uwtable "disable-tail-calls"="false" "less-precise-fpmad"="false" "frame-pointer"="all" "no-infs-fp-math"="false" "no-jump-tables"="false" "no-nans-fp-math"="false" "stack-protector-buffer-size"="8" "target-cpu"="knl" "target-features"="+adx,+aes,+avx,+avx2,+avx512cd,+avx512er,+avx512f,+avx512pf,+bmi,+bmi2,+cx16,+f16c,+fma,+fsgsbase,+fxsr,+lzcnt,+mmx,+movbe,+pclmul,+popcnt,+prefetchwt1,+rdrnd,+rdseed,+rtm,+sse,+sse2,+sse3,+sse4.1,+sse4.2,+ssse3,+x87,+xsave,+xsaveopt" "unsafe-fp-math"="false" "use-soft-float"="false" }
115