Mercurial > vec
comparison src/impl/x86/avx512f.c @ 31:bf6ad516f1e6
Backed out changeset c6c99ab1088a
author | Paper <paper@tflc.us> |
---|---|
date | Fri, 25 Apr 2025 17:40:33 -0400 |
parents | c6c99ab1088a |
children | 99e4539f922f |
comparison
equal
deleted
inserted
replaced
30:641d8c79b1da | 31:bf6ad516f1e6 |
---|---|
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | 21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE |
22 * SOFTWARE. | 22 * SOFTWARE. |
23 **/ | 23 **/ |
24 | 24 |
25 #include "vec/impl/x86/avx512f.h" | 25 #include "vec/impl/x86/avx512f.h" |
26 #include "vec/impl/generic.h" | |
26 | 27 |
27 #include <immintrin.h> | 28 #include <immintrin.h> |
28 | 29 |
29 /* ------------------------------------------------------------------------ */ | 30 // this is a stupid amount of work just to do these operations, is it really worth it ? |
30 | 31 // also same note in avx2.c applies here, these do not handle sign bits properly, which |
31 #define VEC_AVX512F_MINMAX_TEMPLATE(SIGN, BITS, SIZE, INTLSIGN, OP) \ | 32 // isn't that big of a deal for regular arithmetic operations, but matters quite a bit |
32 VEC_FUNC_IMPL v##SIGN##int##BITS##x##SIZE v##SIGN##int##BITS##x##SIZE##_avx512f_##OP(v##SIGN##int##BITS##x##SIZE vec1, v##SIGN##int##BITS##x##SIZE vec2) \ | 33 // when doing things like arithmetic shifts. |
33 { \ | 34 #define VEC_AVX512F_OPERATION_8x64(op, sign) \ |
34 union v##SIGN##int##BITS##x##SIZE##_impl_data *vec1d = (union v##SIGN##int##BITS##x##SIZE##_impl_data *)&vec1; \ | 35 do { \ |
35 union v##SIGN##int##BITS##x##SIZE##_impl_data *vec2d = (union v##SIGN##int##BITS##x##SIZE##_impl_data *)&vec2; \ | 36 union v##sign##int8x64_impl_data *vec1d = (union v##sign##int8x64_impl_data *)&vec1; \ |
36 \ | 37 union v##sign##int8x64_impl_data *vec2d = (union v##sign##int8x64_impl_data *)&vec2; \ |
37 vec1d->avx512f = _mm512_##OP##_ep##INTLSIGN##BITS(vec1d->avx512f, vec2d->avx512f); \ | 38 \ |
38 \ | 39 /* unpack and operate */ \ |
39 return vec1d->vec; \ | 40 __m512i dst_1 = _mm512_##op##_epi32(_mm512_srli_epi32(_mm512_slli_epi32(vec1d->avx512f, 24), 24), _mm512_srli_epi32(_mm512_slli_epi32(vec2d->avx512f, 24), 24)); \ |
40 } | 41 __m512i dst_2 = _mm512_##op##_epi32(_mm512_srli_epi32(_mm512_slli_epi32(vec1d->avx512f, 16), 24), _mm512_srli_epi32(_mm512_slli_epi32(vec2d->avx512f, 16), 24)); \ |
41 | 42 __m512i dst_3 = _mm512_##op##_epi32(_mm512_srli_epi32(_mm512_slli_epi32(vec1d->avx512f, 8), 24), _mm512_srli_epi32(_mm512_slli_epi32(vec2d->avx512f, 8), 24)); \ |
42 #define VEC_AVX512F_MINMAX_32x16(OP) VEC_AVX512F_MINMAX_TEMPLATE( , 32, 16, i, OP) | 43 __m512i dst_4 = _mm512_##op##_epi32(_mm512_srli_epi32(vec1d->avx512f, 24), _mm512_srli_epi32(vec2d->avx512f, 24)); \ |
43 #define VEC_AVX512F_MINMAX_u32x16(OP) VEC_AVX512F_MINMAX_TEMPLATE(u, 32, 16, u, OP) | 44 \ |
44 #define VEC_AVX512F_MINMAX_64x8(OP) VEC_AVX512F_MINMAX_TEMPLATE( , 64, 8, i, OP) | 45 /* repack */ \ |
45 #define VEC_AVX512F_MINMAX_u64x8(OP) VEC_AVX512F_MINMAX_TEMPLATE(u, 64, 8, u, OP) | 46 vec1d->avx512f = _mm512_or_si512( \ |
46 | 47 _mm512_or_si512( \ |
47 #define VEC_AVX512F_STRUCT_MINMAX_32x16(OP, SIGN) v##SIGN##int32x16_avx512f_##OP | 48 _mm512_srli_epi32(_mm512_slli_epi32(dst_1, 24), 24), \ |
48 #define VEC_AVX512F_STRUCT_MINMAX_64x8(OP, SIGN) v##SIGN##int64x8_avx512f_##OP | 49 _mm512_srli_epi32(_mm512_slli_epi32(dst_2, 24), 16) \ |
49 | 50 ), \ |
50 /* ------------------------------------------------------------------------ */ | 51 _mm512_or_si512( \ |
51 | 52 _mm512_srli_epi32(_mm512_slli_epi32(dst_3, 24), 8), \ |
52 #define VEC_AVX512F_OPERATION_EX(name, op, sign, bits, size, secondsign) \ | 53 _mm512_slli_epi32(dst_4, 24) \ |
53 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_##name(v##sign##int##bits##x##size vec1, v##secondsign##int##bits##x##size vec2) \ | 54 ) \ |
54 { \ | 55 ); \ |
55 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ | 56 \ |
56 union v##secondsign##int##bits##x##size##_impl_data *vec2d = (union v##secondsign##int##bits##x##size##_impl_data *)&vec2; \ | 57 return vec1d->vec; \ |
57 \ | 58 } while (0) |
58 vec1d->avx512f = _mm512_##op##_epi##bits(vec1d->avx512f, vec2d->avx512f); \ | 59 |
59 \ | 60 #define VEC_AVX512F_OPERATION_16x32(op, sign) \ |
60 return vec1d->vec; \ | 61 do { \ |
61 } | 62 union v##sign##int16x32_impl_data *vec1d = (union v##sign##int16x32_impl_data *)&vec1; \ |
62 | 63 union v##sign##int16x32_impl_data *vec2d = (union v##sign##int16x32_impl_data *)&vec2; \ |
63 #define VEC_AVX512F_OPERATION(name, op, sign, bits, size) \ | 64 \ |
64 VEC_AVX512F_OPERATION_EX(name, op, sign, bits, size, sign) | 65 /* unpack and operate; it would be nice if we had an _m512_andi_epi32... */ \ |
65 | 66 __m512i dst_1 = _mm512_##op##_epi32(_mm512_srli_epi32(_mm512_slli_epi32(vec1d->avx512f, 16), 16), _mm512_srli_epi32(_mm512_slli_epi32(vec2d->avx512f, 16), 16)); \ |
66 #define VEC_AVX512F_OPERATION_SHIFT(name, op, sign, bits, size) \ | 67 __m512i dst_2 = _mm512_##op##_epi32(_mm512_srli_epi32(vec1d->avx512f, 16), _mm512_srli_epi32(vec2d->avx512f, 16)); \ |
67 VEC_AVX512F_OPERATION_EX(name, op, sign, bits, size, u) | 68 \ |
68 | 69 /* repack */ \ |
69 #define VEC_AVX512F_ADD_32x16(sign) VEC_AVX512F_OPERATION(add, add, sign, 32, 16) | 70 vec1d->avx512f = _mm512_or_si512( \ |
70 #define VEC_AVX512F_ADD_64x8(sign) VEC_AVX512F_OPERATION(add, add, sign, 64, 8) | 71 _mm512_srli_epi32(_mm512_slli_epi32(dst_1, 16), 16), \ |
71 | 72 _mm512_slli_epi32(dst_2, 16) \ |
72 #define VEC_AVX512F_SUB_32x16(sign) VEC_AVX512F_OPERATION(sub, sub, sign, 32, 16) | 73 ); \ |
73 #define VEC_AVX512F_SUB_64x8(sign) VEC_AVX512F_OPERATION(sub, sub, sign, 64, 8) | 74 return vec1d->vec; \ |
74 | 75 } while (0) |
75 #define VEC_AVX512F_MUL_32x16(sign) VEC_AVX512F_OPERATION(mul, mullo, sign, 32, 16) | 76 |
76 #define VEC_AVX512F_MUL_64x8(sign) VEC_AVX512F_OPERATION(mul, mullox, sign, 64, 8) | 77 #define VEC_AVX512F_ADD_8x64(sign) \ |
77 | 78 VEC_AVX512F_OPERATION_8x64(add, sign) |
78 #define VEC_AVX512F_LSHIFT_32x16(sign) VEC_AVX512F_OPERATION_SHIFT(lshift, sllv, sign, 32, 16) | 79 |
79 #define VEC_AVX512F_LSHIFT_64x8(sign) VEC_AVX512F_OPERATION_SHIFT(lshift, sllv, sign, 64, 8) | 80 #define VEC_AVX512F_ADD_16x32(sign) \ |
80 | 81 VEC_AVX512F_OPERATION_16x32(add, sign) |
81 #define VEC_AVX512F_XRSHIFT(name, bits, size, sign, aORl) \ | 82 |
82 VEC_AVX512F_OPERATION_SHIFT(name, sr##aORl##v, sign, bits, size) | 83 #define VEC_AVX512F_ADD_32x16(sign) \ |
83 | 84 do { \ |
84 /* always the same, regardless of signedness */ | 85 union v##sign##int32x16_impl_data *vec1d = (union v##sign##int32x16_impl_data *)&vec1; \ |
85 #define VEC_AVX512F_LRSHIFT_32x16(sign) VEC_AVX512F_XRSHIFT(lrshift, 32, 16, sign, l) | 86 union v##sign##int32x16_impl_data *vec2d = (union v##sign##int32x16_impl_data *)&vec2; \ |
86 #define VEC_AVX512F_LRSHIFT_64x8(sign) VEC_AVX512F_XRSHIFT(lrshift, 64, 8, sign, l) | 87 \ |
87 | 88 vec1d->avx512f = _mm512_add_epi32(vec1d->avx512f, vec2d->avx512f); \ |
88 #define VEC_AVX512F_RSHIFT_32x16(sign) VEC_AVX512F_XRSHIFT(rshift, 32, 16, sign, a) | 89 return vec1d->vec; \ |
89 #define VEC_AVX512F_RSHIFT_64x8(sign) VEC_AVX512F_XRSHIFT(rshift, 64, 8, sign, a) | 90 } while (0) |
90 | 91 |
91 #define VEC_AVX512F_uRSHIFT_32x16(sign) VEC_AVX512F_XRSHIFT(rshift, 32, 16, sign, l) | 92 #define VEC_AVX512F_ADD_64x8(sign) \ |
92 #define VEC_AVX512F_uRSHIFT_64x8(sign) VEC_AVX512F_XRSHIFT(rshift, 64, 8, sign, l) | 93 do { \ |
93 | 94 union v##sign##int64x8_impl_data *vec1d = (union v##sign##int64x8_impl_data *)&vec1; \ |
94 /* ok */ | 95 union v##sign##int64x8_impl_data *vec2d = (union v##sign##int64x8_impl_data *)&vec2; \ |
95 #define VEC_AVX512F_STRUCT_OPERATION_32x16(OP, SIGN) v##SIGN##int32x16_avx512f_##OP | 96 \ |
96 #define VEC_AVX512F_STRUCT_OPERATION_64x8(OP, SIGN) v##SIGN##int64x8_avx512f_##OP | 97 vec1d->avx512f = _mm512_add_epi64(vec1d->avx512f, vec2d->avx512f); \ |
97 | 98 return vec1d->vec; \ |
98 /* ------------------------------------------------------------------------ */ | 99 } while (0) |
100 | |
101 #define VEC_AVX512F_SUB_8x64(sign) \ | |
102 VEC_AVX512F_OPERATION_8x64(sub, sign) | |
103 | |
104 #define VEC_AVX512F_SUB_16x32(sign) \ | |
105 VEC_AVX512F_OPERATION_16x32(sub, sign) | |
106 | |
107 #define VEC_AVX512F_SUB_32x16(sign) \ | |
108 do { \ | |
109 union v##sign##int32x16_impl_data *vec1d = (union v##sign##int32x16_impl_data *)&vec1; \ | |
110 union v##sign##int32x16_impl_data *vec2d = (union v##sign##int32x16_impl_data *)&vec2; \ | |
111 \ | |
112 vec1d->avx512f = _mm512_sub_epi32(vec1d->avx512f, vec2d->avx512f); \ | |
113 return vec1d->vec; \ | |
114 } while (0) | |
115 | |
116 #define VEC_AVX512F_SUB_64x8(sign) \ | |
117 do { \ | |
118 union v##sign##int64x8_impl_data *vec1d = (union v##sign##int64x8_impl_data *)&vec1; \ | |
119 union v##sign##int64x8_impl_data *vec2d = (union v##sign##int64x8_impl_data *)&vec2; \ | |
120 \ | |
121 vec1d->avx512f = _mm512_sub_epi64(vec1d->avx512f, vec2d->avx512f); \ | |
122 return vec1d->vec; \ | |
123 } while (0) | |
124 | |
125 #define VEC_AVX512F_MUL_8x64(sign) \ | |
126 VEC_AVX512F_OPERATION_8x64(mullo, sign) | |
127 | |
128 #define VEC_AVX512F_MUL_16x32(sign) \ | |
129 VEC_AVX512F_OPERATION_16x32(mullo, sign) | |
130 | |
131 #define VEC_AVX512F_MUL_32x16(sign) \ | |
132 do { \ | |
133 union v##sign##int32x16_impl_data *vec1d = (union v##sign##int32x16_impl_data *)&vec1; \ | |
134 union v##sign##int32x16_impl_data *vec2d = (union v##sign##int32x16_impl_data *)&vec2; \ | |
135 \ | |
136 vec1d->avx512f = _mm512_mullo_epi32(vec1d->avx512f, vec2d->avx512f); \ | |
137 return vec1d->vec; \ | |
138 } while (0) | |
139 | |
140 #define VEC_AVX512F_MUL_64x8(sign) \ | |
141 do { \ | |
142 union v##sign##int64x8_impl_data *vec1d = (union v##sign##int64x8_impl_data *)&vec1; \ | |
143 union v##sign##int64x8_impl_data *vec2d = (union v##sign##int64x8_impl_data *)&vec2; \ | |
144 \ | |
145 __m512i ac = _mm512_mul_epu32(vec1d->avx512f, vec2d->avx512f); \ | |
146 __m512i b = _mm512_srli_epi64(vec1d->avx512f, 32); \ | |
147 __m512i bc = _mm512_mul_epu32(b, vec2d->avx512f); \ | |
148 __m512i d = _mm512_srli_epi64(vec2d->avx512f, 32); \ | |
149 __m512i ad = _mm512_mul_epu32(vec1d->avx512f, d); \ | |
150 __m512i hi = _mm512_add_epi64(bc, ad); \ | |
151 hi = _mm512_slli_epi64(hi, 32); \ | |
152 \ | |
153 vec1d->avx512f = _mm512_add_epi64(hi, ac); \ | |
154 return vec1d->vec; \ | |
155 } while (0) | |
156 | |
157 #define VEC_AVX512F_LSHIFT_8x64(sign) \ | |
158 VEC_AVX512F_OPERATION_8x64(sllv, sign) | |
159 | |
160 #define VEC_AVX512F_LSHIFT_16x32(sign) \ | |
161 VEC_AVX512F_OPERATION_16x32(sllv, sign) | |
162 | |
163 #define VEC_AVX512F_LSHIFT_32x16(sign) \ | |
164 do { \ | |
165 union v##sign##int32x16_impl_data *vec1d = (union v##sign##int32x16_impl_data *)&vec1; \ | |
166 union v##sign##int32x16_impl_data *vec2d = (union v##sign##int32x16_impl_data *)&vec2; \ | |
167 \ | |
168 vec1d->avx512f = _mm512_sllv_epi32(vec1d->avx512f, vec2d->avx512f); \ | |
169 return vec1d->vec; \ | |
170 } while (0) | |
171 | |
172 #define VEC_AVX512F_LSHIFT_64x8(sign) \ | |
173 do { \ | |
174 union v##sign##int64x8_impl_data *vec1d = (union v##sign##int64x8_impl_data *)&vec1; \ | |
175 union v##sign##int64x8_impl_data *vec2d = (union v##sign##int64x8_impl_data *)&vec2; \ | |
176 \ | |
177 vec1d->avx512f = _mm512_sllv_epi64(vec1d->avx512f, vec2d->avx512f); \ | |
178 return vec1d->vec; \ | |
179 } while (0) | |
180 | |
181 #define VEC_AVX512F_lRSHIFT_8x64(sign) \ | |
182 VEC_AVX512F_OPERATION_8x64(srlv, sign) | |
183 | |
184 #define VEC_AVX512F_lRSHIFT_16x32(sign) \ | |
185 VEC_AVX512F_OPERATION_16x32(srlv, sign) | |
186 | |
187 #define VEC_AVX512F_aRSHIFT_8x64(sign) \ | |
188 do { \ | |
189 return v##sign##int8x64_generic_rshift(vec1, vec2); \ | |
190 } while (0) | |
191 | |
192 #define VEC_AVX512F_aRSHIFT_16x32(sign) \ | |
193 do { \ | |
194 return v##sign##int16x32_generic_rshift(vec1, vec2); \ | |
195 } while (0) | |
196 | |
197 #define VEC_AVX512F_RSHIFT_8x64(sign, aORl) VEC_AVX512F_##aORl##RSHIFT_8x64(sign) | |
198 #define VEC_AVX512F_RSHIFT_16x32(sign, aORl) VEC_AVX512F_##aORl##RSHIFT_16x32(sign) | |
199 | |
200 #define VEC_AVX512F_RSHIFT_32x16(sign, aORl) \ | |
201 do { \ | |
202 union v##sign##int32x16_impl_data *vec1d = (union v##sign##int32x16_impl_data *)&vec1; \ | |
203 union v##sign##int32x16_impl_data *vec2d = (union v##sign##int32x16_impl_data *)&vec2; \ | |
204 \ | |
205 vec1d->avx512f = _mm512_sr##aORl##v_epi32(vec1d->avx512f, vec2d->avx512f); \ | |
206 return vec1d->vec; \ | |
207 } while (0) | |
208 | |
209 #define VEC_AVX512F_RSHIFT_64x8(sign, aORl) \ | |
210 do { \ | |
211 union v##sign##int64x8_impl_data *vec1d = (union v##sign##int64x8_impl_data *)&vec1; \ | |
212 union v##sign##int64x8_impl_data *vec2d = (union v##sign##int64x8_impl_data *)&vec2; \ | |
213 \ | |
214 vec1d->avx512f = _mm512_sr##aORl##v_epi64(vec1d->avx512f, vec2d->avx512f); \ | |
215 return vec1d->vec; \ | |
216 } while (0) | |
217 | |
218 #define VEC_AVX512F_uRSHIFT_8x64(sign, aORl) VEC_AVX512F_RSHIFT_8x64(sign, l) | |
219 #define VEC_AVX512F_uRSHIFT_16x32(sign, aORl) VEC_AVX512F_RSHIFT_16x32(sign, l) | |
220 #define VEC_AVX512F_uRSHIFT_32x16(sign, aORl) VEC_AVX512F_RSHIFT_32x16(sign, l) | |
221 #define VEC_AVX512F_uRSHIFT_64x8(sign, aORl) VEC_AVX512F_RSHIFT_64x8(sign, l) | |
99 | 222 |
100 #define VEC_AVX512F_DEFINE_OPERATIONS_SIGN(sign, bits, size) \ | 223 #define VEC_AVX512F_DEFINE_OPERATIONS_SIGN(sign, bits, size) \ |
101 union v##sign##int##bits##x##size##_impl_data { \ | 224 union v##sign##int##bits##x##size##_impl_data { \ |
102 v##sign##int##bits##x##size vec; \ | 225 v##sign##int##bits##x##size vec; \ |
103 __m512i avx512f; \ | 226 __m512i avx512f; \ |
104 }; \ | 227 }; \ |
105 \ | 228 \ |
106 VEC_STATIC_ASSERT(VEC_ALIGNOF(__m512i) <= VEC_ALIGNOF(v##sign##int##bits##x##size), "vec: v" #sign "int" #bits "x" #size " alignment needs to be expanded to fit intrinsic type size"); \ | 229 VEC_STATIC_ASSERT(VEC_ALIGNOF(__m512i) <= VEC_ALIGNOF(v##sign##int##bits##x##size), "vec: v" #sign "int" #bits "x" #size " alignment needs to be expanded to fit intrinsic type size"); \ |
107 VEC_STATIC_ASSERT(sizeof(__m512i) <= sizeof(v##sign##int##bits##x##size), "vec: v" #sign "int" #bits "x" #size " needs to be expanded to fit intrinsic type size"); \ | 230 VEC_STATIC_ASSERT(sizeof(__m512i) <= sizeof(v##sign##int##bits##x##size), "vec: v" #sign "int" #bits "x" #size " needs to be expanded to fit intrinsic type size"); \ |
108 \ | 231 \ |
109 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_load_aligned(const vec_##sign##int##bits in[size]) \ | 232 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_load_aligned(const vec_##sign##int##bits in[size]) \ |
110 { \ | 233 { \ |
111 union v##sign##int##bits##x##size##_impl_data vec; \ | 234 union v##sign##int##bits##x##size##_impl_data vec; \ |
112 vec.avx512f = _mm512_load_si512((const __m512i *)in); \ | 235 vec.avx512f = _mm512_load_si512((const __m512i *)in); \ |
113 return vec.vec; \ | 236 return vec.vec; \ |
114 } \ | 237 } \ |
115 \ | 238 \ |
116 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_load(const vec_##sign##int##bits in[size]) \ | 239 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_load(const vec_##sign##int##bits in[size]) \ |
117 { \ | 240 { \ |
118 union v##sign##int##bits##x##size##_impl_data vec; \ | 241 union v##sign##int##bits##x##size##_impl_data vec; \ |
119 vec.avx512f = _mm512_loadu_si512((const __m512i *)in); \ | 242 vec.avx512f = _mm512_loadu_si512((const __m512i *)in); \ |
120 return vec.vec; \ | 243 return vec.vec; \ |
121 } \ | 244 } \ |
122 \ | 245 \ |
123 VEC_FUNC_IMPL void v##sign##int##bits##x##size##_avx512f_store_aligned(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ | 246 static void v##sign##int##bits##x##size##_avx512f_store_aligned(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ |
124 { \ | 247 { \ |
125 _mm512_store_si512((__m512i *)out, ((union v##sign##int##bits##x##size##_impl_data *)&vec)->avx512f); \ | 248 _mm512_store_si512((__m512i *)out, ((union v##sign##int##bits##x##size##_impl_data *)&vec)->avx512f); \ |
126 } \ | 249 } \ |
127 \ | 250 \ |
128 VEC_FUNC_IMPL void v##sign##int##bits##x##size##_avx512f_store(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ | 251 static void v##sign##int##bits##x##size##_avx512f_store(v##sign##int##bits##x##size vec, vec_##sign##int##bits out[size]) \ |
129 { \ | 252 { \ |
130 _mm512_storeu_si512((__m512i *)out, ((union v##sign##int##bits##x##size##_impl_data *)&vec)->avx512f); \ | 253 _mm512_storeu_si512((__m512i *)out, ((union v##sign##int##bits##x##size##_impl_data *)&vec)->avx512f); \ |
131 } \ | 254 } \ |
132 \ | 255 \ |
133 VEC_AVX512F_ADD_##bits##x##size(sign) \ | 256 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_add(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ |
134 VEC_AVX512F_SUB_##bits##x##size(sign) \ | 257 { \ |
135 VEC_AVX512F_MUL_##bits##x##size(sign) \ | 258 VEC_AVX512F_ADD_##bits##x##size(sign); \ |
136 \ | 259 } \ |
137 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_and(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | 260 \ |
261 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_sub(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | |
262 { \ | |
263 VEC_AVX512F_SUB_##bits##x##size(sign); \ | |
264 } \ | |
265 \ | |
266 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_mul(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | |
267 { \ | |
268 VEC_AVX512F_MUL_##bits##x##size(sign); \ | |
269 } \ | |
270 \ | |
271 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_and(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | |
138 { \ | 272 { \ |
139 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ | 273 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ |
140 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ | 274 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ |
141 \ | 275 \ |
142 vec1d->avx512f = _mm512_and_si512(vec1d->avx512f, vec2d->avx512f); \ | 276 vec1d->avx512f = _mm512_and_si512(vec1d->avx512f, vec2d->avx512f); \ |
143 return vec1d->vec; \ | 277 return vec1d->vec; \ |
144 } \ | 278 } \ |
145 \ | 279 \ |
146 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_or(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | 280 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_or(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ |
147 { \ | 281 { \ |
148 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ | 282 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ |
149 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ | 283 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ |
150 \ | 284 \ |
151 vec1d->avx512f = _mm512_or_si512(vec1d->avx512f, vec2d->avx512f); \ | 285 vec1d->avx512f = _mm512_or_si512(vec1d->avx512f, vec2d->avx512f); \ |
152 return vec1d->vec; \ | 286 return vec1d->vec; \ |
153 } \ | 287 } \ |
154 \ | 288 \ |
155 VEC_FUNC_IMPL v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_xor(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ | 289 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_xor(v##sign##int##bits##x##size vec1, v##sign##int##bits##x##size vec2) \ |
156 { \ | 290 { \ |
157 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ | 291 union v##sign##int##bits##x##size##_impl_data *vec1d = (union v##sign##int##bits##x##size##_impl_data *)&vec1; \ |
158 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ | 292 union v##sign##int##bits##x##size##_impl_data *vec2d = (union v##sign##int##bits##x##size##_impl_data *)&vec2; \ |
159 \ | 293 \ |
160 vec1d->avx512f = _mm512_xor_si512(vec1d->avx512f, vec2d->avx512f); \ | 294 vec1d->avx512f = _mm512_xor_si512(vec1d->avx512f, vec2d->avx512f); \ |
161 return vec1d->vec; \ | 295 return vec1d->vec; \ |
162 } \ | 296 } \ |
163 \ | 297 \ |
164 VEC_AVX512F_LSHIFT_##bits##x##size(sign); \ | 298 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_lshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ |
165 VEC_AVX512F_##sign##RSHIFT_##bits##x##size(sign); \ | 299 { \ |
166 VEC_AVX512F_LRSHIFT_##bits##x##size(sign); \ | 300 VEC_AVX512F_LSHIFT_##bits##x##size(sign); \ |
167 \ | 301 } \ |
168 VEC_AVX512F_MINMAX_##sign##bits##x##size(min) \ | 302 \ |
169 VEC_AVX512F_MINMAX_##sign##bits##x##size(max) \ | 303 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_rshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ |
304 { \ | |
305 VEC_AVX512F_##sign##RSHIFT_##bits##x##size(sign, a); \ | |
306 } \ | |
307 \ | |
308 static v##sign##int##bits##x##size v##sign##int##bits##x##size##_avx512f_lrshift(v##sign##int##bits##x##size vec1, vuint##bits##x##size vec2) \ | |
309 { \ | |
310 VEC_AVX512F_RSHIFT_##bits##x##size(sign, l); \ | |
311 } \ | |
170 \ | 312 \ |
171 const v##sign##int##bits##x##size##_impl v##sign##int##bits##x##size##_impl_avx512f = { \ | 313 const v##sign##int##bits##x##size##_impl v##sign##int##bits##x##size##_impl_avx512f = { \ |
172 .load_aligned = v##sign##int##bits##x##size##_avx512f_load_aligned, \ | 314 v##sign##int##bits##x##size##_generic_splat, \ |
173 .load = v##sign##int##bits##x##size##_avx512f_load, \ | 315 v##sign##int##bits##x##size##_avx512f_load_aligned, \ |
174 .store_aligned = v##sign##int##bits##x##size##_avx512f_store_aligned, \ | 316 v##sign##int##bits##x##size##_avx512f_load, \ |
175 .store = v##sign##int##bits##x##size##_avx512f_store, \ | 317 v##sign##int##bits##x##size##_avx512f_store_aligned, \ |
176 .add = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(add, sign), \ | 318 v##sign##int##bits##x##size##_avx512f_store, \ |
177 .sub = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(sub, sign), \ | 319 v##sign##int##bits##x##size##_avx512f_add, \ |
178 .mul = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(mul, sign), \ | 320 v##sign##int##bits##x##size##_avx512f_sub, \ |
179 .band = v##sign##int##bits##x##size##_avx512f_and, \ | 321 v##sign##int##bits##x##size##_avx512f_mul, \ |
180 .bor = v##sign##int##bits##x##size##_avx512f_or, \ | 322 v##sign##int##bits##x##size##_generic_div, \ |
181 .bxor = v##sign##int##bits##x##size##_avx512f_xor, \ | 323 v##sign##int##bits##x##size##_generic_avg, \ |
182 .lshift = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(lshift, sign), \ | 324 v##sign##int##bits##x##size##_avx512f_and, \ |
183 .rshift = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(rshift, sign), \ | 325 v##sign##int##bits##x##size##_avx512f_or, \ |
184 .lrshift = VEC_AVX512F_STRUCT_OPERATION_##bits##x##size(lrshift, sign), \ | 326 v##sign##int##bits##x##size##_avx512f_xor, \ |
185 .min = VEC_AVX512F_STRUCT_MINMAX_##bits##x##size(min, sign), \ | 327 v##sign##int##bits##x##size##_generic_not, \ |
186 .max = VEC_AVX512F_STRUCT_MINMAX_##bits##x##size(max, sign), \ | 328 v##sign##int##bits##x##size##_avx512f_lshift, \ |
329 v##sign##int##bits##x##size##_avx512f_rshift, \ | |
330 v##sign##int##bits##x##size##_avx512f_lrshift, \ | |
331 v##sign##int##bits##x##size##_generic_cmplt, \ | |
332 v##sign##int##bits##x##size##_generic_cmple, \ | |
333 v##sign##int##bits##x##size##_generic_cmpeq, \ | |
334 v##sign##int##bits##x##size##_generic_cmpge, \ | |
335 v##sign##int##bits##x##size##_generic_cmpgt, \ | |
187 }; | 336 }; |
188 | 337 |
189 #define VEC_AVX512F_DEFINE_OPERATIONS(bits, size) \ | 338 #define VEC_AVX512F_DEFINE_OPERATIONS(bits, size) \ |
190 VEC_AVX512F_DEFINE_OPERATIONS_SIGN(u, bits, size) \ | 339 VEC_AVX512F_DEFINE_OPERATIONS_SIGN(u, bits, size) \ |
191 VEC_AVX512F_DEFINE_OPERATIONS_SIGN( , bits, size) | 340 VEC_AVX512F_DEFINE_OPERATIONS_SIGN( , bits, size) |
192 | 341 |
342 VEC_AVX512F_DEFINE_OPERATIONS(8, 64) | |
343 VEC_AVX512F_DEFINE_OPERATIONS(16, 32) | |
193 VEC_AVX512F_DEFINE_OPERATIONS(32, 16) | 344 VEC_AVX512F_DEFINE_OPERATIONS(32, 16) |
194 VEC_AVX512F_DEFINE_OPERATIONS(64, 8) | 345 VEC_AVX512F_DEFINE_OPERATIONS(64, 8) |